00:00:00.000 Started by upstream project "autotest-per-patch" build number 126167 00:00:00.000 originally caused by: 00:00:00.000 Started by upstream project "jbp-per-patch" build number 23924 00:00:00.000 originally caused by: 00:00:00.001 Started by user sys_sgci 00:00:00.015 Checking out git https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool into /var/jenkins_home/workspace/crypto-phy-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4 to read jbp/jenkins/jjb-config/jobs/autotest-downstream/autotest-phy.groovy 00:00:00.015 The recommended git tool is: git 00:00:00.015 using credential 00000000-0000-0000-0000-000000000002 00:00:00.017 > git rev-parse --resolve-git-dir /var/jenkins_home/workspace/crypto-phy-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4/jbp/.git # timeout=10 00:00:00.029 Fetching changes from the remote Git repository 00:00:00.030 > git config remote.origin.url https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool # timeout=10 00:00:00.045 Using shallow fetch with depth 1 00:00:00.045 Fetching upstream changes from https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool 00:00:00.045 > git --version # timeout=10 00:00:00.065 > git --version # 'git version 2.39.2' 00:00:00.065 using GIT_ASKPASS to set credentials SPDKCI HTTPS Credentials 00:00:00.095 Setting http proxy: proxy-dmz.intel.com:911 00:00:00.095 > git fetch --tags --force --progress --depth=1 -- https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool refs/changes/75/21875/22 # timeout=5 00:00:02.560 > git rev-parse origin/FETCH_HEAD^{commit} # timeout=10 00:00:02.571 > git rev-parse FETCH_HEAD^{commit} # timeout=10 00:00:02.582 Checking out Revision 8c6732c9e0fe7c9c74cd1fb560a619e554726af3 (FETCH_HEAD) 00:00:02.582 > git config core.sparsecheckout # timeout=10 00:00:02.593 > git read-tree -mu HEAD # timeout=10 00:00:02.608 > git checkout -f 8c6732c9e0fe7c9c74cd1fb560a619e554726af3 # timeout=5 00:00:02.632 Commit message: "jenkins/jjb-config: Remove SPDK_TEST_RELEASE_BUILD from packaging job" 00:00:02.633 > git rev-list --no-walk 9bf0dabeadcf84e29a3d5dbec2430e38aceadf8d # timeout=10 00:00:02.744 [Pipeline] Start of Pipeline 00:00:02.755 [Pipeline] library 00:00:02.757 Loading library shm_lib@master 00:00:02.757 Library shm_lib@master is cached. Copying from home. 00:00:02.772 [Pipeline] node 00:00:02.787 Running on WFP50 in /var/jenkins/workspace/crypto-phy-autotest 00:00:02.789 [Pipeline] { 00:00:02.801 [Pipeline] catchError 00:00:02.802 [Pipeline] { 00:00:02.816 [Pipeline] wrap 00:00:02.827 [Pipeline] { 00:00:02.835 [Pipeline] stage 00:00:02.837 [Pipeline] { (Prologue) 00:00:03.030 [Pipeline] sh 00:00:03.306 + logger -p user.info -t JENKINS-CI 00:00:03.322 [Pipeline] echo 00:00:03.323 Node: WFP50 00:00:03.330 [Pipeline] sh 00:00:03.617 [Pipeline] setCustomBuildProperty 00:00:03.629 [Pipeline] echo 00:00:03.631 Cleanup processes 00:00:03.637 [Pipeline] sh 00:00:03.917 + sudo pgrep -af /var/jenkins/workspace/crypto-phy-autotest/spdk 00:00:03.917 704699 sudo pgrep -af /var/jenkins/workspace/crypto-phy-autotest/spdk 00:00:03.930 [Pipeline] sh 00:00:04.210 ++ sudo pgrep -af /var/jenkins/workspace/crypto-phy-autotest/spdk 00:00:04.210 ++ grep -v 'sudo pgrep' 00:00:04.210 ++ awk '{print $1}' 00:00:04.210 + sudo kill -9 00:00:04.210 + true 00:00:04.225 [Pipeline] cleanWs 00:00:04.235 [WS-CLEANUP] Deleting project workspace... 00:00:04.235 [WS-CLEANUP] Deferred wipeout is used... 00:00:04.242 [WS-CLEANUP] done 00:00:04.247 [Pipeline] setCustomBuildProperty 00:00:04.264 [Pipeline] sh 00:00:04.543 + sudo git config --global --replace-all safe.directory '*' 00:00:04.605 [Pipeline] httpRequest 00:00:04.618 [Pipeline] echo 00:00:04.619 Sorcerer 10.211.164.101 is alive 00:00:04.624 [Pipeline] httpRequest 00:00:04.627 HttpMethod: GET 00:00:04.628 URL: http://10.211.164.101/packages/jbp_8c6732c9e0fe7c9c74cd1fb560a619e554726af3.tar.gz 00:00:04.628 Sending request to url: http://10.211.164.101/packages/jbp_8c6732c9e0fe7c9c74cd1fb560a619e554726af3.tar.gz 00:00:04.639 Response Code: HTTP/1.1 200 OK 00:00:04.639 Success: Status code 200 is in the accepted range: 200,404 00:00:04.640 Saving response body to /var/jenkins/workspace/crypto-phy-autotest/jbp_8c6732c9e0fe7c9c74cd1fb560a619e554726af3.tar.gz 00:00:06.140 [Pipeline] sh 00:00:06.419 + tar --no-same-owner -xf jbp_8c6732c9e0fe7c9c74cd1fb560a619e554726af3.tar.gz 00:00:06.690 [Pipeline] httpRequest 00:00:06.714 [Pipeline] echo 00:00:06.715 Sorcerer 10.211.164.101 is alive 00:00:06.721 [Pipeline] httpRequest 00:00:06.725 HttpMethod: GET 00:00:06.726 URL: http://10.211.164.101/packages/spdk_e7cce062d7bcec53f8a0237bb456695749792008.tar.gz 00:00:06.726 Sending request to url: http://10.211.164.101/packages/spdk_e7cce062d7bcec53f8a0237bb456695749792008.tar.gz 00:00:06.743 Response Code: HTTP/1.1 200 OK 00:00:06.744 Success: Status code 200 is in the accepted range: 200,404 00:00:06.745 Saving response body to /var/jenkins/workspace/crypto-phy-autotest/spdk_e7cce062d7bcec53f8a0237bb456695749792008.tar.gz 00:01:22.895 [Pipeline] sh 00:01:23.170 + tar --no-same-owner -xf spdk_e7cce062d7bcec53f8a0237bb456695749792008.tar.gz 00:01:26.465 [Pipeline] sh 00:01:26.793 + git -C spdk log --oneline -n5 00:01:26.793 e7cce062d Examples/Perf: correct the calculation of total bandwidth 00:01:26.793 3b4b1d00c libvfio-user: bump MAX_DMA_REGIONS 00:01:26.793 32a79de81 lib/event: add disable_cpumask_locks to spdk_app_opts 00:01:26.793 719d03c6a sock/uring: only register net impl if supported 00:01:26.793 e64f085ad vbdev_lvol_ut: unify usage of dummy base bdev 00:01:26.804 [Pipeline] } 00:01:26.821 [Pipeline] // stage 00:01:26.830 [Pipeline] stage 00:01:26.833 [Pipeline] { (Prepare) 00:01:26.853 [Pipeline] writeFile 00:01:26.872 [Pipeline] sh 00:01:27.152 + logger -p user.info -t JENKINS-CI 00:01:27.164 [Pipeline] sh 00:01:27.444 + logger -p user.info -t JENKINS-CI 00:01:27.460 [Pipeline] sh 00:01:27.744 + cat autorun-spdk.conf 00:01:27.744 SPDK_RUN_FUNCTIONAL_TEST=1 00:01:27.744 SPDK_TEST_BLOCKDEV=1 00:01:27.744 SPDK_TEST_ISAL=1 00:01:27.744 SPDK_TEST_CRYPTO=1 00:01:27.744 SPDK_TEST_REDUCE=1 00:01:27.744 SPDK_TEST_VBDEV_COMPRESS=1 00:01:27.744 SPDK_RUN_UBSAN=1 00:01:27.751 RUN_NIGHTLY=0 00:01:27.756 [Pipeline] readFile 00:01:27.780 [Pipeline] withEnv 00:01:27.782 [Pipeline] { 00:01:27.796 [Pipeline] sh 00:01:28.078 + set -ex 00:01:28.078 + [[ -f /var/jenkins/workspace/crypto-phy-autotest/autorun-spdk.conf ]] 00:01:28.078 + source /var/jenkins/workspace/crypto-phy-autotest/autorun-spdk.conf 00:01:28.078 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:01:28.078 ++ SPDK_TEST_BLOCKDEV=1 00:01:28.078 ++ SPDK_TEST_ISAL=1 00:01:28.078 ++ SPDK_TEST_CRYPTO=1 00:01:28.078 ++ SPDK_TEST_REDUCE=1 00:01:28.078 ++ SPDK_TEST_VBDEV_COMPRESS=1 00:01:28.078 ++ SPDK_RUN_UBSAN=1 00:01:28.078 ++ RUN_NIGHTLY=0 00:01:28.078 + case $SPDK_TEST_NVMF_NICS in 00:01:28.078 + DRIVERS= 00:01:28.078 + [[ -n '' ]] 00:01:28.078 + exit 0 00:01:28.086 [Pipeline] } 00:01:28.105 [Pipeline] // withEnv 00:01:28.111 [Pipeline] } 00:01:28.127 [Pipeline] // stage 00:01:28.137 [Pipeline] catchError 00:01:28.139 [Pipeline] { 00:01:28.155 [Pipeline] timeout 00:01:28.155 Timeout set to expire in 40 min 00:01:28.157 [Pipeline] { 00:01:28.172 [Pipeline] stage 00:01:28.174 [Pipeline] { (Tests) 00:01:28.190 [Pipeline] sh 00:01:28.471 + jbp/jenkins/jjb-config/jobs/scripts/autoruner.sh /var/jenkins/workspace/crypto-phy-autotest 00:01:28.471 ++ readlink -f /var/jenkins/workspace/crypto-phy-autotest 00:01:28.471 + DIR_ROOT=/var/jenkins/workspace/crypto-phy-autotest 00:01:28.471 + [[ -n /var/jenkins/workspace/crypto-phy-autotest ]] 00:01:28.471 + DIR_SPDK=/var/jenkins/workspace/crypto-phy-autotest/spdk 00:01:28.471 + DIR_OUTPUT=/var/jenkins/workspace/crypto-phy-autotest/output 00:01:28.471 + [[ -d /var/jenkins/workspace/crypto-phy-autotest/spdk ]] 00:01:28.471 + [[ ! -d /var/jenkins/workspace/crypto-phy-autotest/output ]] 00:01:28.471 + mkdir -p /var/jenkins/workspace/crypto-phy-autotest/output 00:01:28.471 + [[ -d /var/jenkins/workspace/crypto-phy-autotest/output ]] 00:01:28.471 + [[ crypto-phy-autotest == pkgdep-* ]] 00:01:28.471 + cd /var/jenkins/workspace/crypto-phy-autotest 00:01:28.471 + source /etc/os-release 00:01:28.471 ++ NAME='Fedora Linux' 00:01:28.471 ++ VERSION='38 (Cloud Edition)' 00:01:28.471 ++ ID=fedora 00:01:28.471 ++ VERSION_ID=38 00:01:28.471 ++ VERSION_CODENAME= 00:01:28.471 ++ PLATFORM_ID=platform:f38 00:01:28.471 ++ PRETTY_NAME='Fedora Linux 38 (Cloud Edition)' 00:01:28.471 ++ ANSI_COLOR='0;38;2;60;110;180' 00:01:28.471 ++ LOGO=fedora-logo-icon 00:01:28.471 ++ CPE_NAME=cpe:/o:fedoraproject:fedora:38 00:01:28.471 ++ HOME_URL=https://fedoraproject.org/ 00:01:28.471 ++ DOCUMENTATION_URL=https://docs.fedoraproject.org/en-US/fedora/f38/system-administrators-guide/ 00:01:28.471 ++ SUPPORT_URL=https://ask.fedoraproject.org/ 00:01:28.471 ++ BUG_REPORT_URL=https://bugzilla.redhat.com/ 00:01:28.471 ++ REDHAT_BUGZILLA_PRODUCT=Fedora 00:01:28.471 ++ REDHAT_BUGZILLA_PRODUCT_VERSION=38 00:01:28.471 ++ REDHAT_SUPPORT_PRODUCT=Fedora 00:01:28.471 ++ REDHAT_SUPPORT_PRODUCT_VERSION=38 00:01:28.471 ++ SUPPORT_END=2024-05-14 00:01:28.471 ++ VARIANT='Cloud Edition' 00:01:28.471 ++ VARIANT_ID=cloud 00:01:28.471 + uname -a 00:01:28.471 Linux spdk-wfp-50 6.7.0-68.fc38.x86_64 #1 SMP PREEMPT_DYNAMIC Mon Jan 15 00:59:40 UTC 2024 x86_64 GNU/Linux 00:01:28.471 + sudo /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh status 00:01:31.754 0000:d7:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:d7:05.5 00:01:31.754 0000:85:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:85:05.5 00:01:31.754 Hugepages 00:01:31.754 node hugesize free / total 00:01:31.754 node0 1048576kB 0 / 0 00:01:31.754 node0 2048kB 0 / 0 00:01:31.754 node1 1048576kB 0 / 0 00:01:31.754 node1 2048kB 0 / 0 00:01:31.754 00:01:31.754 Type BDF Vendor Device NUMA Driver Device Block devices 00:01:31.754 I/OAT 0000:00:04.0 8086 2021 0 ioatdma - - 00:01:31.754 I/OAT 0000:00:04.1 8086 2021 0 ioatdma - - 00:01:31.754 I/OAT 0000:00:04.2 8086 2021 0 ioatdma - - 00:01:31.754 I/OAT 0000:00:04.3 8086 2021 0 ioatdma - - 00:01:31.754 I/OAT 0000:00:04.4 8086 2021 0 ioatdma - - 00:01:31.754 I/OAT 0000:00:04.5 8086 2021 0 ioatdma - - 00:01:31.754 I/OAT 0000:00:04.6 8086 2021 0 ioatdma - - 00:01:31.754 I/OAT 0000:00:04.7 8086 2021 0 ioatdma - - 00:01:31.754 NVMe 0000:5e:00.0 8086 0b60 0 nvme nvme0 nvme0n1 00:01:31.754 I/OAT 0000:80:04.0 8086 2021 1 ioatdma - - 00:01:31.754 I/OAT 0000:80:04.1 8086 2021 1 ioatdma - - 00:01:31.754 I/OAT 0000:80:04.2 8086 2021 1 ioatdma - - 00:01:31.754 I/OAT 0000:80:04.3 8086 2021 1 ioatdma - - 00:01:31.754 I/OAT 0000:80:04.4 8086 2021 1 ioatdma - - 00:01:31.754 I/OAT 0000:80:04.5 8086 2021 1 ioatdma - - 00:01:31.754 I/OAT 0000:80:04.6 8086 2021 1 ioatdma - - 00:01:31.754 I/OAT 0000:80:04.7 8086 2021 1 ioatdma - - 00:01:31.754 VMD 0000:85:05.5 8086 201d 1 vfio-pci - - 00:01:31.754 VMD 0000:d7:05.5 8086 201d 1 vfio-pci - - 00:01:31.754 + rm -f /tmp/spdk-ld-path 00:01:31.754 + source autorun-spdk.conf 00:01:31.754 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:01:31.754 ++ SPDK_TEST_BLOCKDEV=1 00:01:31.754 ++ SPDK_TEST_ISAL=1 00:01:31.754 ++ SPDK_TEST_CRYPTO=1 00:01:31.754 ++ SPDK_TEST_REDUCE=1 00:01:31.754 ++ SPDK_TEST_VBDEV_COMPRESS=1 00:01:31.754 ++ SPDK_RUN_UBSAN=1 00:01:31.754 ++ RUN_NIGHTLY=0 00:01:31.754 + (( SPDK_TEST_NVME_CMB == 1 || SPDK_TEST_NVME_PMR == 1 )) 00:01:31.754 + [[ -n '' ]] 00:01:31.754 + sudo git config --global --add safe.directory /var/jenkins/workspace/crypto-phy-autotest/spdk 00:01:31.754 + for M in /var/spdk/build-*-manifest.txt 00:01:31.754 + [[ -f /var/spdk/build-pkg-manifest.txt ]] 00:01:31.754 + cp /var/spdk/build-pkg-manifest.txt /var/jenkins/workspace/crypto-phy-autotest/output/ 00:01:31.754 + for M in /var/spdk/build-*-manifest.txt 00:01:31.754 + [[ -f /var/spdk/build-repo-manifest.txt ]] 00:01:31.754 + cp /var/spdk/build-repo-manifest.txt /var/jenkins/workspace/crypto-phy-autotest/output/ 00:01:31.754 ++ uname 00:01:31.754 + [[ Linux == \L\i\n\u\x ]] 00:01:31.754 + sudo dmesg -T 00:01:31.754 + sudo dmesg --clear 00:01:31.754 + dmesg_pid=706194 00:01:31.754 + [[ Fedora Linux == FreeBSD ]] 00:01:31.754 + export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:01:31.754 + UNBIND_ENTIRE_IOMMU_GROUP=yes 00:01:31.754 + [[ -e /var/spdk/dependencies/vhost/spdk_test_image.qcow2 ]] 00:01:31.754 + [[ -x /usr/src/fio-static/fio ]] 00:01:31.754 + sudo dmesg -Tw 00:01:31.754 + export FIO_BIN=/usr/src/fio-static/fio 00:01:31.754 + FIO_BIN=/usr/src/fio-static/fio 00:01:31.754 + [[ '' == \/\v\a\r\/\j\e\n\k\i\n\s\/\w\o\r\k\s\p\a\c\e\/\c\r\y\p\t\o\-\p\h\y\-\a\u\t\o\t\e\s\t\/\q\e\m\u\_\v\f\i\o\/* ]] 00:01:31.754 + [[ ! -v VFIO_QEMU_BIN ]] 00:01:31.754 + [[ -e /usr/local/qemu/vfio-user-latest ]] 00:01:31.754 + export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:01:31.754 + VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:01:31.754 + [[ -e /usr/local/qemu/vanilla-latest ]] 00:01:31.754 + export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:01:31.754 + QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:01:31.754 + spdk/autorun.sh /var/jenkins/workspace/crypto-phy-autotest/autorun-spdk.conf 00:01:31.754 Test configuration: 00:01:31.754 SPDK_RUN_FUNCTIONAL_TEST=1 00:01:31.754 SPDK_TEST_BLOCKDEV=1 00:01:31.754 SPDK_TEST_ISAL=1 00:01:31.754 SPDK_TEST_CRYPTO=1 00:01:31.754 SPDK_TEST_REDUCE=1 00:01:31.754 SPDK_TEST_VBDEV_COMPRESS=1 00:01:31.754 SPDK_RUN_UBSAN=1 00:01:31.754 RUN_NIGHTLY=0 11:12:15 -- common/autobuild_common.sh@15 -- $ source /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/common.sh 00:01:31.754 11:12:15 -- scripts/common.sh@508 -- $ [[ -e /bin/wpdk_common.sh ]] 00:01:31.754 11:12:15 -- scripts/common.sh@516 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:01:31.754 11:12:15 -- scripts/common.sh@517 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:01:31.754 11:12:15 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:01:31.754 11:12:15 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:01:31.754 11:12:15 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:01:31.754 11:12:15 -- paths/export.sh@5 -- $ export PATH 00:01:31.754 11:12:15 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:01:32.013 11:12:15 -- common/autobuild_common.sh@443 -- $ out=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:01:32.013 11:12:15 -- common/autobuild_common.sh@444 -- $ date +%s 00:01:32.013 11:12:15 -- common/autobuild_common.sh@444 -- $ mktemp -dt spdk_1721034735.XXXXXX 00:01:32.013 11:12:15 -- common/autobuild_common.sh@444 -- $ SPDK_WORKSPACE=/tmp/spdk_1721034735.sMmieS 00:01:32.013 11:12:15 -- common/autobuild_common.sh@446 -- $ [[ -n '' ]] 00:01:32.013 11:12:15 -- common/autobuild_common.sh@450 -- $ '[' -n '' ']' 00:01:32.013 11:12:15 -- common/autobuild_common.sh@453 -- $ scanbuild_exclude='--exclude /var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/' 00:01:32.013 11:12:15 -- common/autobuild_common.sh@457 -- $ scanbuild_exclude+=' --exclude /var/jenkins/workspace/crypto-phy-autotest/spdk/xnvme --exclude /tmp' 00:01:32.013 11:12:15 -- common/autobuild_common.sh@459 -- $ scanbuild='scan-build -o /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/scan-build-tmp --exclude /var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/ --exclude /var/jenkins/workspace/crypto-phy-autotest/spdk/xnvme --exclude /tmp --status-bugs' 00:01:32.013 11:12:15 -- common/autobuild_common.sh@460 -- $ get_config_params 00:01:32.013 11:12:15 -- common/autotest_common.sh@396 -- $ xtrace_disable 00:01:32.013 11:12:15 -- common/autotest_common.sh@10 -- $ set +x 00:01:32.013 11:12:15 -- common/autobuild_common.sh@460 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --with-vbdev-compress --with-dpdk-compressdev --with-crypto --enable-ubsan --enable-coverage --with-ublk' 00:01:32.013 11:12:15 -- common/autobuild_common.sh@462 -- $ start_monitor_resources 00:01:32.013 11:12:15 -- pm/common@17 -- $ local monitor 00:01:32.013 11:12:15 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:01:32.013 11:12:15 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:01:32.013 11:12:15 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:01:32.013 11:12:15 -- pm/common@21 -- $ date +%s 00:01:32.013 11:12:15 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:01:32.013 11:12:15 -- pm/common@21 -- $ date +%s 00:01:32.013 11:12:15 -- pm/common@25 -- $ sleep 1 00:01:32.013 11:12:15 -- pm/common@21 -- $ date +%s 00:01:32.013 11:12:15 -- pm/common@21 -- $ date +%s 00:01:32.013 11:12:15 -- pm/common@21 -- $ /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm/collect-cpu-temp -d /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1721034735 00:01:32.013 11:12:15 -- pm/common@21 -- $ /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm/collect-cpu-load -d /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1721034735 00:01:32.013 11:12:15 -- pm/common@21 -- $ /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm/collect-vmstat -d /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1721034735 00:01:32.013 11:12:15 -- pm/common@21 -- $ sudo -E /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm/collect-bmc-pm -d /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1721034735 00:01:32.014 Redirecting to /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1721034735_collect-vmstat.pm.log 00:01:32.014 Redirecting to /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1721034735_collect-cpu-temp.pm.log 00:01:32.014 Redirecting to /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1721034735_collect-cpu-load.pm.log 00:01:32.014 Redirecting to /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1721034735_collect-bmc-pm.bmc.pm.log 00:01:32.949 11:12:16 -- common/autobuild_common.sh@463 -- $ trap stop_monitor_resources EXIT 00:01:32.949 11:12:16 -- spdk/autobuild.sh@11 -- $ SPDK_TEST_AUTOBUILD= 00:01:32.949 11:12:16 -- spdk/autobuild.sh@12 -- $ umask 022 00:01:32.949 11:12:16 -- spdk/autobuild.sh@13 -- $ cd /var/jenkins/workspace/crypto-phy-autotest/spdk 00:01:32.949 11:12:16 -- spdk/autobuild.sh@16 -- $ date -u 00:01:32.949 Mon Jul 15 09:12:16 AM UTC 2024 00:01:32.949 11:12:16 -- spdk/autobuild.sh@17 -- $ git describe --tags 00:01:32.949 v24.09-pre-205-ge7cce062d 00:01:32.949 11:12:16 -- spdk/autobuild.sh@19 -- $ '[' 0 -eq 1 ']' 00:01:32.949 11:12:16 -- spdk/autobuild.sh@23 -- $ '[' 1 -eq 1 ']' 00:01:32.949 11:12:16 -- spdk/autobuild.sh@24 -- $ run_test ubsan echo 'using ubsan' 00:01:32.949 11:12:16 -- common/autotest_common.sh@1099 -- $ '[' 3 -le 1 ']' 00:01:32.949 11:12:16 -- common/autotest_common.sh@1105 -- $ xtrace_disable 00:01:32.949 11:12:16 -- common/autotest_common.sh@10 -- $ set +x 00:01:32.949 ************************************ 00:01:32.949 START TEST ubsan 00:01:32.949 ************************************ 00:01:32.949 11:12:16 ubsan -- common/autotest_common.sh@1123 -- $ echo 'using ubsan' 00:01:32.949 using ubsan 00:01:32.949 00:01:32.949 real 0m0.001s 00:01:32.949 user 0m0.000s 00:01:32.949 sys 0m0.000s 00:01:32.949 11:12:16 ubsan -- common/autotest_common.sh@1124 -- $ xtrace_disable 00:01:32.949 11:12:16 ubsan -- common/autotest_common.sh@10 -- $ set +x 00:01:32.949 ************************************ 00:01:32.949 END TEST ubsan 00:01:32.949 ************************************ 00:01:32.949 11:12:16 -- common/autotest_common.sh@1142 -- $ return 0 00:01:32.949 11:12:16 -- spdk/autobuild.sh@27 -- $ '[' -n '' ']' 00:01:32.949 11:12:16 -- spdk/autobuild.sh@31 -- $ case "$SPDK_TEST_AUTOBUILD" in 00:01:32.949 11:12:16 -- spdk/autobuild.sh@47 -- $ [[ 0 -eq 1 ]] 00:01:32.949 11:12:16 -- spdk/autobuild.sh@51 -- $ [[ 0 -eq 1 ]] 00:01:32.949 11:12:16 -- spdk/autobuild.sh@55 -- $ [[ -n '' ]] 00:01:32.949 11:12:16 -- spdk/autobuild.sh@57 -- $ [[ 0 -eq 1 ]] 00:01:32.949 11:12:16 -- spdk/autobuild.sh@59 -- $ [[ 0 -eq 1 ]] 00:01:32.949 11:12:16 -- spdk/autobuild.sh@62 -- $ [[ 0 -eq 1 ]] 00:01:32.949 11:12:16 -- spdk/autobuild.sh@67 -- $ /var/jenkins/workspace/crypto-phy-autotest/spdk/configure --enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --with-vbdev-compress --with-dpdk-compressdev --with-crypto --enable-ubsan --enable-coverage --with-ublk --with-shared 00:01:33.207 Using default SPDK env in /var/jenkins/workspace/crypto-phy-autotest/spdk/lib/env_dpdk 00:01:33.207 Using default DPDK in /var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build 00:01:33.465 Using 'verbs' RDMA provider 00:01:49.713 Configuring ISA-L (logfile: /var/jenkins/workspace/crypto-phy-autotest/spdk/.spdk-isal.log)...done. 00:02:04.591 Configuring ISA-L-crypto (logfile: /var/jenkins/workspace/crypto-phy-autotest/spdk/.spdk-isal-crypto.log)...done. 00:02:04.591 Creating mk/config.mk...done. 00:02:04.591 Creating mk/cc.flags.mk...done. 00:02:04.591 Type 'make' to build. 00:02:04.591 11:12:46 -- spdk/autobuild.sh@69 -- $ run_test make make -j72 00:02:04.591 11:12:46 -- common/autotest_common.sh@1099 -- $ '[' 3 -le 1 ']' 00:02:04.591 11:12:46 -- common/autotest_common.sh@1105 -- $ xtrace_disable 00:02:04.591 11:12:46 -- common/autotest_common.sh@10 -- $ set +x 00:02:04.591 ************************************ 00:02:04.591 START TEST make 00:02:04.591 ************************************ 00:02:04.591 11:12:46 make -- common/autotest_common.sh@1123 -- $ make -j72 00:02:04.591 make[1]: Nothing to be done for 'all'. 00:02:43.324 The Meson build system 00:02:43.324 Version: 1.3.1 00:02:43.324 Source dir: /var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk 00:02:43.324 Build dir: /var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build-tmp 00:02:43.324 Build type: native build 00:02:43.324 Program cat found: YES (/usr/bin/cat) 00:02:43.324 Project name: DPDK 00:02:43.324 Project version: 24.03.0 00:02:43.324 C compiler for the host machine: cc (gcc 13.2.1 "cc (GCC) 13.2.1 20231011 (Red Hat 13.2.1-4)") 00:02:43.324 C linker for the host machine: cc ld.bfd 2.39-16 00:02:43.324 Host machine cpu family: x86_64 00:02:43.324 Host machine cpu: x86_64 00:02:43.324 Message: ## Building in Developer Mode ## 00:02:43.324 Program pkg-config found: YES (/usr/bin/pkg-config) 00:02:43.324 Program check-symbols.sh found: YES (/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/buildtools/check-symbols.sh) 00:02:43.324 Program options-ibverbs-static.sh found: YES (/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/buildtools/options-ibverbs-static.sh) 00:02:43.324 Program python3 found: YES (/usr/bin/python3) 00:02:43.324 Program cat found: YES (/usr/bin/cat) 00:02:43.324 Compiler for C supports arguments -march=native: YES 00:02:43.324 Checking for size of "void *" : 8 00:02:43.324 Checking for size of "void *" : 8 (cached) 00:02:43.324 Compiler for C supports link arguments -Wl,--undefined-version: NO 00:02:43.324 Library m found: YES 00:02:43.324 Library numa found: YES 00:02:43.324 Has header "numaif.h" : YES 00:02:43.324 Library fdt found: NO 00:02:43.324 Library execinfo found: NO 00:02:43.324 Has header "execinfo.h" : YES 00:02:43.324 Found pkg-config: YES (/usr/bin/pkg-config) 1.8.0 00:02:43.324 Run-time dependency libarchive found: NO (tried pkgconfig) 00:02:43.324 Run-time dependency libbsd found: NO (tried pkgconfig) 00:02:43.324 Run-time dependency jansson found: NO (tried pkgconfig) 00:02:43.324 Run-time dependency openssl found: YES 3.0.9 00:02:43.324 Run-time dependency libpcap found: YES 1.10.4 00:02:43.324 Has header "pcap.h" with dependency libpcap: YES 00:02:43.324 Compiler for C supports arguments -Wcast-qual: YES 00:02:43.324 Compiler for C supports arguments -Wdeprecated: YES 00:02:43.324 Compiler for C supports arguments -Wformat: YES 00:02:43.324 Compiler for C supports arguments -Wformat-nonliteral: NO 00:02:43.324 Compiler for C supports arguments -Wformat-security: NO 00:02:43.324 Compiler for C supports arguments -Wmissing-declarations: YES 00:02:43.324 Compiler for C supports arguments -Wmissing-prototypes: YES 00:02:43.324 Compiler for C supports arguments -Wnested-externs: YES 00:02:43.324 Compiler for C supports arguments -Wold-style-definition: YES 00:02:43.324 Compiler for C supports arguments -Wpointer-arith: YES 00:02:43.324 Compiler for C supports arguments -Wsign-compare: YES 00:02:43.324 Compiler for C supports arguments -Wstrict-prototypes: YES 00:02:43.324 Compiler for C supports arguments -Wundef: YES 00:02:43.324 Compiler for C supports arguments -Wwrite-strings: YES 00:02:43.324 Compiler for C supports arguments -Wno-address-of-packed-member: YES 00:02:43.324 Compiler for C supports arguments -Wno-packed-not-aligned: YES 00:02:43.324 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:02:43.324 Compiler for C supports arguments -Wno-zero-length-bounds: YES 00:02:43.324 Program objdump found: YES (/usr/bin/objdump) 00:02:43.324 Compiler for C supports arguments -mavx512f: YES 00:02:43.324 Checking if "AVX512 checking" compiles: YES 00:02:43.324 Fetching value of define "__SSE4_2__" : 1 00:02:43.324 Fetching value of define "__AES__" : 1 00:02:43.324 Fetching value of define "__AVX__" : 1 00:02:43.324 Fetching value of define "__AVX2__" : 1 00:02:43.324 Fetching value of define "__AVX512BW__" : 1 00:02:43.324 Fetching value of define "__AVX512CD__" : 1 00:02:43.324 Fetching value of define "__AVX512DQ__" : 1 00:02:43.324 Fetching value of define "__AVX512F__" : 1 00:02:43.324 Fetching value of define "__AVX512VL__" : 1 00:02:43.324 Fetching value of define "__PCLMUL__" : 1 00:02:43.324 Fetching value of define "__RDRND__" : 1 00:02:43.324 Fetching value of define "__RDSEED__" : 1 00:02:43.324 Fetching value of define "__VPCLMULQDQ__" : (undefined) 00:02:43.324 Fetching value of define "__znver1__" : (undefined) 00:02:43.324 Fetching value of define "__znver2__" : (undefined) 00:02:43.324 Fetching value of define "__znver3__" : (undefined) 00:02:43.324 Fetching value of define "__znver4__" : (undefined) 00:02:43.324 Compiler for C supports arguments -Wno-format-truncation: YES 00:02:43.324 Message: lib/log: Defining dependency "log" 00:02:43.324 Message: lib/kvargs: Defining dependency "kvargs" 00:02:43.324 Message: lib/telemetry: Defining dependency "telemetry" 00:02:43.324 Checking for function "getentropy" : NO 00:02:43.324 Message: lib/eal: Defining dependency "eal" 00:02:43.324 Message: lib/ring: Defining dependency "ring" 00:02:43.324 Message: lib/rcu: Defining dependency "rcu" 00:02:43.324 Message: lib/mempool: Defining dependency "mempool" 00:02:43.324 Message: lib/mbuf: Defining dependency "mbuf" 00:02:43.324 Fetching value of define "__PCLMUL__" : 1 (cached) 00:02:43.324 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:43.324 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:43.324 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:43.324 Fetching value of define "__AVX512VL__" : 1 (cached) 00:02:43.324 Fetching value of define "__VPCLMULQDQ__" : (undefined) (cached) 00:02:43.324 Compiler for C supports arguments -mpclmul: YES 00:02:43.324 Compiler for C supports arguments -maes: YES 00:02:43.324 Compiler for C supports arguments -mavx512f: YES (cached) 00:02:43.324 Compiler for C supports arguments -mavx512bw: YES 00:02:43.324 Compiler for C supports arguments -mavx512dq: YES 00:02:43.324 Compiler for C supports arguments -mavx512vl: YES 00:02:43.324 Compiler for C supports arguments -mvpclmulqdq: YES 00:02:43.324 Compiler for C supports arguments -mavx2: YES 00:02:43.324 Compiler for C supports arguments -mavx: YES 00:02:43.324 Message: lib/net: Defining dependency "net" 00:02:43.324 Message: lib/meter: Defining dependency "meter" 00:02:43.324 Message: lib/ethdev: Defining dependency "ethdev" 00:02:43.324 Message: lib/pci: Defining dependency "pci" 00:02:43.324 Message: lib/cmdline: Defining dependency "cmdline" 00:02:43.324 Message: lib/hash: Defining dependency "hash" 00:02:43.324 Message: lib/timer: Defining dependency "timer" 00:02:43.324 Message: lib/compressdev: Defining dependency "compressdev" 00:02:43.324 Message: lib/cryptodev: Defining dependency "cryptodev" 00:02:43.324 Message: lib/dmadev: Defining dependency "dmadev" 00:02:43.324 Compiler for C supports arguments -Wno-cast-qual: YES 00:02:43.324 Message: lib/power: Defining dependency "power" 00:02:43.324 Message: lib/reorder: Defining dependency "reorder" 00:02:43.324 Message: lib/security: Defining dependency "security" 00:02:43.324 Has header "linux/userfaultfd.h" : YES 00:02:43.324 Has header "linux/vduse.h" : YES 00:02:43.324 Message: lib/vhost: Defining dependency "vhost" 00:02:43.324 Compiler for C supports arguments -Wno-format-truncation: YES (cached) 00:02:43.324 Message: drivers/bus/auxiliary: Defining dependency "bus_auxiliary" 00:02:43.324 Message: drivers/bus/pci: Defining dependency "bus_pci" 00:02:43.324 Message: drivers/bus/vdev: Defining dependency "bus_vdev" 00:02:43.324 Compiler for C supports arguments -std=c11: YES 00:02:43.324 Compiler for C supports arguments -Wno-strict-prototypes: YES 00:02:43.324 Compiler for C supports arguments -D_BSD_SOURCE: YES 00:02:43.324 Compiler for C supports arguments -D_DEFAULT_SOURCE: YES 00:02:43.324 Compiler for C supports arguments -D_XOPEN_SOURCE=600: YES 00:02:43.324 Run-time dependency libmlx5 found: YES 1.24.44.0 00:02:43.324 Run-time dependency libibverbs found: YES 1.14.44.0 00:02:43.324 Library mtcr_ul found: NO 00:02:43.324 Header "infiniband/verbs.h" has symbol "IBV_FLOW_SPEC_ESP" with dependencies libmlx5, libibverbs: YES 00:02:43.324 Header "infiniband/verbs.h" has symbol "IBV_RX_HASH_IPSEC_SPI" with dependencies libmlx5, libibverbs: YES 00:02:43.324 Header "infiniband/verbs.h" has symbol "IBV_ACCESS_RELAXED_ORDERING " with dependencies libmlx5, libibverbs: YES 00:02:43.324 Header "infiniband/mlx5dv.h" has symbol "MLX5DV_CQE_RES_FORMAT_CSUM_STRIDX" with dependencies libmlx5, libibverbs: YES 00:02:43.324 Header "infiniband/mlx5dv.h" has symbol "MLX5DV_CONTEXT_MASK_TUNNEL_OFFLOADS" with dependencies libmlx5, libibverbs: YES 00:02:43.324 Header "infiniband/mlx5dv.h" has symbol "MLX5DV_CONTEXT_FLAGS_MPW_ALLOWED" with dependencies libmlx5, libibverbs: YES 00:02:43.325 Header "infiniband/mlx5dv.h" has symbol "MLX5DV_CONTEXT_FLAGS_CQE_128B_COMP" with dependencies libmlx5, libibverbs: YES 00:02:43.325 Header "infiniband/mlx5dv.h" has symbol "MLX5DV_CQ_INIT_ATTR_FLAGS_CQE_PAD" with dependencies libmlx5, libibverbs: YES 00:02:43.325 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_create_flow_action_packet_reformat" with dependencies libmlx5, libibverbs: YES 00:02:43.325 Header "infiniband/verbs.h" has symbol "IBV_FLOW_SPEC_MPLS" with dependencies libmlx5, libibverbs: YES 00:02:43.325 Header "infiniband/verbs.h" has symbol "IBV_WQ_FLAGS_PCI_WRITE_END_PADDING" with dependencies libmlx5, libibverbs: YES 00:02:43.325 Header "infiniband/verbs.h" has symbol "IBV_WQ_FLAG_RX_END_PADDING" with dependencies libmlx5, libibverbs: NO 00:02:43.325 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_query_devx_port" with dependencies libmlx5, libibverbs: NO 00:02:43.325 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_query_port" with dependencies libmlx5, libibverbs: YES 00:02:43.325 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_dr_action_create_dest_ib_port" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_devx_obj_create" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "MLX5DV_FLOW_ACTION_COUNTERS_DEVX" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "MLX5DV_FLOW_ACTION_DEFAULT_MISS" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_devx_obj_query_async" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_devx_qp_query" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_pp_alloc" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_dr_action_create_dest_devx_tir" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_devx_get_event" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_dr_action_create_flow_meter" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "MLX5_MMAP_GET_NC_PAGES_CMD" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "MLX5DV_DR_DOMAIN_TYPE_NIC_RX" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "MLX5DV_DR_DOMAIN_TYPE_FDB" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_dr_action_create_push_vlan" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_alloc_var" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "MLX5_OPCODE_ENHANCED_MPSW" with dependencies libmlx5, libibverbs: NO 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "MLX5_OPCODE_SEND_EN" with dependencies libmlx5, libibverbs: NO 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "MLX5_OPCODE_WAIT" with dependencies libmlx5, libibverbs: NO 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "MLX5_OPCODE_ACCESS_ASO" with dependencies libmlx5, libibverbs: NO 00:02:44.703 Header "linux/if_link.h" has symbol "IFLA_NUM_VF" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "linux/if_link.h" has symbol "IFLA_EXT_MASK" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "linux/if_link.h" has symbol "IFLA_PHYS_SWITCH_ID" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "linux/if_link.h" has symbol "IFLA_PHYS_PORT_NAME" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "rdma/rdma_netlink.h" has symbol "RDMA_NL_NLDEV" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "rdma/rdma_netlink.h" has symbol "RDMA_NLDEV_CMD_GET" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "rdma/rdma_netlink.h" has symbol "RDMA_NLDEV_CMD_PORT_GET" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "rdma/rdma_netlink.h" has symbol "RDMA_NLDEV_ATTR_DEV_INDEX" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "rdma/rdma_netlink.h" has symbol "RDMA_NLDEV_ATTR_DEV_NAME" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "rdma/rdma_netlink.h" has symbol "RDMA_NLDEV_ATTR_PORT_INDEX" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "rdma/rdma_netlink.h" has symbol "RDMA_NLDEV_ATTR_PORT_STATE" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "rdma/rdma_netlink.h" has symbol "RDMA_NLDEV_ATTR_NDEV_INDEX" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_dump_dr_domain" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_dr_action_create_flow_sampler" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_dr_domain_set_reclaim_device_memory" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_dr_action_create_dest_array" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "linux/devlink.h" has symbol "DEVLINK_GENL_NAME" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_dr_action_create_aso" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/verbs.h" has symbol "INFINIBAND_VERBS_H" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "MLX5_WQE_UMR_CTRL_FLAG_INLINE" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_dump_dr_rule" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "MLX5DV_DR_ACTION_FLAGS_ASO_CT_DIRECTION_INITIATOR" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_dr_domain_allow_duplicate_rules" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/verbs.h" has symbol "ibv_reg_mr_iova" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/verbs.h" has symbol "ibv_import_device" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_dr_action_create_dest_root_table" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/mlx5dv.h" has symbol "mlx5dv_create_steering_anchor" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Header "infiniband/verbs.h" has symbol "ibv_is_fork_initialized" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Checking whether type "struct mlx5dv_sw_parsing_caps" has member "sw_parsing_offloads" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Checking whether type "struct ibv_counter_set_init_attr" has member "counter_set_id" with dependencies libmlx5, libibverbs: NO 00:02:44.703 Checking whether type "struct ibv_counters_init_attr" has member "comp_mask" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Checking whether type "struct mlx5dv_devx_uar" has member "mmap_off" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Checking whether type "struct mlx5dv_flow_matcher_attr" has member "ft_type" with dependencies libmlx5, libibverbs: YES 00:02:44.703 Configuring mlx5_autoconf.h using configuration 00:02:44.703 Message: drivers/common/mlx5: Defining dependency "common_mlx5" 00:02:44.703 Run-time dependency libcrypto found: YES 3.0.9 00:02:44.703 Library IPSec_MB found: YES 00:02:44.703 Fetching value of define "IMB_VERSION_STR" : "1.5.0" 00:02:44.703 Message: drivers/common/qat: Defining dependency "common_qat" 00:02:44.703 Message: drivers/mempool/ring: Defining dependency "mempool_ring" 00:02:44.703 Message: Disabling raw/* drivers: missing internal dependency "rawdev" 00:02:44.703 Library IPSec_MB found: YES 00:02:44.703 Fetching value of define "IMB_VERSION_STR" : "1.5.0" (cached) 00:02:44.703 Message: drivers/crypto/ipsec_mb: Defining dependency "crypto_ipsec_mb" 00:02:44.703 Compiler for C supports arguments -std=c11: YES (cached) 00:02:44.703 Compiler for C supports arguments -Wno-strict-prototypes: YES (cached) 00:02:44.703 Compiler for C supports arguments -D_BSD_SOURCE: YES (cached) 00:02:44.703 Compiler for C supports arguments -D_DEFAULT_SOURCE: YES (cached) 00:02:44.703 Compiler for C supports arguments -D_XOPEN_SOURCE=600: YES (cached) 00:02:44.703 Message: drivers/crypto/mlx5: Defining dependency "crypto_mlx5" 00:02:44.703 Run-time dependency libisal found: NO (tried pkgconfig) 00:02:44.703 Library libisal found: NO 00:02:44.703 Message: drivers/compress/isal: Defining dependency "compress_isal" 00:02:44.703 Compiler for C supports arguments -std=c11: YES (cached) 00:02:44.703 Compiler for C supports arguments -Wno-strict-prototypes: YES (cached) 00:02:44.703 Compiler for C supports arguments -D_BSD_SOURCE: YES (cached) 00:02:44.703 Compiler for C supports arguments -D_DEFAULT_SOURCE: YES (cached) 00:02:44.703 Compiler for C supports arguments -D_XOPEN_SOURCE=600: YES (cached) 00:02:44.703 Message: drivers/compress/mlx5: Defining dependency "compress_mlx5" 00:02:44.703 Message: Disabling regex/* drivers: missing internal dependency "regexdev" 00:02:44.703 Message: Disabling ml/* drivers: missing internal dependency "mldev" 00:02:44.703 Message: Disabling event/* drivers: missing internal dependency "eventdev" 00:02:44.703 Message: Disabling baseband/* drivers: missing internal dependency "bbdev" 00:02:44.703 Message: Disabling gpu/* drivers: missing internal dependency "gpudev" 00:02:44.703 Program doxygen found: YES (/usr/bin/doxygen) 00:02:44.703 Configuring doxy-api-html.conf using configuration 00:02:44.703 Configuring doxy-api-man.conf using configuration 00:02:44.703 Program mandb found: YES (/usr/bin/mandb) 00:02:44.703 Program sphinx-build found: NO 00:02:44.703 Configuring rte_build_config.h using configuration 00:02:44.703 Message: 00:02:44.703 ================= 00:02:44.703 Applications Enabled 00:02:44.703 ================= 00:02:44.703 00:02:44.703 apps: 00:02:44.703 00:02:44.703 00:02:44.703 Message: 00:02:44.703 ================= 00:02:44.703 Libraries Enabled 00:02:44.703 ================= 00:02:44.703 00:02:44.703 libs: 00:02:44.703 log, kvargs, telemetry, eal, ring, rcu, mempool, mbuf, 00:02:44.703 net, meter, ethdev, pci, cmdline, hash, timer, compressdev, 00:02:44.703 cryptodev, dmadev, power, reorder, security, vhost, 00:02:44.703 00:02:44.703 Message: 00:02:44.703 =============== 00:02:44.703 Drivers Enabled 00:02:44.703 =============== 00:02:44.703 00:02:44.703 common: 00:02:44.703 mlx5, qat, 00:02:44.703 bus: 00:02:44.703 auxiliary, pci, vdev, 00:02:44.703 mempool: 00:02:44.703 ring, 00:02:44.703 dma: 00:02:44.703 00:02:44.703 net: 00:02:44.703 00:02:44.703 crypto: 00:02:44.703 ipsec_mb, mlx5, 00:02:44.703 compress: 00:02:44.703 isal, mlx5, 00:02:44.703 vdpa: 00:02:44.703 00:02:44.703 00:02:44.703 Message: 00:02:44.703 ================= 00:02:44.703 Content Skipped 00:02:44.703 ================= 00:02:44.703 00:02:44.703 apps: 00:02:44.703 dumpcap: explicitly disabled via build config 00:02:44.703 graph: explicitly disabled via build config 00:02:44.703 pdump: explicitly disabled via build config 00:02:44.703 proc-info: explicitly disabled via build config 00:02:44.703 test-acl: explicitly disabled via build config 00:02:44.703 test-bbdev: explicitly disabled via build config 00:02:44.703 test-cmdline: explicitly disabled via build config 00:02:44.703 test-compress-perf: explicitly disabled via build config 00:02:44.703 test-crypto-perf: explicitly disabled via build config 00:02:44.703 test-dma-perf: explicitly disabled via build config 00:02:44.703 test-eventdev: explicitly disabled via build config 00:02:44.703 test-fib: explicitly disabled via build config 00:02:44.703 test-flow-perf: explicitly disabled via build config 00:02:44.703 test-gpudev: explicitly disabled via build config 00:02:44.703 test-mldev: explicitly disabled via build config 00:02:44.703 test-pipeline: explicitly disabled via build config 00:02:44.703 test-pmd: explicitly disabled via build config 00:02:44.703 test-regex: explicitly disabled via build config 00:02:44.703 test-sad: explicitly disabled via build config 00:02:44.703 test-security-perf: explicitly disabled via build config 00:02:44.703 00:02:44.703 libs: 00:02:44.703 argparse: explicitly disabled via build config 00:02:44.703 metrics: explicitly disabled via build config 00:02:44.703 acl: explicitly disabled via build config 00:02:44.703 bbdev: explicitly disabled via build config 00:02:44.703 bitratestats: explicitly disabled via build config 00:02:44.703 bpf: explicitly disabled via build config 00:02:44.703 cfgfile: explicitly disabled via build config 00:02:44.703 distributor: explicitly disabled via build config 00:02:44.703 efd: explicitly disabled via build config 00:02:44.704 eventdev: explicitly disabled via build config 00:02:44.704 dispatcher: explicitly disabled via build config 00:02:44.704 gpudev: explicitly disabled via build config 00:02:44.704 gro: explicitly disabled via build config 00:02:44.704 gso: explicitly disabled via build config 00:02:44.704 ip_frag: explicitly disabled via build config 00:02:44.704 jobstats: explicitly disabled via build config 00:02:44.704 latencystats: explicitly disabled via build config 00:02:44.704 lpm: explicitly disabled via build config 00:02:44.704 member: explicitly disabled via build config 00:02:44.704 pcapng: explicitly disabled via build config 00:02:44.704 rawdev: explicitly disabled via build config 00:02:44.704 regexdev: explicitly disabled via build config 00:02:44.704 mldev: explicitly disabled via build config 00:02:44.704 rib: explicitly disabled via build config 00:02:44.704 sched: explicitly disabled via build config 00:02:44.704 stack: explicitly disabled via build config 00:02:44.704 ipsec: explicitly disabled via build config 00:02:44.704 pdcp: explicitly disabled via build config 00:02:44.704 fib: explicitly disabled via build config 00:02:44.704 port: explicitly disabled via build config 00:02:44.704 pdump: explicitly disabled via build config 00:02:44.704 table: explicitly disabled via build config 00:02:44.704 pipeline: explicitly disabled via build config 00:02:44.704 graph: explicitly disabled via build config 00:02:44.704 node: explicitly disabled via build config 00:02:44.704 00:02:44.704 drivers: 00:02:44.704 common/cpt: not in enabled drivers build config 00:02:44.704 common/dpaax: not in enabled drivers build config 00:02:44.704 common/iavf: not in enabled drivers build config 00:02:44.704 common/idpf: not in enabled drivers build config 00:02:44.704 common/ionic: not in enabled drivers build config 00:02:44.704 common/mvep: not in enabled drivers build config 00:02:44.704 common/octeontx: not in enabled drivers build config 00:02:44.704 bus/cdx: not in enabled drivers build config 00:02:44.704 bus/dpaa: not in enabled drivers build config 00:02:44.704 bus/fslmc: not in enabled drivers build config 00:02:44.704 bus/ifpga: not in enabled drivers build config 00:02:44.704 bus/platform: not in enabled drivers build config 00:02:44.704 bus/uacce: not in enabled drivers build config 00:02:44.704 bus/vmbus: not in enabled drivers build config 00:02:44.704 common/cnxk: not in enabled drivers build config 00:02:44.704 common/nfp: not in enabled drivers build config 00:02:44.704 common/nitrox: not in enabled drivers build config 00:02:44.704 common/sfc_efx: not in enabled drivers build config 00:02:44.704 mempool/bucket: not in enabled drivers build config 00:02:44.704 mempool/cnxk: not in enabled drivers build config 00:02:44.704 mempool/dpaa: not in enabled drivers build config 00:02:44.704 mempool/dpaa2: not in enabled drivers build config 00:02:44.704 mempool/octeontx: not in enabled drivers build config 00:02:44.704 mempool/stack: not in enabled drivers build config 00:02:44.704 dma/cnxk: not in enabled drivers build config 00:02:44.704 dma/dpaa: not in enabled drivers build config 00:02:44.704 dma/dpaa2: not in enabled drivers build config 00:02:44.704 dma/hisilicon: not in enabled drivers build config 00:02:44.704 dma/idxd: not in enabled drivers build config 00:02:44.704 dma/ioat: not in enabled drivers build config 00:02:44.704 dma/skeleton: not in enabled drivers build config 00:02:44.704 net/af_packet: not in enabled drivers build config 00:02:44.704 net/af_xdp: not in enabled drivers build config 00:02:44.704 net/ark: not in enabled drivers build config 00:02:44.704 net/atlantic: not in enabled drivers build config 00:02:44.704 net/avp: not in enabled drivers build config 00:02:44.704 net/axgbe: not in enabled drivers build config 00:02:44.704 net/bnx2x: not in enabled drivers build config 00:02:44.704 net/bnxt: not in enabled drivers build config 00:02:44.704 net/bonding: not in enabled drivers build config 00:02:44.704 net/cnxk: not in enabled drivers build config 00:02:44.704 net/cpfl: not in enabled drivers build config 00:02:44.704 net/cxgbe: not in enabled drivers build config 00:02:44.704 net/dpaa: not in enabled drivers build config 00:02:44.704 net/dpaa2: not in enabled drivers build config 00:02:44.704 net/e1000: not in enabled drivers build config 00:02:44.704 net/ena: not in enabled drivers build config 00:02:44.704 net/enetc: not in enabled drivers build config 00:02:44.704 net/enetfec: not in enabled drivers build config 00:02:44.704 net/enic: not in enabled drivers build config 00:02:44.704 net/failsafe: not in enabled drivers build config 00:02:44.704 net/fm10k: not in enabled drivers build config 00:02:44.704 net/gve: not in enabled drivers build config 00:02:44.704 net/hinic: not in enabled drivers build config 00:02:44.704 net/hns3: not in enabled drivers build config 00:02:44.704 net/i40e: not in enabled drivers build config 00:02:44.704 net/iavf: not in enabled drivers build config 00:02:44.704 net/ice: not in enabled drivers build config 00:02:44.704 net/idpf: not in enabled drivers build config 00:02:44.704 net/igc: not in enabled drivers build config 00:02:44.704 net/ionic: not in enabled drivers build config 00:02:44.704 net/ipn3ke: not in enabled drivers build config 00:02:44.704 net/ixgbe: not in enabled drivers build config 00:02:44.704 net/mana: not in enabled drivers build config 00:02:44.704 net/memif: not in enabled drivers build config 00:02:44.704 net/mlx4: not in enabled drivers build config 00:02:44.704 net/mlx5: not in enabled drivers build config 00:02:44.704 net/mvneta: not in enabled drivers build config 00:02:44.704 net/mvpp2: not in enabled drivers build config 00:02:44.704 net/netvsc: not in enabled drivers build config 00:02:44.704 net/nfb: not in enabled drivers build config 00:02:44.704 net/nfp: not in enabled drivers build config 00:02:44.704 net/ngbe: not in enabled drivers build config 00:02:44.704 net/null: not in enabled drivers build config 00:02:44.704 net/octeontx: not in enabled drivers build config 00:02:44.704 net/octeon_ep: not in enabled drivers build config 00:02:44.704 net/pcap: not in enabled drivers build config 00:02:44.704 net/pfe: not in enabled drivers build config 00:02:44.704 net/qede: not in enabled drivers build config 00:02:44.704 net/ring: not in enabled drivers build config 00:02:44.704 net/sfc: not in enabled drivers build config 00:02:44.704 net/softnic: not in enabled drivers build config 00:02:44.704 net/tap: not in enabled drivers build config 00:02:44.704 net/thunderx: not in enabled drivers build config 00:02:44.704 net/txgbe: not in enabled drivers build config 00:02:44.704 net/vdev_netvsc: not in enabled drivers build config 00:02:44.704 net/vhost: not in enabled drivers build config 00:02:44.704 net/virtio: not in enabled drivers build config 00:02:44.704 net/vmxnet3: not in enabled drivers build config 00:02:44.704 raw/*: missing internal dependency, "rawdev" 00:02:44.704 crypto/armv8: not in enabled drivers build config 00:02:44.704 crypto/bcmfs: not in enabled drivers build config 00:02:44.704 crypto/caam_jr: not in enabled drivers build config 00:02:44.704 crypto/ccp: not in enabled drivers build config 00:02:44.704 crypto/cnxk: not in enabled drivers build config 00:02:44.704 crypto/dpaa_sec: not in enabled drivers build config 00:02:44.704 crypto/dpaa2_sec: not in enabled drivers build config 00:02:44.704 crypto/mvsam: not in enabled drivers build config 00:02:44.704 crypto/nitrox: not in enabled drivers build config 00:02:44.704 crypto/null: not in enabled drivers build config 00:02:44.704 crypto/octeontx: not in enabled drivers build config 00:02:44.704 crypto/openssl: not in enabled drivers build config 00:02:44.704 crypto/scheduler: not in enabled drivers build config 00:02:44.704 crypto/uadk: not in enabled drivers build config 00:02:44.704 crypto/virtio: not in enabled drivers build config 00:02:44.704 compress/nitrox: not in enabled drivers build config 00:02:44.704 compress/octeontx: not in enabled drivers build config 00:02:44.704 compress/zlib: not in enabled drivers build config 00:02:44.704 regex/*: missing internal dependency, "regexdev" 00:02:44.704 ml/*: missing internal dependency, "mldev" 00:02:44.704 vdpa/ifc: not in enabled drivers build config 00:02:44.704 vdpa/mlx5: not in enabled drivers build config 00:02:44.704 vdpa/nfp: not in enabled drivers build config 00:02:44.704 vdpa/sfc: not in enabled drivers build config 00:02:44.704 event/*: missing internal dependency, "eventdev" 00:02:44.704 baseband/*: missing internal dependency, "bbdev" 00:02:44.704 gpu/*: missing internal dependency, "gpudev" 00:02:44.704 00:02:44.704 00:02:45.273 Build targets in project: 115 00:02:45.273 00:02:45.273 DPDK 24.03.0 00:02:45.273 00:02:45.273 User defined options 00:02:45.273 buildtype : debug 00:02:45.273 default_library : shared 00:02:45.273 libdir : lib 00:02:45.273 prefix : /var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build 00:02:45.273 c_args : -Wno-stringop-overflow -fcommon -Wno-stringop-overread -Wno-array-bounds -I/var/jenkins/workspace/crypto-phy-autotest/spdk/intel-ipsec-mb/lib -DNO_COMPAT_IMB_API_053 -I/var/jenkins/workspace/crypto-phy-autotest/spdk/isa-l -I/var/jenkins/workspace/crypto-phy-autotest/spdk/isalbuild -fPIC -Werror 00:02:45.273 c_link_args : -L/var/jenkins/workspace/crypto-phy-autotest/spdk/intel-ipsec-mb/lib -L/var/jenkins/workspace/crypto-phy-autotest/spdk/isa-l/.libs -lisal 00:02:45.273 cpu_instruction_set: native 00:02:45.273 disable_apps : dumpcap,graph,pdump,proc-info,test-acl,test-bbdev,test-cmdline,test-compress-perf,test-crypto-perf,test-dma-perf,test-eventdev,test-fib,test-flow-perf,test-gpudev,test-mldev,test-pipeline,test-pmd,test-regex,test-sad,test-security-perf,test 00:02:45.273 disable_libs : acl,argparse,bbdev,bitratestats,bpf,cfgfile,dispatcher,distributor,efd,eventdev,fib,gpudev,graph,gro,gso,ip_frag,ipsec,jobstats,latencystats,lpm,member,metrics,mldev,node,pcapng,pdcp,pdump,pipeline,port,rawdev,regexdev,rib,sched,stack,table 00:02:45.273 enable_docs : false 00:02:45.273 enable_drivers : bus,bus/pci,bus/vdev,mempool/ring,crypto/qat,compress/qat,common/qat,common/mlx5,bus/auxiliary,crypto,crypto/aesni_mb,crypto/mlx5,crypto/ipsec_mb,compress,compress/isal,compress/mlx5 00:02:45.273 enable_kmods : false 00:02:45.273 max_lcores : 128 00:02:45.273 tests : false 00:02:45.273 00:02:45.273 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:02:45.532 ninja: Entering directory `/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build-tmp' 00:02:45.800 [1/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hypervisor.c.o 00:02:45.800 [2/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_errno.c.o 00:02:45.800 [3/378] Compiling C object lib/librte_kvargs.a.p/kvargs_rte_kvargs.c.o 00:02:45.800 [4/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hexdump.c.o 00:02:45.800 [5/378] Compiling C object lib/librte_log.a.p/log_log_linux.c.o 00:02:45.800 [6/378] Compiling C object lib/librte_eal.a.p/eal_common_rte_reciprocal.c.o 00:02:45.800 [7/378] Linking static target lib/librte_kvargs.a 00:02:45.800 [8/378] Compiling C object lib/librte_eal.a.p/eal_common_rte_version.c.o 00:02:45.800 [9/378] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_data.c.o 00:02:45.800 [10/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_uuid.c.o 00:02:45.800 [11/378] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cpuflags.c.o 00:02:45.800 [12/378] Compiling C object lib/librte_eal.a.p/eal_linux_eal_cpuflags.c.o 00:02:45.800 [13/378] Compiling C object lib/librte_eal.a.p/eal_x86_rte_spinlock.c.o 00:02:45.800 [14/378] Compiling C object lib/librte_log.a.p/log_log.c.o 00:02:45.800 [15/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_string_fns.c.o 00:02:45.800 [16/378] Compiling C object lib/librte_eal.a.p/eal_x86_rte_hypervisor.c.o 00:02:46.061 [17/378] Linking static target lib/librte_log.a 00:02:46.061 [18/378] Compiling C object lib/librte_eal.a.p/eal_unix_eal_debug.c.o 00:02:46.061 [19/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_cpuflags.c.o 00:02:46.333 [20/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_class.c.o 00:02:46.333 [21/378] Compiling C object lib/librte_eal.a.p/eal_linux_eal_timer.c.o 00:02:46.333 [22/378] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_timer.c.o 00:02:46.333 [23/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_config.c.o 00:02:46.333 [24/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_debug.c.o 00:02:46.333 [25/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_fbarray.c.o 00:02:46.333 [26/378] Generating lib/kvargs.sym_chk with a custom command (wrapped by meson to capture output) 00:02:46.333 [27/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_lcore.c.o 00:02:46.333 [28/378] Compiling C object lib/librte_eal.a.p/eal_common_rte_malloc.c.o 00:02:46.333 [29/378] Compiling C object lib/librte_eal.a.p/eal_common_rte_random.c.o 00:02:46.333 [30/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_tailqs.c.o 00:02:46.333 [31/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_bus.c.o 00:02:46.333 [32/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_mcfg.c.o 00:02:46.333 [33/378] Compiling C object lib/librte_eal.a.p/eal_linux_eal_lcore.c.o 00:02:46.333 [34/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace.c.o 00:02:46.333 [35/378] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops_default.c.o 00:02:46.333 [36/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memalloc.c.o 00:02:46.333 [37/378] Compiling C object lib/librte_eal.a.p/eal_unix_eal_firmware.c.o 00:02:46.333 [38/378] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_thread.c.o 00:02:46.333 [39/378] Compiling C object lib/librte_eal.a.p/eal_x86_rte_power_intrinsics.c.o 00:02:46.333 [40/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dev.c.o 00:02:46.333 [41/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memory.c.o 00:02:46.333 [42/378] Compiling C object lib/librte_eal.a.p/eal_common_hotplug_mp.c.o 00:02:46.333 [43/378] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio_mp_sync.c.o 00:02:46.333 [44/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_interrupts.c.o 00:02:46.333 [45/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_timer.c.o 00:02:46.333 [46/378] Compiling C object lib/librte_eal.a.p/eal_common_rte_keepalive.c.o 00:02:46.333 [47/378] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_legacy.c.o 00:02:46.333 [48/378] Compiling C object lib/librte_eal.a.p/eal_unix_eal_file.c.o 00:02:46.333 [49/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memzone.c.o 00:02:46.333 [50/378] Compiling C object lib/librte_eal.a.p/eal_linux_eal_thread.c.o 00:02:46.333 [51/378] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cycles.c.o 00:02:46.594 [52/378] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_memory.c.o 00:02:46.594 [53/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_devargs.c.o 00:02:46.594 [54/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_ctf.c.o 00:02:46.594 [55/378] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_pool_ops.c.o 00:02:46.594 [56/378] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_dyn.c.o 00:02:46.594 [57/378] Compiling C object lib/librte_eal.a.p/eal_common_malloc_elem.c.o 00:02:46.594 [58/378] Compiling C object lib/librte_eal.a.p/eal_unix_eal_filesystem.c.o 00:02:46.594 [59/378] Compiling C object lib/librte_eal.a.p/eal_unix_rte_thread.c.o 00:02:46.594 [60/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dynmem.c.o 00:02:46.594 [61/378] Compiling C object lib/librte_eal.a.p/eal_linux_eal_hugepage_info.c.o 00:02:46.594 [62/378] Compiling C object lib/librte_eal.a.p/eal_linux_eal_alarm.c.o 00:02:46.594 [63/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_points.c.o 00:02:46.594 [64/378] Compiling C object lib/librte_eal.a.p/eal_common_malloc_mp.c.o 00:02:46.594 [65/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_thread.c.o 00:02:46.594 [66/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_utils.c.o 00:02:46.594 [67/378] Compiling C object lib/librte_eal.a.p/eal_linux_eal_dev.c.o 00:02:46.594 [68/378] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_ptype.c.o 00:02:46.594 [69/378] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry.c.o 00:02:46.594 [70/378] Compiling C object lib/librte_ring.a.p/ring_rte_ring.c.o 00:02:46.594 [71/378] Compiling C object lib/librte_pci.a.p/pci_rte_pci.c.o 00:02:46.594 [72/378] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_cirbuf.c.o 00:02:46.594 [73/378] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_portlist.c.o 00:02:46.594 [74/378] Linking static target lib/librte_telemetry.a 00:02:46.594 [75/378] Compiling C object lib/librte_net.a.p/net_rte_net_crc.c.o 00:02:46.594 [76/378] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_num.c.o 00:02:46.594 [77/378] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_string.c.o 00:02:46.594 [78/378] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_vt100.c.o 00:02:46.594 [79/378] Linking static target lib/librte_ring.a 00:02:46.594 [80/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_proc.c.o 00:02:46.594 [81/378] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline.c.o 00:02:46.594 [82/378] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse.c.o 00:02:46.594 [83/378] Linking static target lib/librte_pci.a 00:02:46.594 [84/378] Compiling C object lib/librte_eal.a.p/eal_linux_eal.c.o 00:02:46.594 [85/378] Compiling C object lib/librte_mempool.a.p/mempool_mempool_trace_points.c.o 00:02:46.594 [86/378] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops.c.o 00:02:46.594 [87/378] Compiling C object lib/librte_eal.a.p/eal_common_malloc_heap.c.o 00:02:46.594 [88/378] Compiling C object lib/net/libnet_crc_avx512_lib.a.p/net_crc_avx512.c.o 00:02:46.594 [89/378] Compiling C object lib/librte_hash.a.p/hash_rte_hash_crc.c.o 00:02:46.594 [90/378] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memory.c.o 00:02:46.594 [91/378] Compiling C object lib/librte_eal.a.p/eal_common_rte_service.c.o 00:02:46.594 [92/378] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_socket.c.o 00:02:46.594 [93/378] Linking static target lib/net/libnet_crc_avx512_lib.a 00:02:46.594 [94/378] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memalloc.c.o 00:02:46.594 [95/378] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_os_unix.c.o 00:02:46.594 [96/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_options.c.o 00:02:46.594 [97/378] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_rdline.c.o 00:02:46.594 [98/378] Compiling C object lib/librte_power.a.p/power_power_common.c.o 00:02:46.594 [99/378] Compiling C object lib/librte_eal.a.p/eal_linux_eal_interrupts.c.o 00:02:46.594 [100/378] Compiling C object lib/librte_power.a.p/power_power_kvm_vm.c.o 00:02:46.594 [101/378] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio.c.o 00:02:46.594 [102/378] Compiling C object lib/librte_power.a.p/power_guest_channel.c.o 00:02:46.861 [103/378] Compiling C object lib/librte_net.a.p/net_rte_net.c.o 00:02:46.861 [104/378] Compiling C object lib/librte_vhost.a.p/vhost_fd_man.c.o 00:02:46.861 [105/378] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool.c.o 00:02:46.861 [106/378] Compiling C object lib/librte_meter.a.p/meter_rte_meter.c.o 00:02:46.861 [107/378] Linking static target lib/librte_mempool.a 00:02:46.861 [108/378] Compiling C object lib/librte_rcu.a.p/rcu_rte_rcu_qsbr.c.o 00:02:46.861 [109/378] Linking static target lib/librte_meter.a 00:02:46.861 [110/378] Linking static target lib/librte_rcu.a 00:02:46.861 [111/378] Compiling C object lib/librte_net.a.p/net_rte_arp.c.o 00:02:46.861 [112/378] Compiling C object drivers/libtmp_rte_bus_auxiliary.a.p/bus_auxiliary_auxiliary_params.c.o 00:02:46.861 [113/378] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_params.c.o 00:02:46.861 [114/378] Compiling C object lib/librte_net.a.p/net_net_crc_sse.c.o 00:02:46.861 [115/378] Generating lib/log.sym_chk with a custom command (wrapped by meson to capture output) 00:02:46.861 [116/378] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_launch.c.o 00:02:46.861 [117/378] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev_params.c.o 00:02:46.861 [118/378] Linking static target lib/librte_eal.a 00:02:46.861 [119/378] Linking target lib/librte_log.so.24.1 00:02:46.861 [120/378] Generating lib/pci.sym_chk with a custom command (wrapped by meson to capture output) 00:02:47.136 [121/378] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_ipaddr.c.o 00:02:47.136 [122/378] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_private.c.o 00:02:47.136 [123/378] Compiling C object lib/librte_hash.a.p/hash_rte_thash_gfni.c.o 00:02:47.136 [124/378] Generating lib/ring.sym_chk with a custom command (wrapped by meson to capture output) 00:02:47.136 [125/378] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf.c.o 00:02:47.136 [126/378] Linking static target lib/librte_mbuf.a 00:02:47.136 [127/378] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_profile.c.o 00:02:47.136 [128/378] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_etheraddr.c.o 00:02:47.136 [129/378] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_class_eth.c.o 00:02:47.136 [130/378] Linking static target lib/librte_cmdline.a 00:02:47.136 [131/378] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_telemetry.c.o 00:02:47.136 [132/378] Compiling C object lib/librte_hash.a.p/hash_rte_fbk_hash.c.o 00:02:47.136 [133/378] Compiling C object drivers/libtmp_rte_bus_auxiliary.a.p/bus_auxiliary_linux_auxiliary.c.o 00:02:47.136 [134/378] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev_trace_points.c.o 00:02:47.136 [135/378] Generating lib/meter.sym_chk with a custom command (wrapped by meson to capture output) 00:02:47.136 [136/378] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev_pmd.c.o 00:02:47.136 [137/378] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8472.c.o 00:02:47.136 [138/378] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_pmd.c.o 00:02:47.136 [139/378] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_linux_ethtool.c.o 00:02:47.136 [140/378] Compiling C object lib/librte_timer.a.p/timer_rte_timer.c.o 00:02:47.136 [141/378] Compiling C object lib/librte_power.a.p/power_rte_power_uncore.c.o 00:02:47.136 [142/378] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8079.c.o 00:02:47.136 [143/378] Compiling C object lib/librte_power.a.p/power_power_acpi_cpufreq.c.o 00:02:47.136 [144/378] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_driver.c.o 00:02:47.136 [145/378] Linking static target lib/librte_timer.a 00:02:47.136 [146/378] Compiling C object lib/librte_power.a.p/power_power_amd_pstate_cpufreq.c.o 00:02:47.136 [147/378] Compiling C object lib/librte_vhost.a.p/vhost_vdpa.c.o 00:02:47.136 [148/378] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_common.c.o 00:02:47.136 [149/378] Compiling C object lib/librte_power.a.p/power_rte_power.c.o 00:02:47.136 [150/378] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_mtr.c.o 00:02:47.136 [151/378] Compiling C object lib/librte_power.a.p/power_power_intel_uncore.c.o 00:02:47.136 [152/378] Generating lib/rcu.sym_chk with a custom command (wrapped by meson to capture output) 00:02:47.404 [153/378] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev.c.o 00:02:47.404 [154/378] Compiling C object drivers/libtmp_rte_bus_auxiliary.a.p/bus_auxiliary_auxiliary_common.c.o 00:02:47.404 [155/378] Generating symbol file lib/librte_log.so.24.1.p/librte_log.so.24.1.symbols 00:02:47.404 [156/378] Compiling C object lib/librte_power.a.p/power_power_cppc_cpufreq.c.o 00:02:47.404 [157/378] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_cman.c.o 00:02:47.404 [158/378] Linking static target drivers/libtmp_rte_bus_auxiliary.a 00:02:47.404 [159/378] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_telemetry.c.o 00:02:47.404 [160/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/common_qat_qat_logs.c.o 00:02:47.404 [161/378] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8636.c.o 00:02:47.404 [162/378] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev.c.o 00:02:47.404 [163/378] Linking static target lib/librte_dmadev.a 00:02:47.404 [164/378] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_trace_points.c.o 00:02:47.404 [165/378] Generating lib/telemetry.sym_chk with a custom command (wrapped by meson to capture output) 00:02:47.404 [166/378] Compiling C object lib/librte_hash.a.p/hash_rte_thash.c.o 00:02:47.404 [167/378] Compiling C object lib/librte_power.a.p/power_power_pstate_cpufreq.c.o 00:02:47.404 [168/378] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_vfio.c.o 00:02:47.404 [169/378] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common_uio.c.o 00:02:47.404 [170/378] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_comp.c.o 00:02:47.404 [171/378] Linking target lib/librte_kvargs.so.24.1 00:02:47.404 [172/378] Linking static target lib/librte_compressdev.a 00:02:47.404 [173/378] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common.c.o 00:02:47.404 [174/378] Compiling C object lib/librte_power.a.p/power_rte_power_pmd_mgmt.c.o 00:02:47.404 [175/378] Compiling C object lib/librte_vhost.a.p/vhost_iotlb.c.o 00:02:47.404 [176/378] Linking static target lib/librte_power.a 00:02:47.404 [177/378] Linking target lib/librte_telemetry.so.24.1 00:02:47.404 [178/378] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci.c.o 00:02:47.404 [179/378] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net_ctrl.c.o 00:02:47.404 [180/378] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_uio.c.o 00:02:47.404 [181/378] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev.c.o 00:02:47.404 [182/378] Compiling C object lib/librte_net.a.p/net_rte_ether.c.o 00:02:47.404 [183/378] Linking static target drivers/libtmp_rte_bus_pci.a 00:02:47.404 [184/378] Linking static target drivers/libtmp_rte_bus_vdev.a 00:02:47.404 [185/378] Linking static target lib/librte_net.a 00:02:47.404 [186/378] Compiling C object lib/librte_reorder.a.p/reorder_rte_reorder.c.o 00:02:47.404 [187/378] Linking static target lib/librte_reorder.a 00:02:47.404 [188/378] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_tm.c.o 00:02:47.672 [189/378] Compiling C object lib/librte_vhost.a.p/vhost_socket.c.o 00:02:47.672 [190/378] Compiling C object drivers/libtmp_rte_common_mlx5.a.p/common_mlx5_linux_mlx5_glue.c.o 00:02:47.672 [191/378] Compiling C object lib/librte_vhost.a.p/vhost_vduse.c.o 00:02:47.672 [192/378] Compiling C object lib/librte_security.a.p/security_rte_security.c.o 00:02:47.672 [193/378] Generating symbol file lib/librte_kvargs.so.24.1.p/librte_kvargs.so.24.1.symbols 00:02:47.672 [194/378] Linking static target lib/librte_security.a 00:02:47.672 [195/378] Compiling C object drivers/libtmp_rte_common_mlx5.a.p/common_mlx5_mlx5_common.c.o 00:02:47.672 [196/378] Generating symbol file lib/librte_telemetry.so.24.1.p/librte_telemetry.so.24.1.symbols 00:02:47.672 [197/378] Generating drivers/rte_bus_auxiliary.pmd.c with a custom command 00:02:47.672 [198/378] Compiling C object drivers/libtmp_rte_common_mlx5.a.p/common_mlx5_mlx5_common_mp.c.o 00:02:47.672 [199/378] Compiling C object drivers/librte_bus_auxiliary.a.p/meson-generated_.._rte_bus_auxiliary.pmd.c.o 00:02:47.672 [200/378] Linking static target drivers/librte_bus_auxiliary.a 00:02:47.672 [201/378] Compiling C object drivers/librte_bus_auxiliary.so.24.1.p/meson-generated_.._rte_bus_auxiliary.pmd.c.o 00:02:47.672 [202/378] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_trace_points.c.o 00:02:47.933 [203/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/common_qat_dev_qat_dev_gen_lce.c.o 00:02:47.933 [204/378] Generating drivers/rte_bus_vdev.pmd.c with a custom command 00:02:47.933 [205/378] Compiling C object drivers/libtmp_rte_common_mlx5.a.p/common_mlx5_mlx5_malloc.c.o 00:02:47.933 [206/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/compress_qat_dev_qat_comp_pmd_gen5.c.o 00:02:47.933 [207/378] Compiling C object lib/librte_hash.a.p/hash_rte_cuckoo_hash.c.o 00:02:47.933 [208/378] Compiling C object drivers/libtmp_rte_common_mlx5.a.p/common_mlx5_linux_mlx5_common_auxiliary.c.o 00:02:47.933 [209/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/compress_qat_dev_qat_comp_pmd_gen3.c.o 00:02:47.933 [210/378] Linking static target lib/librte_hash.a 00:02:47.933 [211/378] Compiling C object lib/librte_vhost.a.p/vhost_vhost_user.c.o 00:02:47.933 [212/378] Compiling C object lib/librte_vhost.a.p/vhost_vhost.c.o 00:02:47.933 [213/378] Compiling C object drivers/librte_bus_vdev.a.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:02:47.933 [214/378] Compiling C object drivers/librte_bus_vdev.so.24.1.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:02:47.933 [215/378] Generating drivers/rte_bus_pci.pmd.c with a custom command 00:02:47.933 [216/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/common_qat_qat_common.c.o 00:02:47.933 [217/378] Linking static target drivers/librte_bus_vdev.a 00:02:47.933 [218/378] Compiling C object drivers/libtmp_rte_common_mlx5.a.p/common_mlx5_mlx5_common_pci.c.o 00:02:47.933 [219/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/common_qat_dev_qat_dev_gen3.c.o 00:02:47.933 [220/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/common_qat_dev_qat_dev_gen2.c.o 00:02:47.933 [221/378] Compiling C object drivers/libtmp_rte_common_mlx5.a.p/common_mlx5_mlx5_common_devx.c.o 00:02:47.933 [222/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/common_qat_dev_qat_dev_gen5.c.o 00:02:47.933 [223/378] Compiling C object drivers/librte_bus_pci.a.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:02:47.934 [224/378] Compiling C object drivers/librte_bus_pci.so.24.1.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:02:47.934 [225/378] Generating lib/timer.sym_chk with a custom command (wrapped by meson to capture output) 00:02:47.934 [226/378] Linking static target drivers/librte_bus_pci.a 00:02:47.934 [227/378] Compiling C object drivers/libtmp_rte_common_mlx5.a.p/common_mlx5_mlx5_common_utils.c.o 00:02:47.934 [228/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/compress_qat_dev_qat_comp_pmd_gen2.c.o 00:02:47.934 [229/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/common_qat_dev_qat_dev_gen4.c.o 00:02:47.934 [230/378] Generating lib/mempool.sym_chk with a custom command (wrapped by meson to capture output) 00:02:47.934 [231/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/crypto_qat_qat_crypto.c.o 00:02:47.934 [232/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/common_qat_dev_qat_dev_gen1.c.o 00:02:47.934 [233/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/common_qat_qat_pf2vf.c.o 00:02:47.934 [234/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/compress_qat_dev_qat_comp_pmd_gen4.c.o 00:02:47.934 [235/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/crypto_qat_dev_qat_crypto_pmd_gen2.c.o 00:02:47.934 [236/378] Generating lib/net.sym_chk with a custom command (wrapped by meson to capture output) 00:02:47.934 [237/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/compress_qat_dev_qat_comp_pmd_gen1.c.o 00:02:47.934 [238/378] Compiling C object drivers/libtmp_rte_common_mlx5.a.p/common_mlx5_linux_mlx5_common_verbs.c.o 00:02:47.934 [239/378] Generating lib/mbuf.sym_chk with a custom command (wrapped by meson to capture output) 00:02:48.192 [240/378] Generating lib/reorder.sym_chk with a custom command (wrapped by meson to capture output) 00:02:48.192 [241/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/compress_qat_qat_comp_pmd.c.o 00:02:48.192 [242/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/common_qat_qat_device.c.o 00:02:48.192 [243/378] Generating drivers/rte_bus_auxiliary.sym_chk with a custom command (wrapped by meson to capture output) 00:02:48.192 [244/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/crypto_qat_qat_sym.c.o 00:02:48.192 [245/378] Compiling C object drivers/libtmp_rte_common_mlx5.a.p/common_mlx5_linux_mlx5_nl.c.o 00:02:48.192 [246/378] Compiling C object drivers/libtmp_rte_common_mlx5.a.p/common_mlx5_mlx5_common_mr.c.o 00:02:48.192 [247/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/crypto_qat_dev_qat_asym_pmd_gen1.c.o 00:02:48.192 [248/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/crypto_qat_dev_qat_crypto_pmd_gen5.c.o 00:02:48.192 [249/378] Generating lib/dmadev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:48.192 [250/378] Compiling C object lib/librte_cryptodev.a.p/cryptodev_rte_cryptodev.c.o 00:02:48.192 [251/378] Compiling C object drivers/libtmp_rte_crypto_ipsec_mb.a.p/crypto_ipsec_mb_ipsec_mb_ops.c.o 00:02:48.192 [252/378] Linking static target lib/librte_cryptodev.a 00:02:48.192 [253/378] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_flow.c.o 00:02:48.192 [254/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/crypto_qat_dev_qat_crypto_pmd_gen_lce.c.o 00:02:48.192 [255/378] Generating lib/compressdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:48.192 [256/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/common_qat_qat_qp.c.o 00:02:48.192 [257/378] Compiling C object drivers/libtmp_rte_mempool_ring.a.p/mempool_ring_rte_mempool_ring.c.o 00:02:48.192 [258/378] Linking static target drivers/libtmp_rte_mempool_ring.a 00:02:48.192 [259/378] Generating lib/security.sym_chk with a custom command (wrapped by meson to capture output) 00:02:48.192 [260/378] Compiling C object drivers/libtmp_rte_crypto_ipsec_mb.a.p/crypto_ipsec_mb_ipsec_mb_private.c.o 00:02:48.192 [261/378] Generating drivers/rte_bus_vdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:48.497 [262/378] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev.c.o 00:02:48.497 [263/378] Generating lib/power.sym_chk with a custom command (wrapped by meson to capture output) 00:02:48.497 [264/378] Compiling C object drivers/libtmp_rte_crypto_mlx5.a.p/crypto_mlx5_mlx5_crypto_dek.c.o 00:02:48.497 [265/378] Linking static target lib/librte_ethdev.a 00:02:48.497 [266/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/crypto_qat_dev_qat_crypto_pmd_gen4.c.o 00:02:48.497 [267/378] Compiling C object drivers/libtmp_rte_crypto_mlx5.a.p/crypto_mlx5_mlx5_crypto.c.o 00:02:48.497 [268/378] Compiling C object drivers/libtmp_rte_crypto_ipsec_mb.a.p/crypto_ipsec_mb_pmd_aesni_gcm.c.o 00:02:48.497 [269/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/crypto_qat_qat_sym_session.c.o 00:02:48.497 [270/378] Generating drivers/rte_mempool_ring.pmd.c with a custom command 00:02:48.497 [271/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/compress_qat_qat_comp.c.o 00:02:48.497 [272/378] Compiling C object drivers/libtmp_rte_crypto_ipsec_mb.a.p/crypto_ipsec_mb_pmd_kasumi.c.o 00:02:48.497 [273/378] Compiling C object drivers/librte_mempool_ring.so.24.1.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:02:48.497 [274/378] Compiling C object drivers/libtmp_rte_crypto_mlx5.a.p/crypto_mlx5_mlx5_crypto_xts.c.o 00:02:48.497 [275/378] Compiling C object drivers/libtmp_rte_crypto_ipsec_mb.a.p/crypto_ipsec_mb_pmd_chacha_poly.c.o 00:02:48.497 [276/378] Compiling C object drivers/libtmp_rte_common_mlx5.a.p/common_mlx5_mlx5_devx_cmds.c.o 00:02:48.497 [277/378] Compiling C object drivers/librte_mempool_ring.a.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:02:48.497 [278/378] Linking static target drivers/librte_mempool_ring.a 00:02:48.497 [279/378] Generating lib/cmdline.sym_chk with a custom command (wrapped by meson to capture output) 00:02:48.498 [280/378] Compiling C object drivers/libtmp_rte_compress_isal.a.p/compress_isal_isal_compress_pmd_ops.c.o 00:02:48.498 [281/378] Compiling C object drivers/libtmp_rte_compress_isal.a.p/compress_isal_isal_compress_pmd.c.o 00:02:48.498 [282/378] Linking static target drivers/libtmp_rte_compress_isal.a 00:02:48.498 [283/378] Compiling C object drivers/libtmp_rte_crypto_ipsec_mb.a.p/crypto_ipsec_mb_pmd_aesni_mb.c.o 00:02:48.498 [284/378] Compiling C object lib/librte_vhost.a.p/vhost_vhost_crypto.c.o 00:02:48.498 [285/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/crypto_qat_dev_qat_crypto_pmd_gen3.c.o 00:02:48.498 [286/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/crypto_qat_dev_qat_sym_pmd_gen1.c.o 00:02:48.756 [287/378] Compiling C object drivers/libtmp_rte_crypto_ipsec_mb.a.p/crypto_ipsec_mb_pmd_zuc.c.o 00:02:48.756 [288/378] Compiling C object drivers/libtmp_rte_compress_mlx5.a.p/compress_mlx5_mlx5_compress.c.o 00:02:48.756 [289/378] Compiling C object drivers/libtmp_rte_crypto_mlx5.a.p/crypto_mlx5_mlx5_crypto_gcm.c.o 00:02:48.756 [290/378] Linking static target drivers/libtmp_rte_compress_mlx5.a 00:02:48.756 [291/378] Linking static target drivers/libtmp_rte_crypto_mlx5.a 00:02:48.756 [292/378] Generating drivers/rte_bus_pci.sym_chk with a custom command (wrapped by meson to capture output) 00:02:48.756 [293/378] Compiling C object drivers/libtmp_rte_common_mlx5.a.p/common_mlx5_linux_mlx5_common_os.c.o 00:02:48.756 [294/378] Compiling C object drivers/libtmp_rte_crypto_ipsec_mb.a.p/crypto_ipsec_mb_pmd_snow3g.c.o 00:02:48.756 [295/378] Linking static target drivers/libtmp_rte_common_mlx5.a 00:02:48.756 [296/378] Linking static target drivers/libtmp_rte_crypto_ipsec_mb.a 00:02:48.756 [297/378] Generating drivers/rte_compress_isal.pmd.c with a custom command 00:02:48.756 [298/378] Generating lib/hash.sym_chk with a custom command (wrapped by meson to capture output) 00:02:48.756 [299/378] Compiling C object drivers/librte_compress_isal.so.24.1.p/meson-generated_.._rte_compress_isal.pmd.c.o 00:02:48.756 [300/378] Compiling C object drivers/librte_compress_isal.a.p/meson-generated_.._rte_compress_isal.pmd.c.o 00:02:48.756 [301/378] Linking static target drivers/librte_compress_isal.a 00:02:49.014 [302/378] Generating drivers/rte_compress_mlx5.pmd.c with a custom command 00:02:49.015 [303/378] Generating drivers/rte_crypto_mlx5.pmd.c with a custom command 00:02:49.015 [304/378] Compiling C object drivers/librte_compress_mlx5.a.p/meson-generated_.._rte_compress_mlx5.pmd.c.o 00:02:49.015 [305/378] Compiling C object drivers/librte_compress_mlx5.so.24.1.p/meson-generated_.._rte_compress_mlx5.pmd.c.o 00:02:49.015 [306/378] Compiling C object drivers/librte_crypto_mlx5.so.24.1.p/meson-generated_.._rte_crypto_mlx5.pmd.c.o 00:02:49.015 [307/378] Compiling C object drivers/librte_crypto_mlx5.a.p/meson-generated_.._rte_crypto_mlx5.pmd.c.o 00:02:49.015 [308/378] Linking static target drivers/librte_compress_mlx5.a 00:02:49.015 [309/378] Linking static target drivers/librte_crypto_mlx5.a 00:02:49.015 [310/378] Generating drivers/rte_common_mlx5.pmd.c with a custom command 00:02:49.015 [311/378] Compiling C object drivers/librte_common_mlx5.so.24.1.p/meson-generated_.._rte_common_mlx5.pmd.c.o 00:02:49.015 [312/378] Compiling C object drivers/librte_common_mlx5.a.p/meson-generated_.._rte_common_mlx5.pmd.c.o 00:02:49.015 [313/378] Generating drivers/rte_crypto_ipsec_mb.pmd.c with a custom command 00:02:49.015 [314/378] Linking static target drivers/librte_common_mlx5.a 00:02:49.015 [315/378] Compiling C object drivers/librte_crypto_ipsec_mb.so.24.1.p/meson-generated_.._rte_crypto_ipsec_mb.pmd.c.o 00:02:49.015 [316/378] Compiling C object drivers/librte_crypto_ipsec_mb.a.p/meson-generated_.._rte_crypto_ipsec_mb.pmd.c.o 00:02:49.015 [317/378] Linking static target drivers/librte_crypto_ipsec_mb.a 00:02:49.582 [318/378] Compiling C object drivers/libtmp_rte_common_qat.a.p/crypto_qat_qat_asym.c.o 00:02:49.582 [319/378] Linking static target drivers/libtmp_rte_common_qat.a 00:02:49.839 [320/378] Generating drivers/rte_common_qat.pmd.c with a custom command 00:02:49.839 [321/378] Compiling C object drivers/librte_common_qat.so.24.1.p/meson-generated_.._rte_common_qat.pmd.c.o 00:02:49.839 [322/378] Compiling C object drivers/librte_common_qat.a.p/meson-generated_.._rte_common_qat.pmd.c.o 00:02:49.839 [323/378] Linking static target drivers/librte_common_qat.a 00:02:50.097 [324/378] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net.c.o 00:02:50.097 [325/378] Linking static target lib/librte_vhost.a 00:02:50.354 [326/378] Generating lib/cryptodev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:52.882 [327/378] Generating lib/vhost.sym_chk with a custom command (wrapped by meson to capture output) 00:02:54.780 [328/378] Generating drivers/rte_common_mlx5.sym_chk with a custom command (wrapped by meson to capture output) 00:02:58.056 [329/378] Generating lib/ethdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:59.430 [330/378] Generating lib/eal.sym_chk with a custom command (wrapped by meson to capture output) 00:02:59.430 [331/378] Linking target lib/librte_eal.so.24.1 00:02:59.689 [332/378] Generating symbol file lib/librte_eal.so.24.1.p/librte_eal.so.24.1.symbols 00:02:59.689 [333/378] Linking target lib/librte_meter.so.24.1 00:02:59.689 [334/378] Linking target lib/librte_pci.so.24.1 00:02:59.689 [335/378] Linking target lib/librte_ring.so.24.1 00:02:59.689 [336/378] Linking target lib/librte_timer.so.24.1 00:02:59.689 [337/378] Linking target lib/librte_dmadev.so.24.1 00:02:59.689 [338/378] Linking target drivers/librte_bus_auxiliary.so.24.1 00:02:59.689 [339/378] Linking target drivers/librte_bus_vdev.so.24.1 00:02:59.947 [340/378] Generating symbol file lib/librte_ring.so.24.1.p/librte_ring.so.24.1.symbols 00:02:59.947 [341/378] Generating symbol file lib/librte_pci.so.24.1.p/librte_pci.so.24.1.symbols 00:02:59.947 [342/378] Generating symbol file drivers/librte_bus_vdev.so.24.1.p/librte_bus_vdev.so.24.1.symbols 00:02:59.947 [343/378] Generating symbol file lib/librte_timer.so.24.1.p/librte_timer.so.24.1.symbols 00:02:59.947 [344/378] Generating symbol file drivers/librte_bus_auxiliary.so.24.1.p/librte_bus_auxiliary.so.24.1.symbols 00:02:59.947 [345/378] Generating symbol file lib/librte_meter.so.24.1.p/librte_meter.so.24.1.symbols 00:02:59.947 [346/378] Linking target lib/librte_mempool.so.24.1 00:02:59.947 [347/378] Generating symbol file lib/librte_dmadev.so.24.1.p/librte_dmadev.so.24.1.symbols 00:02:59.947 [348/378] Linking target lib/librte_rcu.so.24.1 00:02:59.947 [349/378] Linking target drivers/librte_bus_pci.so.24.1 00:02:59.947 [350/378] Generating symbol file lib/librte_mempool.so.24.1.p/librte_mempool.so.24.1.symbols 00:02:59.947 [351/378] Generating symbol file lib/librte_rcu.so.24.1.p/librte_rcu.so.24.1.symbols 00:02:59.947 [352/378] Generating symbol file drivers/librte_bus_pci.so.24.1.p/librte_bus_pci.so.24.1.symbols 00:03:00.205 [353/378] Linking target drivers/librte_mempool_ring.so.24.1 00:03:00.205 [354/378] Linking target lib/librte_mbuf.so.24.1 00:03:00.205 [355/378] Generating symbol file lib/librte_mbuf.so.24.1.p/librte_mbuf.so.24.1.symbols 00:03:00.205 [356/378] Linking target lib/librte_reorder.so.24.1 00:03:00.205 [357/378] Linking target lib/librte_net.so.24.1 00:03:00.205 [358/378] Linking target lib/librte_compressdev.so.24.1 00:03:00.205 [359/378] Linking target lib/librte_cryptodev.so.24.1 00:03:00.463 [360/378] Generating symbol file lib/librte_net.so.24.1.p/librte_net.so.24.1.symbols 00:03:00.463 [361/378] Generating symbol file lib/librte_compressdev.so.24.1.p/librte_compressdev.so.24.1.symbols 00:03:00.463 [362/378] Generating symbol file lib/librte_cryptodev.so.24.1.p/librte_cryptodev.so.24.1.symbols 00:03:00.463 [363/378] Linking target lib/librte_cmdline.so.24.1 00:03:00.463 [364/378] Linking target lib/librte_hash.so.24.1 00:03:00.463 [365/378] Linking target drivers/librte_compress_isal.so.24.1 00:03:00.463 [366/378] Linking target lib/librte_security.so.24.1 00:03:00.721 [367/378] Linking target lib/librte_ethdev.so.24.1 00:03:00.721 [368/378] Generating symbol file lib/librte_security.so.24.1.p/librte_security.so.24.1.symbols 00:03:00.721 [369/378] Generating symbol file lib/librte_hash.so.24.1.p/librte_hash.so.24.1.symbols 00:03:00.721 [370/378] Generating symbol file lib/librte_ethdev.so.24.1.p/librte_ethdev.so.24.1.symbols 00:03:00.721 [371/378] Linking target drivers/librte_common_mlx5.so.24.1 00:03:00.721 [372/378] Linking target lib/librte_power.so.24.1 00:03:00.721 [373/378] Linking target lib/librte_vhost.so.24.1 00:03:00.978 [374/378] Generating symbol file drivers/librte_common_mlx5.so.24.1.p/librte_common_mlx5.so.24.1.symbols 00:03:00.978 [375/378] Linking target drivers/librte_crypto_mlx5.so.24.1 00:03:00.978 [376/378] Linking target drivers/librte_compress_mlx5.so.24.1 00:03:00.978 [377/378] Linking target drivers/librte_crypto_ipsec_mb.so.24.1 00:03:00.978 [378/378] Linking target drivers/librte_common_qat.so.24.1 00:03:00.978 INFO: autodetecting backend as ninja 00:03:00.978 INFO: calculating backend command to run: /usr/local/bin/ninja -C /var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build-tmp -j 72 00:03:02.355 CC lib/ut/ut.o 00:03:02.356 CC lib/log/log.o 00:03:02.356 CC lib/log/log_flags.o 00:03:02.356 CC lib/log/log_deprecated.o 00:03:02.356 CC lib/ut_mock/mock.o 00:03:02.613 LIB libspdk_log.a 00:03:02.613 LIB libspdk_ut.a 00:03:02.613 LIB libspdk_ut_mock.a 00:03:02.613 SO libspdk_log.so.7.0 00:03:02.613 SO libspdk_ut.so.2.0 00:03:02.613 SO libspdk_ut_mock.so.6.0 00:03:02.613 SYMLINK libspdk_log.so 00:03:02.613 SYMLINK libspdk_ut.so 00:03:02.613 SYMLINK libspdk_ut_mock.so 00:03:03.180 CC lib/ioat/ioat.o 00:03:03.180 CC lib/util/base64.o 00:03:03.180 CC lib/util/bit_array.o 00:03:03.180 CC lib/util/cpuset.o 00:03:03.180 CC lib/dma/dma.o 00:03:03.180 CC lib/util/crc16.o 00:03:03.180 CC lib/util/crc32.o 00:03:03.180 CC lib/util/crc32c.o 00:03:03.180 CC lib/util/crc32_ieee.o 00:03:03.180 CC lib/util/crc64.o 00:03:03.180 CXX lib/trace_parser/trace.o 00:03:03.180 CC lib/util/dif.o 00:03:03.180 CC lib/util/fd.o 00:03:03.180 CC lib/util/file.o 00:03:03.180 CC lib/util/hexlify.o 00:03:03.180 CC lib/util/iov.o 00:03:03.180 CC lib/util/math.o 00:03:03.180 CC lib/util/pipe.o 00:03:03.180 CC lib/util/strerror_tls.o 00:03:03.180 CC lib/util/string.o 00:03:03.180 CC lib/util/uuid.o 00:03:03.180 CC lib/util/fd_group.o 00:03:03.180 CC lib/util/xor.o 00:03:03.180 CC lib/util/zipf.o 00:03:03.180 CC lib/vfio_user/host/vfio_user_pci.o 00:03:03.180 CC lib/vfio_user/host/vfio_user.o 00:03:03.180 LIB libspdk_dma.a 00:03:03.180 SO libspdk_dma.so.4.0 00:03:03.438 LIB libspdk_ioat.a 00:03:03.438 SYMLINK libspdk_dma.so 00:03:03.438 SO libspdk_ioat.so.7.0 00:03:03.438 SYMLINK libspdk_ioat.so 00:03:03.438 LIB libspdk_vfio_user.a 00:03:03.438 SO libspdk_vfio_user.so.5.0 00:03:03.695 LIB libspdk_util.a 00:03:03.695 SYMLINK libspdk_vfio_user.so 00:03:03.695 SO libspdk_util.so.9.1 00:03:03.953 SYMLINK libspdk_util.so 00:03:03.953 LIB libspdk_trace_parser.a 00:03:03.953 SO libspdk_trace_parser.so.5.0 00:03:04.212 SYMLINK libspdk_trace_parser.so 00:03:04.212 CC lib/vmd/vmd.o 00:03:04.212 CC lib/vmd/led.o 00:03:04.212 CC lib/conf/conf.o 00:03:04.212 CC lib/json/json_parse.o 00:03:04.212 CC lib/json/json_write.o 00:03:04.212 CC lib/json/json_util.o 00:03:04.212 CC lib/rdma_provider/common.o 00:03:04.212 CC lib/rdma_provider/rdma_provider_verbs.o 00:03:04.212 CC lib/idxd/idxd.o 00:03:04.212 CC lib/env_dpdk/env.o 00:03:04.212 CC lib/reduce/reduce.o 00:03:04.212 CC lib/idxd/idxd_user.o 00:03:04.212 CC lib/rdma_utils/rdma_utils.o 00:03:04.212 CC lib/idxd/idxd_kernel.o 00:03:04.212 CC lib/env_dpdk/memory.o 00:03:04.212 CC lib/env_dpdk/pci.o 00:03:04.212 CC lib/env_dpdk/init.o 00:03:04.212 CC lib/env_dpdk/threads.o 00:03:04.212 CC lib/env_dpdk/pci_ioat.o 00:03:04.212 CC lib/env_dpdk/pci_virtio.o 00:03:04.212 CC lib/env_dpdk/pci_vmd.o 00:03:04.212 CC lib/env_dpdk/pci_idxd.o 00:03:04.212 CC lib/env_dpdk/pci_event.o 00:03:04.212 CC lib/env_dpdk/sigbus_handler.o 00:03:04.212 CC lib/env_dpdk/pci_dpdk.o 00:03:04.212 CC lib/env_dpdk/pci_dpdk_2207.o 00:03:04.212 CC lib/env_dpdk/pci_dpdk_2211.o 00:03:04.469 LIB libspdk_conf.a 00:03:04.469 LIB libspdk_rdma_provider.a 00:03:04.469 SO libspdk_conf.so.6.0 00:03:04.469 SO libspdk_rdma_provider.so.6.0 00:03:04.469 LIB libspdk_json.a 00:03:04.469 LIB libspdk_rdma_utils.a 00:03:04.469 SYMLINK libspdk_conf.so 00:03:04.469 SYMLINK libspdk_rdma_provider.so 00:03:04.469 SO libspdk_json.so.6.0 00:03:04.727 SO libspdk_rdma_utils.so.1.0 00:03:04.727 SYMLINK libspdk_rdma_utils.so 00:03:04.727 SYMLINK libspdk_json.so 00:03:04.727 LIB libspdk_idxd.a 00:03:04.727 SO libspdk_idxd.so.12.0 00:03:04.985 LIB libspdk_vmd.a 00:03:04.985 SYMLINK libspdk_idxd.so 00:03:04.985 SO libspdk_vmd.so.6.0 00:03:04.985 LIB libspdk_reduce.a 00:03:04.985 SO libspdk_reduce.so.6.0 00:03:04.985 SYMLINK libspdk_vmd.so 00:03:04.985 SYMLINK libspdk_reduce.so 00:03:04.985 CC lib/jsonrpc/jsonrpc_server.o 00:03:04.985 CC lib/jsonrpc/jsonrpc_client.o 00:03:04.985 CC lib/jsonrpc/jsonrpc_server_tcp.o 00:03:04.985 CC lib/jsonrpc/jsonrpc_client_tcp.o 00:03:05.243 LIB libspdk_jsonrpc.a 00:03:05.243 LIB libspdk_env_dpdk.a 00:03:05.501 SO libspdk_jsonrpc.so.6.0 00:03:05.501 SO libspdk_env_dpdk.so.14.1 00:03:05.501 SYMLINK libspdk_jsonrpc.so 00:03:05.501 SYMLINK libspdk_env_dpdk.so 00:03:05.814 CC lib/rpc/rpc.o 00:03:06.072 LIB libspdk_rpc.a 00:03:06.072 SO libspdk_rpc.so.6.0 00:03:06.330 SYMLINK libspdk_rpc.so 00:03:06.588 CC lib/keyring/keyring_rpc.o 00:03:06.588 CC lib/keyring/keyring.o 00:03:06.588 CC lib/trace/trace.o 00:03:06.588 CC lib/trace/trace_flags.o 00:03:06.588 CC lib/trace/trace_rpc.o 00:03:06.588 CC lib/notify/notify_rpc.o 00:03:06.588 CC lib/notify/notify.o 00:03:06.847 LIB libspdk_notify.a 00:03:06.847 SO libspdk_notify.so.6.0 00:03:06.847 LIB libspdk_keyring.a 00:03:06.847 SO libspdk_keyring.so.1.0 00:03:06.847 LIB libspdk_trace.a 00:03:06.847 SYMLINK libspdk_notify.so 00:03:06.847 SO libspdk_trace.so.10.0 00:03:06.847 SYMLINK libspdk_keyring.so 00:03:06.847 SYMLINK libspdk_trace.so 00:03:07.415 CC lib/sock/sock_rpc.o 00:03:07.415 CC lib/sock/sock.o 00:03:07.415 CC lib/thread/thread.o 00:03:07.415 CC lib/thread/iobuf.o 00:03:07.674 LIB libspdk_sock.a 00:03:07.674 SO libspdk_sock.so.10.0 00:03:07.933 SYMLINK libspdk_sock.so 00:03:08.191 CC lib/nvme/nvme_ctrlr_cmd.o 00:03:08.191 CC lib/nvme/nvme_ctrlr.o 00:03:08.191 CC lib/nvme/nvme_ns.o 00:03:08.191 CC lib/nvme/nvme_fabric.o 00:03:08.191 CC lib/nvme/nvme_ns_cmd.o 00:03:08.191 CC lib/nvme/nvme_pcie_common.o 00:03:08.191 CC lib/nvme/nvme_pcie.o 00:03:08.191 CC lib/nvme/nvme_qpair.o 00:03:08.191 CC lib/nvme/nvme.o 00:03:08.191 CC lib/nvme/nvme_transport.o 00:03:08.191 CC lib/nvme/nvme_quirks.o 00:03:08.191 CC lib/nvme/nvme_ctrlr_ocssd_cmd.o 00:03:08.191 CC lib/nvme/nvme_discovery.o 00:03:08.191 CC lib/nvme/nvme_ns_ocssd_cmd.o 00:03:08.191 CC lib/nvme/nvme_tcp.o 00:03:08.191 CC lib/nvme/nvme_opal.o 00:03:08.191 CC lib/nvme/nvme_io_msg.o 00:03:08.191 CC lib/nvme/nvme_zns.o 00:03:08.191 CC lib/nvme/nvme_stubs.o 00:03:08.191 CC lib/nvme/nvme_poll_group.o 00:03:08.191 CC lib/nvme/nvme_auth.o 00:03:08.191 CC lib/nvme/nvme_cuse.o 00:03:08.191 CC lib/nvme/nvme_rdma.o 00:03:08.757 LIB libspdk_thread.a 00:03:09.016 SO libspdk_thread.so.10.1 00:03:09.016 SYMLINK libspdk_thread.so 00:03:09.274 CC lib/accel/accel.o 00:03:09.274 CC lib/accel/accel_rpc.o 00:03:09.274 CC lib/accel/accel_sw.o 00:03:09.274 CC lib/blob/blobstore.o 00:03:09.274 CC lib/blob/request.o 00:03:09.274 CC lib/blob/zeroes.o 00:03:09.274 CC lib/blob/blob_bs_dev.o 00:03:09.274 CC lib/init/json_config.o 00:03:09.274 CC lib/init/subsystem.o 00:03:09.274 CC lib/init/subsystem_rpc.o 00:03:09.274 CC lib/init/rpc.o 00:03:09.274 CC lib/virtio/virtio.o 00:03:09.274 CC lib/virtio/virtio_vhost_user.o 00:03:09.274 CC lib/virtio/virtio_vfio_user.o 00:03:09.274 CC lib/virtio/virtio_pci.o 00:03:09.532 LIB libspdk_init.a 00:03:09.791 SO libspdk_init.so.5.0 00:03:09.791 LIB libspdk_virtio.a 00:03:09.791 SYMLINK libspdk_init.so 00:03:09.791 SO libspdk_virtio.so.7.0 00:03:09.791 SYMLINK libspdk_virtio.so 00:03:10.050 CC lib/event/app.o 00:03:10.050 CC lib/event/log_rpc.o 00:03:10.050 CC lib/event/reactor.o 00:03:10.050 CC lib/event/app_rpc.o 00:03:10.050 CC lib/event/scheduler_static.o 00:03:10.308 LIB libspdk_accel.a 00:03:10.309 SO libspdk_accel.so.15.1 00:03:10.567 SYMLINK libspdk_accel.so 00:03:10.567 LIB libspdk_event.a 00:03:10.567 SO libspdk_event.so.14.0 00:03:10.567 SYMLINK libspdk_event.so 00:03:10.824 CC lib/bdev/bdev.o 00:03:10.824 CC lib/bdev/bdev_rpc.o 00:03:10.824 CC lib/bdev/bdev_zone.o 00:03:10.824 CC lib/bdev/part.o 00:03:10.824 CC lib/bdev/scsi_nvme.o 00:03:12.723 LIB libspdk_blob.a 00:03:12.723 SO libspdk_blob.so.11.0 00:03:12.723 SYMLINK libspdk_blob.so 00:03:12.723 LIB libspdk_nvme.a 00:03:12.723 CC lib/blobfs/blobfs.o 00:03:12.981 CC lib/blobfs/tree.o 00:03:12.981 CC lib/lvol/lvol.o 00:03:12.981 SO libspdk_nvme.so.13.1 00:03:13.238 SYMLINK libspdk_nvme.so 00:03:13.496 LIB libspdk_blobfs.a 00:03:13.496 SO libspdk_blobfs.so.10.0 00:03:13.496 LIB libspdk_bdev.a 00:03:13.496 SYMLINK libspdk_blobfs.so 00:03:13.753 SO libspdk_bdev.so.15.1 00:03:13.753 SYMLINK libspdk_bdev.so 00:03:13.753 LIB libspdk_lvol.a 00:03:13.753 SO libspdk_lvol.so.10.0 00:03:14.017 SYMLINK libspdk_lvol.so 00:03:14.017 CC lib/ftl/ftl_core.o 00:03:14.017 CC lib/ftl/ftl_layout.o 00:03:14.017 CC lib/ftl/ftl_init.o 00:03:14.017 CC lib/ftl/ftl_debug.o 00:03:14.017 CC lib/ublk/ublk_rpc.o 00:03:14.017 CC lib/ftl/ftl_io.o 00:03:14.017 CC lib/ublk/ublk.o 00:03:14.017 CC lib/ftl/ftl_sb.o 00:03:14.017 CC lib/ftl/ftl_l2p.o 00:03:14.017 CC lib/ftl/ftl_l2p_flat.o 00:03:14.017 CC lib/ftl/ftl_band.o 00:03:14.017 CC lib/ftl/ftl_nv_cache.o 00:03:14.017 CC lib/ftl/ftl_band_ops.o 00:03:14.017 CC lib/ftl/ftl_writer.o 00:03:14.017 CC lib/ftl/ftl_rq.o 00:03:14.017 CC lib/ftl/ftl_reloc.o 00:03:14.017 CC lib/nbd/nbd.o 00:03:14.017 CC lib/ftl/mngt/ftl_mngt.o 00:03:14.017 CC lib/ftl/ftl_l2p_cache.o 00:03:14.017 CC lib/ftl/ftl_p2l.o 00:03:14.017 CC lib/nbd/nbd_rpc.o 00:03:14.017 CC lib/ftl/mngt/ftl_mngt_startup.o 00:03:14.017 CC lib/ftl/mngt/ftl_mngt_bdev.o 00:03:14.017 CC lib/ftl/mngt/ftl_mngt_shutdown.o 00:03:14.017 CC lib/ftl/mngt/ftl_mngt_md.o 00:03:14.017 CC lib/ftl/mngt/ftl_mngt_misc.o 00:03:14.017 CC lib/ftl/mngt/ftl_mngt_ioch.o 00:03:14.017 CC lib/ftl/mngt/ftl_mngt_l2p.o 00:03:14.017 CC lib/ftl/mngt/ftl_mngt_band.o 00:03:14.017 CC lib/ftl/mngt/ftl_mngt_self_test.o 00:03:14.017 CC lib/ftl/mngt/ftl_mngt_p2l.o 00:03:14.017 CC lib/ftl/mngt/ftl_mngt_recovery.o 00:03:14.017 CC lib/ftl/mngt/ftl_mngt_upgrade.o 00:03:14.017 CC lib/scsi/lun.o 00:03:14.017 CC lib/scsi/dev.o 00:03:14.017 CC lib/ftl/utils/ftl_conf.o 00:03:14.017 CC lib/scsi/port.o 00:03:14.017 CC lib/ftl/utils/ftl_md.o 00:03:14.017 CC lib/scsi/scsi.o 00:03:14.017 CC lib/ftl/utils/ftl_mempool.o 00:03:14.017 CC lib/scsi/scsi_bdev.o 00:03:14.017 CC lib/scsi/scsi_pr.o 00:03:14.017 CC lib/nvmf/ctrlr.o 00:03:14.017 CC lib/ftl/utils/ftl_bitmap.o 00:03:14.017 CC lib/scsi/scsi_rpc.o 00:03:14.017 CC lib/scsi/task.o 00:03:14.017 CC lib/nvmf/ctrlr_discovery.o 00:03:14.017 CC lib/ftl/utils/ftl_property.o 00:03:14.017 CC lib/nvmf/ctrlr_bdev.o 00:03:14.017 CC lib/ftl/utils/ftl_layout_tracker_bdev.o 00:03:14.017 CC lib/nvmf/subsystem.o 00:03:14.017 CC lib/ftl/upgrade/ftl_layout_upgrade.o 00:03:14.017 CC lib/nvmf/nvmf_rpc.o 00:03:14.017 CC lib/ftl/upgrade/ftl_sb_upgrade.o 00:03:14.017 CC lib/ftl/upgrade/ftl_p2l_upgrade.o 00:03:14.017 CC lib/nvmf/tcp.o 00:03:14.017 CC lib/nvmf/transport.o 00:03:14.017 CC lib/nvmf/nvmf.o 00:03:14.017 CC lib/ftl/upgrade/ftl_band_upgrade.o 00:03:14.017 CC lib/ftl/upgrade/ftl_trim_upgrade.o 00:03:14.017 CC lib/ftl/upgrade/ftl_chunk_upgrade.o 00:03:14.017 CC lib/nvmf/stubs.o 00:03:14.017 CC lib/ftl/upgrade/ftl_sb_v3.o 00:03:14.017 CC lib/nvmf/mdns_server.o 00:03:14.017 CC lib/nvmf/rdma.o 00:03:14.017 CC lib/ftl/nvc/ftl_nvc_dev.o 00:03:14.017 CC lib/nvmf/auth.o 00:03:14.017 CC lib/ftl/upgrade/ftl_sb_v5.o 00:03:14.017 CC lib/ftl/base/ftl_base_dev.o 00:03:14.017 CC lib/ftl/nvc/ftl_nvc_bdev_vss.o 00:03:14.017 CC lib/ftl/base/ftl_base_bdev.o 00:03:14.279 CC lib/ftl/ftl_trace.o 00:03:14.845 LIB libspdk_scsi.a 00:03:14.845 SO libspdk_scsi.so.9.0 00:03:15.103 LIB libspdk_ublk.a 00:03:15.103 SYMLINK libspdk_scsi.so 00:03:15.103 LIB libspdk_nbd.a 00:03:15.103 SO libspdk_ublk.so.3.0 00:03:15.103 SO libspdk_nbd.so.7.0 00:03:15.103 SYMLINK libspdk_ublk.so 00:03:15.103 SYMLINK libspdk_nbd.so 00:03:15.361 CC lib/vhost/vhost_scsi.o 00:03:15.361 CC lib/vhost/vhost.o 00:03:15.361 CC lib/vhost/vhost_rpc.o 00:03:15.361 CC lib/vhost/vhost_blk.o 00:03:15.361 CC lib/vhost/rte_vhost_user.o 00:03:15.361 CC lib/iscsi/conn.o 00:03:15.361 CC lib/iscsi/init_grp.o 00:03:15.361 CC lib/iscsi/iscsi.o 00:03:15.361 CC lib/iscsi/md5.o 00:03:15.361 CC lib/iscsi/param.o 00:03:15.361 LIB libspdk_ftl.a 00:03:15.361 CC lib/iscsi/portal_grp.o 00:03:15.361 CC lib/iscsi/tgt_node.o 00:03:15.361 CC lib/iscsi/iscsi_rpc.o 00:03:15.361 CC lib/iscsi/iscsi_subsystem.o 00:03:15.361 CC lib/iscsi/task.o 00:03:15.618 SO libspdk_ftl.so.9.0 00:03:15.876 SYMLINK libspdk_ftl.so 00:03:16.442 LIB libspdk_nvmf.a 00:03:16.442 LIB libspdk_vhost.a 00:03:16.442 SO libspdk_vhost.so.8.0 00:03:16.701 SO libspdk_nvmf.so.18.1 00:03:16.701 SYMLINK libspdk_vhost.so 00:03:16.701 SYMLINK libspdk_nvmf.so 00:03:16.701 LIB libspdk_iscsi.a 00:03:16.959 SO libspdk_iscsi.so.8.0 00:03:16.959 SYMLINK libspdk_iscsi.so 00:03:17.526 CC module/env_dpdk/env_dpdk_rpc.o 00:03:17.785 CC module/accel/dpdk_compressdev/accel_dpdk_compressdev.o 00:03:17.785 CC module/accel/error/accel_error.o 00:03:17.785 LIB libspdk_env_dpdk_rpc.a 00:03:17.785 CC module/accel/dpdk_compressdev/accel_dpdk_compressdev_rpc.o 00:03:17.785 CC module/accel/error/accel_error_rpc.o 00:03:17.785 CC module/keyring/file/keyring.o 00:03:17.785 CC module/keyring/file/keyring_rpc.o 00:03:17.785 CC module/scheduler/dpdk_governor/dpdk_governor.o 00:03:17.785 CC module/sock/posix/posix.o 00:03:17.785 CC module/scheduler/dynamic/scheduler_dynamic.o 00:03:17.785 CC module/accel/ioat/accel_ioat_rpc.o 00:03:17.785 CC module/accel/ioat/accel_ioat.o 00:03:17.785 CC module/keyring/linux/keyring.o 00:03:17.785 CC module/keyring/linux/keyring_rpc.o 00:03:17.785 CC module/scheduler/gscheduler/gscheduler.o 00:03:17.785 CC module/accel/iaa/accel_iaa.o 00:03:17.785 CC module/accel/dpdk_cryptodev/accel_dpdk_cryptodev.o 00:03:17.785 CC module/accel/iaa/accel_iaa_rpc.o 00:03:17.785 CC module/blob/bdev/blob_bdev.o 00:03:17.785 CC module/accel/dpdk_cryptodev/accel_dpdk_cryptodev_rpc.o 00:03:17.785 CC module/accel/dsa/accel_dsa.o 00:03:17.785 CC module/accel/dsa/accel_dsa_rpc.o 00:03:17.785 SO libspdk_env_dpdk_rpc.so.6.0 00:03:17.785 SYMLINK libspdk_env_dpdk_rpc.so 00:03:18.043 LIB libspdk_accel_error.a 00:03:18.043 LIB libspdk_keyring_linux.a 00:03:18.043 LIB libspdk_keyring_file.a 00:03:18.043 LIB libspdk_scheduler_dpdk_governor.a 00:03:18.043 LIB libspdk_scheduler_gscheduler.a 00:03:18.043 SO libspdk_accel_error.so.2.0 00:03:18.043 SO libspdk_keyring_linux.so.1.0 00:03:18.043 SO libspdk_keyring_file.so.1.0 00:03:18.043 SO libspdk_scheduler_dpdk_governor.so.4.0 00:03:18.043 LIB libspdk_accel_ioat.a 00:03:18.043 SO libspdk_scheduler_gscheduler.so.4.0 00:03:18.043 LIB libspdk_scheduler_dynamic.a 00:03:18.043 LIB libspdk_accel_iaa.a 00:03:18.043 SYMLINK libspdk_accel_error.so 00:03:18.043 SO libspdk_accel_ioat.so.6.0 00:03:18.043 SYMLINK libspdk_keyring_linux.so 00:03:18.043 SO libspdk_scheduler_dynamic.so.4.0 00:03:18.043 SO libspdk_accel_iaa.so.3.0 00:03:18.043 SYMLINK libspdk_keyring_file.so 00:03:18.043 SYMLINK libspdk_scheduler_dpdk_governor.so 00:03:18.043 LIB libspdk_accel_dsa.a 00:03:18.043 LIB libspdk_blob_bdev.a 00:03:18.043 SYMLINK libspdk_scheduler_gscheduler.so 00:03:18.043 SO libspdk_blob_bdev.so.11.0 00:03:18.043 SYMLINK libspdk_scheduler_dynamic.so 00:03:18.043 SYMLINK libspdk_accel_iaa.so 00:03:18.043 SO libspdk_accel_dsa.so.5.0 00:03:18.043 SYMLINK libspdk_accel_ioat.so 00:03:18.302 SYMLINK libspdk_blob_bdev.so 00:03:18.302 SYMLINK libspdk_accel_dsa.so 00:03:18.560 LIB libspdk_sock_posix.a 00:03:18.560 SO libspdk_sock_posix.so.6.0 00:03:18.560 CC module/bdev/gpt/vbdev_gpt.o 00:03:18.560 CC module/bdev/gpt/gpt.o 00:03:18.560 CC module/bdev/malloc/bdev_malloc.o 00:03:18.560 CC module/bdev/malloc/bdev_malloc_rpc.o 00:03:18.560 CC module/bdev/nvme/bdev_nvme.o 00:03:18.560 CC module/bdev/nvme/bdev_nvme_rpc.o 00:03:18.560 CC module/bdev/nvme/bdev_mdns_client.o 00:03:18.560 CC module/bdev/zone_block/vbdev_zone_block.o 00:03:18.560 CC module/bdev/nvme/nvme_rpc.o 00:03:18.560 CC module/bdev/nvme/bdev_nvme_cuse_rpc.o 00:03:18.560 CC module/bdev/nvme/vbdev_opal.o 00:03:18.560 CC module/bdev/nvme/vbdev_opal_rpc.o 00:03:18.560 CC module/bdev/zone_block/vbdev_zone_block_rpc.o 00:03:18.560 CC module/bdev/raid/bdev_raid.o 00:03:18.560 CC module/bdev/raid/bdev_raid_rpc.o 00:03:18.560 CC module/bdev/raid/bdev_raid_sb.o 00:03:18.560 CC module/bdev/raid/raid0.o 00:03:18.560 CC module/bdev/raid/raid1.o 00:03:18.560 CC module/bdev/raid/concat.o 00:03:18.560 CC module/bdev/error/vbdev_error.o 00:03:18.560 CC module/bdev/passthru/vbdev_passthru.o 00:03:18.560 CC module/bdev/iscsi/bdev_iscsi.o 00:03:18.560 CC module/bdev/passthru/vbdev_passthru_rpc.o 00:03:18.560 CC module/bdev/crypto/vbdev_crypto.o 00:03:18.560 CC module/blobfs/bdev/blobfs_bdev_rpc.o 00:03:18.560 CC module/bdev/error/vbdev_error_rpc.o 00:03:18.560 CC module/blobfs/bdev/blobfs_bdev.o 00:03:18.560 SYMLINK libspdk_sock_posix.so 00:03:18.560 CC module/bdev/crypto/vbdev_crypto_rpc.o 00:03:18.560 CC module/bdev/iscsi/bdev_iscsi_rpc.o 00:03:18.560 CC module/bdev/null/bdev_null.o 00:03:18.560 CC module/bdev/ftl/bdev_ftl.o 00:03:18.560 CC module/bdev/lvol/vbdev_lvol.o 00:03:18.560 CC module/bdev/null/bdev_null_rpc.o 00:03:18.560 CC module/bdev/ftl/bdev_ftl_rpc.o 00:03:18.560 CC module/bdev/lvol/vbdev_lvol_rpc.o 00:03:18.560 CC module/bdev/delay/vbdev_delay.o 00:03:18.560 CC module/bdev/delay/vbdev_delay_rpc.o 00:03:18.560 CC module/bdev/virtio/bdev_virtio_scsi.o 00:03:18.818 CC module/bdev/compress/vbdev_compress.o 00:03:18.818 CC module/bdev/virtio/bdev_virtio_blk.o 00:03:18.818 CC module/bdev/compress/vbdev_compress_rpc.o 00:03:18.818 CC module/bdev/aio/bdev_aio_rpc.o 00:03:18.818 CC module/bdev/virtio/bdev_virtio_rpc.o 00:03:18.818 CC module/bdev/aio/bdev_aio.o 00:03:18.818 CC module/bdev/split/vbdev_split.o 00:03:18.818 CC module/bdev/split/vbdev_split_rpc.o 00:03:18.818 LIB libspdk_accel_dpdk_compressdev.a 00:03:19.076 LIB libspdk_bdev_error.a 00:03:19.076 SO libspdk_accel_dpdk_compressdev.so.3.0 00:03:19.076 LIB libspdk_bdev_ftl.a 00:03:19.076 SO libspdk_bdev_error.so.6.0 00:03:19.076 SO libspdk_bdev_ftl.so.6.0 00:03:19.076 LIB libspdk_blobfs_bdev.a 00:03:19.076 SYMLINK libspdk_accel_dpdk_compressdev.so 00:03:19.076 LIB libspdk_bdev_passthru.a 00:03:19.076 SO libspdk_blobfs_bdev.so.6.0 00:03:19.076 LIB libspdk_bdev_zone_block.a 00:03:19.076 LIB libspdk_bdev_gpt.a 00:03:19.076 SYMLINK libspdk_bdev_error.so 00:03:19.076 SYMLINK libspdk_bdev_ftl.so 00:03:19.076 LIB libspdk_bdev_malloc.a 00:03:19.076 LIB libspdk_bdev_split.a 00:03:19.076 LIB libspdk_bdev_null.a 00:03:19.076 SO libspdk_bdev_zone_block.so.6.0 00:03:19.076 SO libspdk_bdev_passthru.so.6.0 00:03:19.076 LIB libspdk_bdev_crypto.a 00:03:19.076 SO libspdk_bdev_gpt.so.6.0 00:03:19.076 SO libspdk_bdev_malloc.so.6.0 00:03:19.076 SO libspdk_bdev_split.so.6.0 00:03:19.076 SYMLINK libspdk_blobfs_bdev.so 00:03:19.076 LIB libspdk_bdev_compress.a 00:03:19.076 SO libspdk_bdev_null.so.6.0 00:03:19.076 SO libspdk_bdev_crypto.so.6.0 00:03:19.076 LIB libspdk_bdev_aio.a 00:03:19.076 SYMLINK libspdk_bdev_passthru.so 00:03:19.076 SYMLINK libspdk_bdev_zone_block.so 00:03:19.076 SO libspdk_bdev_compress.so.6.0 00:03:19.335 LIB libspdk_bdev_delay.a 00:03:19.335 SYMLINK libspdk_bdev_gpt.so 00:03:19.335 SYMLINK libspdk_bdev_malloc.so 00:03:19.335 SYMLINK libspdk_bdev_split.so 00:03:19.335 LIB libspdk_accel_dpdk_cryptodev.a 00:03:19.335 SYMLINK libspdk_bdev_null.so 00:03:19.335 SO libspdk_bdev_aio.so.6.0 00:03:19.335 SYMLINK libspdk_bdev_crypto.so 00:03:19.335 SO libspdk_bdev_delay.so.6.0 00:03:19.335 LIB libspdk_bdev_iscsi.a 00:03:19.335 SO libspdk_accel_dpdk_cryptodev.so.3.0 00:03:19.335 SYMLINK libspdk_bdev_compress.so 00:03:19.335 SO libspdk_bdev_iscsi.so.6.0 00:03:19.335 SYMLINK libspdk_bdev_aio.so 00:03:19.335 SYMLINK libspdk_bdev_delay.so 00:03:19.335 LIB libspdk_bdev_lvol.a 00:03:19.335 SYMLINK libspdk_accel_dpdk_cryptodev.so 00:03:19.335 LIB libspdk_bdev_virtio.a 00:03:19.335 SYMLINK libspdk_bdev_iscsi.so 00:03:19.335 SO libspdk_bdev_lvol.so.6.0 00:03:19.335 SO libspdk_bdev_virtio.so.6.0 00:03:19.593 SYMLINK libspdk_bdev_lvol.so 00:03:19.593 LIB libspdk_bdev_raid.a 00:03:19.593 SYMLINK libspdk_bdev_virtio.so 00:03:19.593 SO libspdk_bdev_raid.so.6.0 00:03:19.593 SYMLINK libspdk_bdev_raid.so 00:03:20.528 LIB libspdk_bdev_nvme.a 00:03:20.528 SO libspdk_bdev_nvme.so.7.0 00:03:20.528 SYMLINK libspdk_bdev_nvme.so 00:03:21.493 CC module/event/subsystems/sock/sock.o 00:03:21.493 CC module/event/subsystems/vhost_blk/vhost_blk.o 00:03:21.494 CC module/event/subsystems/keyring/keyring.o 00:03:21.494 CC module/event/subsystems/iobuf/iobuf.o 00:03:21.494 CC module/event/subsystems/scheduler/scheduler.o 00:03:21.494 CC module/event/subsystems/iobuf/iobuf_rpc.o 00:03:21.494 CC module/event/subsystems/vmd/vmd.o 00:03:21.494 CC module/event/subsystems/vmd/vmd_rpc.o 00:03:21.494 LIB libspdk_event_keyring.a 00:03:21.494 LIB libspdk_event_sock.a 00:03:21.494 LIB libspdk_event_scheduler.a 00:03:21.494 LIB libspdk_event_vhost_blk.a 00:03:21.494 SO libspdk_event_keyring.so.1.0 00:03:21.494 LIB libspdk_event_iobuf.a 00:03:21.494 LIB libspdk_event_vmd.a 00:03:21.494 SO libspdk_event_sock.so.5.0 00:03:21.494 SO libspdk_event_vhost_blk.so.3.0 00:03:21.494 SO libspdk_event_scheduler.so.4.0 00:03:21.494 SO libspdk_event_iobuf.so.3.0 00:03:21.494 SO libspdk_event_vmd.so.6.0 00:03:21.752 SYMLINK libspdk_event_keyring.so 00:03:21.752 SYMLINK libspdk_event_scheduler.so 00:03:21.752 SYMLINK libspdk_event_sock.so 00:03:21.752 SYMLINK libspdk_event_vhost_blk.so 00:03:21.752 SYMLINK libspdk_event_iobuf.so 00:03:21.752 SYMLINK libspdk_event_vmd.so 00:03:22.011 CC module/event/subsystems/accel/accel.o 00:03:22.270 LIB libspdk_event_accel.a 00:03:22.270 SO libspdk_event_accel.so.6.0 00:03:22.270 SYMLINK libspdk_event_accel.so 00:03:22.836 CC module/event/subsystems/bdev/bdev.o 00:03:22.836 LIB libspdk_event_bdev.a 00:03:22.836 SO libspdk_event_bdev.so.6.0 00:03:23.094 SYMLINK libspdk_event_bdev.so 00:03:23.352 CC module/event/subsystems/nvmf/nvmf_rpc.o 00:03:23.352 CC module/event/subsystems/nvmf/nvmf_tgt.o 00:03:23.352 CC module/event/subsystems/scsi/scsi.o 00:03:23.352 CC module/event/subsystems/nbd/nbd.o 00:03:23.352 CC module/event/subsystems/ublk/ublk.o 00:03:23.610 LIB libspdk_event_nbd.a 00:03:23.610 LIB libspdk_event_scsi.a 00:03:23.610 LIB libspdk_event_ublk.a 00:03:23.610 SO libspdk_event_nbd.so.6.0 00:03:23.610 SO libspdk_event_ublk.so.3.0 00:03:23.610 SO libspdk_event_scsi.so.6.0 00:03:23.610 SYMLINK libspdk_event_scsi.so 00:03:23.610 SYMLINK libspdk_event_nbd.so 00:03:23.610 SYMLINK libspdk_event_ublk.so 00:03:23.610 LIB libspdk_event_nvmf.a 00:03:23.868 SO libspdk_event_nvmf.so.6.0 00:03:23.868 SYMLINK libspdk_event_nvmf.so 00:03:23.868 CC module/event/subsystems/iscsi/iscsi.o 00:03:23.868 CC module/event/subsystems/vhost_scsi/vhost_scsi.o 00:03:24.126 LIB libspdk_event_iscsi.a 00:03:24.126 LIB libspdk_event_vhost_scsi.a 00:03:24.126 SO libspdk_event_iscsi.so.6.0 00:03:24.126 SO libspdk_event_vhost_scsi.so.3.0 00:03:24.126 SYMLINK libspdk_event_vhost_scsi.so 00:03:24.126 SYMLINK libspdk_event_iscsi.so 00:03:24.385 SO libspdk.so.6.0 00:03:24.385 SYMLINK libspdk.so 00:03:24.643 CC app/spdk_nvme_identify/identify.o 00:03:24.903 CC app/spdk_nvme_discover/discovery_aer.o 00:03:24.903 CXX app/trace/trace.o 00:03:24.903 CC app/spdk_nvme_perf/perf.o 00:03:24.903 CC app/spdk_lspci/spdk_lspci.o 00:03:24.903 CC test/rpc_client/rpc_client_test.o 00:03:24.903 CC app/trace_record/trace_record.o 00:03:24.903 TEST_HEADER include/spdk/accel.h 00:03:24.903 TEST_HEADER include/spdk/accel_module.h 00:03:24.903 TEST_HEADER include/spdk/barrier.h 00:03:24.903 TEST_HEADER include/spdk/assert.h 00:03:24.903 TEST_HEADER include/spdk/base64.h 00:03:24.903 TEST_HEADER include/spdk/bdev.h 00:03:24.903 TEST_HEADER include/spdk/bdev_module.h 00:03:24.903 CC app/spdk_top/spdk_top.o 00:03:24.903 TEST_HEADER include/spdk/bdev_zone.h 00:03:24.903 TEST_HEADER include/spdk/bit_array.h 00:03:24.903 TEST_HEADER include/spdk/bit_pool.h 00:03:24.903 TEST_HEADER include/spdk/blob_bdev.h 00:03:24.903 TEST_HEADER include/spdk/blobfs.h 00:03:24.903 TEST_HEADER include/spdk/blobfs_bdev.h 00:03:24.903 TEST_HEADER include/spdk/blob.h 00:03:24.903 TEST_HEADER include/spdk/conf.h 00:03:24.903 TEST_HEADER include/spdk/config.h 00:03:24.903 TEST_HEADER include/spdk/cpuset.h 00:03:24.903 TEST_HEADER include/spdk/crc16.h 00:03:24.903 TEST_HEADER include/spdk/crc32.h 00:03:24.903 TEST_HEADER include/spdk/crc64.h 00:03:24.903 TEST_HEADER include/spdk/dif.h 00:03:24.903 TEST_HEADER include/spdk/dma.h 00:03:24.903 TEST_HEADER include/spdk/endian.h 00:03:24.903 TEST_HEADER include/spdk/env_dpdk.h 00:03:24.903 TEST_HEADER include/spdk/env.h 00:03:24.903 TEST_HEADER include/spdk/event.h 00:03:24.903 TEST_HEADER include/spdk/fd_group.h 00:03:24.903 TEST_HEADER include/spdk/fd.h 00:03:24.903 TEST_HEADER include/spdk/file.h 00:03:24.904 TEST_HEADER include/spdk/ftl.h 00:03:24.904 TEST_HEADER include/spdk/gpt_spec.h 00:03:24.904 TEST_HEADER include/spdk/hexlify.h 00:03:24.904 TEST_HEADER include/spdk/histogram_data.h 00:03:24.904 TEST_HEADER include/spdk/idxd_spec.h 00:03:24.904 TEST_HEADER include/spdk/idxd.h 00:03:24.904 TEST_HEADER include/spdk/init.h 00:03:24.904 TEST_HEADER include/spdk/ioat.h 00:03:24.904 TEST_HEADER include/spdk/ioat_spec.h 00:03:24.904 TEST_HEADER include/spdk/iscsi_spec.h 00:03:24.904 TEST_HEADER include/spdk/json.h 00:03:24.904 TEST_HEADER include/spdk/jsonrpc.h 00:03:24.904 TEST_HEADER include/spdk/keyring.h 00:03:24.904 TEST_HEADER include/spdk/keyring_module.h 00:03:24.904 TEST_HEADER include/spdk/likely.h 00:03:24.904 TEST_HEADER include/spdk/log.h 00:03:24.904 TEST_HEADER include/spdk/lvol.h 00:03:24.904 TEST_HEADER include/spdk/memory.h 00:03:24.904 TEST_HEADER include/spdk/mmio.h 00:03:24.904 TEST_HEADER include/spdk/nbd.h 00:03:24.904 TEST_HEADER include/spdk/nvme.h 00:03:24.904 TEST_HEADER include/spdk/notify.h 00:03:24.904 TEST_HEADER include/spdk/nvme_intel.h 00:03:24.904 TEST_HEADER include/spdk/nvme_ocssd.h 00:03:24.904 TEST_HEADER include/spdk/nvme_ocssd_spec.h 00:03:24.904 TEST_HEADER include/spdk/nvme_spec.h 00:03:24.904 TEST_HEADER include/spdk/nvme_zns.h 00:03:24.904 CC examples/interrupt_tgt/interrupt_tgt.o 00:03:24.904 TEST_HEADER include/spdk/nvmf_cmd.h 00:03:24.904 CC app/nvmf_tgt/nvmf_main.o 00:03:24.904 TEST_HEADER include/spdk/nvmf_fc_spec.h 00:03:24.904 TEST_HEADER include/spdk/nvmf.h 00:03:24.904 TEST_HEADER include/spdk/nvmf_spec.h 00:03:24.904 TEST_HEADER include/spdk/nvmf_transport.h 00:03:24.904 TEST_HEADER include/spdk/opal.h 00:03:24.904 TEST_HEADER include/spdk/opal_spec.h 00:03:24.904 TEST_HEADER include/spdk/pci_ids.h 00:03:24.904 CC app/iscsi_tgt/iscsi_tgt.o 00:03:24.904 TEST_HEADER include/spdk/pipe.h 00:03:24.904 TEST_HEADER include/spdk/queue.h 00:03:24.904 TEST_HEADER include/spdk/reduce.h 00:03:24.904 TEST_HEADER include/spdk/rpc.h 00:03:24.904 CC app/spdk_dd/spdk_dd.o 00:03:24.904 TEST_HEADER include/spdk/scsi.h 00:03:24.904 TEST_HEADER include/spdk/scheduler.h 00:03:24.904 TEST_HEADER include/spdk/scsi_spec.h 00:03:24.904 TEST_HEADER include/spdk/sock.h 00:03:24.904 TEST_HEADER include/spdk/stdinc.h 00:03:24.904 TEST_HEADER include/spdk/string.h 00:03:24.904 TEST_HEADER include/spdk/thread.h 00:03:24.904 TEST_HEADER include/spdk/trace_parser.h 00:03:24.904 TEST_HEADER include/spdk/ublk.h 00:03:24.904 TEST_HEADER include/spdk/tree.h 00:03:24.904 TEST_HEADER include/spdk/trace.h 00:03:24.904 TEST_HEADER include/spdk/util.h 00:03:24.904 TEST_HEADER include/spdk/uuid.h 00:03:24.904 TEST_HEADER include/spdk/version.h 00:03:24.904 TEST_HEADER include/spdk/vfio_user_pci.h 00:03:24.904 TEST_HEADER include/spdk/vfio_user_spec.h 00:03:24.904 TEST_HEADER include/spdk/vhost.h 00:03:24.904 TEST_HEADER include/spdk/vmd.h 00:03:24.904 TEST_HEADER include/spdk/zipf.h 00:03:24.904 TEST_HEADER include/spdk/xor.h 00:03:24.904 CXX test/cpp_headers/accel.o 00:03:24.904 CXX test/cpp_headers/accel_module.o 00:03:24.904 CXX test/cpp_headers/assert.o 00:03:24.904 CXX test/cpp_headers/barrier.o 00:03:24.904 CXX test/cpp_headers/base64.o 00:03:24.904 CXX test/cpp_headers/bdev.o 00:03:24.904 CXX test/cpp_headers/bdev_zone.o 00:03:24.904 CXX test/cpp_headers/bdev_module.o 00:03:24.904 CXX test/cpp_headers/bit_array.o 00:03:24.904 CXX test/cpp_headers/bit_pool.o 00:03:24.904 CXX test/cpp_headers/blob_bdev.o 00:03:24.904 CXX test/cpp_headers/blob.o 00:03:24.904 CXX test/cpp_headers/blobfs_bdev.o 00:03:24.904 CXX test/cpp_headers/blobfs.o 00:03:24.904 CXX test/cpp_headers/conf.o 00:03:24.904 CXX test/cpp_headers/config.o 00:03:24.904 CXX test/cpp_headers/cpuset.o 00:03:24.904 CXX test/cpp_headers/crc16.o 00:03:24.904 CXX test/cpp_headers/crc64.o 00:03:24.904 CXX test/cpp_headers/crc32.o 00:03:24.904 CXX test/cpp_headers/dif.o 00:03:24.904 CXX test/cpp_headers/dma.o 00:03:24.904 CXX test/cpp_headers/endian.o 00:03:24.904 CXX test/cpp_headers/env_dpdk.o 00:03:24.904 CC app/spdk_tgt/spdk_tgt.o 00:03:24.904 CXX test/cpp_headers/event.o 00:03:24.904 CXX test/cpp_headers/env.o 00:03:24.904 CXX test/cpp_headers/fd_group.o 00:03:24.904 CXX test/cpp_headers/fd.o 00:03:24.904 CXX test/cpp_headers/ftl.o 00:03:24.904 CXX test/cpp_headers/file.o 00:03:24.904 CXX test/cpp_headers/hexlify.o 00:03:24.904 CXX test/cpp_headers/histogram_data.o 00:03:24.904 CXX test/cpp_headers/idxd.o 00:03:24.904 CXX test/cpp_headers/idxd_spec.o 00:03:24.904 CXX test/cpp_headers/gpt_spec.o 00:03:24.904 CXX test/cpp_headers/init.o 00:03:24.904 CXX test/cpp_headers/ioat_spec.o 00:03:24.904 CXX test/cpp_headers/ioat.o 00:03:24.904 CXX test/cpp_headers/iscsi_spec.o 00:03:24.904 CXX test/cpp_headers/json.o 00:03:24.904 CXX test/cpp_headers/jsonrpc.o 00:03:24.904 CXX test/cpp_headers/keyring.o 00:03:24.904 CC examples/ioat/verify/verify.o 00:03:24.904 CC test/env/pci/pci_ut.o 00:03:24.904 CC test/thread/poller_perf/poller_perf.o 00:03:24.904 CXX test/cpp_headers/keyring_module.o 00:03:24.904 CC app/fio/nvme/fio_plugin.o 00:03:24.904 CC test/env/vtophys/vtophys.o 00:03:24.904 CC examples/ioat/perf/perf.o 00:03:24.904 CC test/env/env_dpdk_post_init/env_dpdk_post_init.o 00:03:24.904 CC examples/util/zipf/zipf.o 00:03:24.904 CC test/env/memory/memory_ut.o 00:03:24.904 CC test/app/histogram_perf/histogram_perf.o 00:03:24.904 CC test/app/jsoncat/jsoncat.o 00:03:25.169 CC test/dma/test_dma/test_dma.o 00:03:25.170 CC test/app/stub/stub.o 00:03:25.170 LINK spdk_lspci 00:03:25.170 CC app/fio/bdev/fio_plugin.o 00:03:25.170 CC test/app/bdev_svc/bdev_svc.o 00:03:25.170 LINK rpc_client_test 00:03:25.170 LINK spdk_nvme_discover 00:03:25.431 LINK interrupt_tgt 00:03:25.431 CC test/env/mem_callbacks/mem_callbacks.o 00:03:25.431 LINK vtophys 00:03:25.431 LINK poller_perf 00:03:25.431 CC test/app/fuzz/nvme_fuzz/nvme_fuzz.o 00:03:25.431 CXX test/cpp_headers/likely.o 00:03:25.431 CXX test/cpp_headers/log.o 00:03:25.431 CXX test/cpp_headers/lvol.o 00:03:25.431 CXX test/cpp_headers/memory.o 00:03:25.431 CXX test/cpp_headers/mmio.o 00:03:25.431 LINK iscsi_tgt 00:03:25.431 LINK spdk_trace_record 00:03:25.431 LINK nvmf_tgt 00:03:25.431 LINK zipf 00:03:25.431 CXX test/cpp_headers/nbd.o 00:03:25.431 CXX test/cpp_headers/notify.o 00:03:25.431 CXX test/cpp_headers/nvme.o 00:03:25.431 CXX test/cpp_headers/nvme_intel.o 00:03:25.431 CXX test/cpp_headers/nvme_ocssd.o 00:03:25.431 CXX test/cpp_headers/nvme_ocssd_spec.o 00:03:25.431 CXX test/cpp_headers/nvme_spec.o 00:03:25.431 CXX test/cpp_headers/nvme_zns.o 00:03:25.431 CXX test/cpp_headers/nvmf_cmd.o 00:03:25.431 CXX test/cpp_headers/nvmf_fc_spec.o 00:03:25.431 CXX test/cpp_headers/nvmf.o 00:03:25.431 CXX test/cpp_headers/nvmf_spec.o 00:03:25.431 LINK jsoncat 00:03:25.431 LINK histogram_perf 00:03:25.431 CXX test/cpp_headers/nvmf_transport.o 00:03:25.431 CXX test/cpp_headers/opal.o 00:03:25.431 CXX test/cpp_headers/opal_spec.o 00:03:25.431 LINK ioat_perf 00:03:25.431 CXX test/cpp_headers/pci_ids.o 00:03:25.431 CXX test/cpp_headers/pipe.o 00:03:25.431 CXX test/cpp_headers/queue.o 00:03:25.431 CXX test/cpp_headers/reduce.o 00:03:25.431 LINK stub 00:03:25.431 CXX test/cpp_headers/scheduler.o 00:03:25.431 CXX test/cpp_headers/rpc.o 00:03:25.692 CXX test/cpp_headers/scsi.o 00:03:25.692 LINK spdk_tgt 00:03:25.692 CXX test/cpp_headers/scsi_spec.o 00:03:25.692 CXX test/cpp_headers/sock.o 00:03:25.692 LINK verify 00:03:25.692 CXX test/cpp_headers/stdinc.o 00:03:25.692 CXX test/cpp_headers/string.o 00:03:25.692 CXX test/cpp_headers/thread.o 00:03:25.692 CXX test/cpp_headers/trace.o 00:03:25.692 CXX test/cpp_headers/trace_parser.o 00:03:25.692 CXX test/cpp_headers/tree.o 00:03:25.692 CC test/app/fuzz/iscsi_fuzz/iscsi_fuzz.o 00:03:25.692 CXX test/cpp_headers/ublk.o 00:03:25.692 CXX test/cpp_headers/util.o 00:03:25.692 CXX test/cpp_headers/uuid.o 00:03:25.692 CC test/app/fuzz/vhost_fuzz/vhost_fuzz_rpc.o 00:03:25.692 CXX test/cpp_headers/version.o 00:03:25.692 CXX test/cpp_headers/vfio_user_pci.o 00:03:25.692 CC test/app/fuzz/vhost_fuzz/vhost_fuzz.o 00:03:25.692 CXX test/cpp_headers/vfio_user_spec.o 00:03:25.692 LINK bdev_svc 00:03:25.692 CXX test/cpp_headers/vhost.o 00:03:25.692 CXX test/cpp_headers/vmd.o 00:03:25.692 CXX test/cpp_headers/xor.o 00:03:25.692 CXX test/cpp_headers/zipf.o 00:03:25.692 LINK spdk_dd 00:03:25.692 LINK env_dpdk_post_init 00:03:25.949 LINK spdk_trace 00:03:25.949 LINK pci_ut 00:03:25.949 LINK test_dma 00:03:25.949 LINK nvme_fuzz 00:03:25.949 CC test/event/reactor/reactor.o 00:03:25.949 CC test/event/reactor_perf/reactor_perf.o 00:03:25.949 CC test/event/event_perf/event_perf.o 00:03:25.949 CC test/event/app_repeat/app_repeat.o 00:03:26.207 CC examples/vmd/lsvmd/lsvmd.o 00:03:26.207 CC examples/vmd/led/led.o 00:03:26.207 CC examples/idxd/perf/perf.o 00:03:26.207 CC test/event/scheduler/scheduler.o 00:03:26.207 CC examples/sock/hello_world/hello_sock.o 00:03:26.207 LINK spdk_bdev 00:03:26.207 LINK spdk_nvme 00:03:26.207 CC examples/thread/thread/thread_ex.o 00:03:26.207 LINK spdk_nvme_perf 00:03:26.207 LINK reactor 00:03:26.207 LINK event_perf 00:03:26.207 LINK reactor_perf 00:03:26.207 LINK spdk_nvme_identify 00:03:26.207 LINK app_repeat 00:03:26.207 LINK lsvmd 00:03:26.207 LINK led 00:03:26.207 LINK vhost_fuzz 00:03:26.207 LINK spdk_top 00:03:26.207 LINK mem_callbacks 00:03:26.464 LINK hello_sock 00:03:26.464 CC app/vhost/vhost.o 00:03:26.464 LINK scheduler 00:03:26.464 LINK memory_ut 00:03:26.464 LINK idxd_perf 00:03:26.464 LINK thread 00:03:26.464 CC test/nvme/aer/aer.o 00:03:26.464 CC test/nvme/reset/reset.o 00:03:26.464 CC test/nvme/sgl/sgl.o 00:03:26.464 CC test/nvme/boot_partition/boot_partition.o 00:03:26.464 CC test/nvme/e2edp/nvme_dp.o 00:03:26.464 CC test/nvme/connect_stress/connect_stress.o 00:03:26.464 CC test/nvme/startup/startup.o 00:03:26.464 CC test/nvme/doorbell_aers/doorbell_aers.o 00:03:26.464 CC test/nvme/overhead/overhead.o 00:03:26.464 CC test/nvme/simple_copy/simple_copy.o 00:03:26.464 CC test/nvme/err_injection/err_injection.o 00:03:26.464 CC test/nvme/reserve/reserve.o 00:03:26.464 CC test/nvme/compliance/nvme_compliance.o 00:03:26.464 CC test/nvme/fused_ordering/fused_ordering.o 00:03:26.464 CC test/nvme/fdp/fdp.o 00:03:26.464 CC test/nvme/cuse/cuse.o 00:03:26.464 CC test/accel/dif/dif.o 00:03:26.464 CC test/blobfs/mkfs/mkfs.o 00:03:26.722 LINK vhost 00:03:26.722 CC test/lvol/esnap/esnap.o 00:03:26.722 LINK startup 00:03:26.722 LINK connect_stress 00:03:26.722 LINK boot_partition 00:03:26.722 LINK doorbell_aers 00:03:26.722 LINK err_injection 00:03:26.722 LINK fused_ordering 00:03:26.722 LINK reserve 00:03:26.722 LINK mkfs 00:03:26.722 LINK simple_copy 00:03:26.722 LINK sgl 00:03:26.722 LINK reset 00:03:26.722 LINK nvme_dp 00:03:26.722 CC examples/nvme/cmb_copy/cmb_copy.o 00:03:26.722 CC examples/nvme/hello_world/hello_world.o 00:03:26.722 LINK aer 00:03:26.722 CC examples/nvme/hotplug/hotplug.o 00:03:26.722 CC examples/nvme/abort/abort.o 00:03:26.722 CC examples/nvme/nvme_manage/nvme_manage.o 00:03:26.722 CC examples/nvme/reconnect/reconnect.o 00:03:26.722 CC examples/nvme/arbitration/arbitration.o 00:03:26.979 CC examples/nvme/pmr_persistence/pmr_persistence.o 00:03:26.979 LINK overhead 00:03:26.979 LINK nvme_compliance 00:03:26.979 LINK fdp 00:03:26.979 LINK dif 00:03:26.979 LINK hello_world 00:03:26.979 LINK pmr_persistence 00:03:26.979 LINK cmb_copy 00:03:26.979 LINK hotplug 00:03:26.979 CC examples/accel/perf/accel_perf.o 00:03:27.238 CC examples/blob/hello_world/hello_blob.o 00:03:27.238 CC examples/blob/cli/blobcli.o 00:03:27.238 LINK reconnect 00:03:27.238 LINK arbitration 00:03:27.238 LINK abort 00:03:27.238 LINK hello_blob 00:03:27.496 LINK nvme_manage 00:03:27.496 LINK accel_perf 00:03:27.754 LINK iscsi_fuzz 00:03:27.754 LINK blobcli 00:03:27.754 CC test/bdev/bdevio/bdevio.o 00:03:28.011 LINK cuse 00:03:28.012 LINK bdevio 00:03:28.578 CC examples/bdev/hello_world/hello_bdev.o 00:03:28.578 CC examples/bdev/bdevperf/bdevperf.o 00:03:28.578 LINK hello_bdev 00:03:29.145 LINK bdevperf 00:03:30.080 CC examples/nvmf/nvmf/nvmf.o 00:03:30.337 LINK nvmf 00:03:30.337 LINK esnap 00:03:30.903 00:03:30.903 real 1m27.651s 00:03:30.903 user 17m17.595s 00:03:30.903 sys 4m14.497s 00:03:30.903 11:14:14 make -- common/autotest_common.sh@1124 -- $ xtrace_disable 00:03:30.903 11:14:14 make -- common/autotest_common.sh@10 -- $ set +x 00:03:30.903 ************************************ 00:03:30.903 END TEST make 00:03:30.903 ************************************ 00:03:30.903 11:14:14 -- common/autotest_common.sh@1142 -- $ return 0 00:03:30.903 11:14:14 -- spdk/autobuild.sh@1 -- $ stop_monitor_resources 00:03:30.903 11:14:14 -- pm/common@29 -- $ signal_monitor_resources TERM 00:03:30.903 11:14:14 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:03:30.903 11:14:14 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:03:30.904 11:14:14 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/collect-cpu-load.pid ]] 00:03:30.904 11:14:14 -- pm/common@44 -- $ pid=706240 00:03:30.904 11:14:14 -- pm/common@50 -- $ kill -TERM 706240 00:03:30.904 11:14:14 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:03:30.904 11:14:14 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/collect-vmstat.pid ]] 00:03:30.904 11:14:14 -- pm/common@44 -- $ pid=706242 00:03:30.904 11:14:14 -- pm/common@50 -- $ kill -TERM 706242 00:03:30.904 11:14:14 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:03:30.904 11:14:14 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/collect-cpu-temp.pid ]] 00:03:30.904 11:14:14 -- pm/common@44 -- $ pid=706244 00:03:30.904 11:14:14 -- pm/common@50 -- $ kill -TERM 706244 00:03:30.904 11:14:14 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:03:30.904 11:14:14 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/collect-bmc-pm.pid ]] 00:03:30.904 11:14:14 -- pm/common@44 -- $ pid=706270 00:03:30.904 11:14:14 -- pm/common@50 -- $ sudo -E kill -TERM 706270 00:03:30.904 11:14:14 -- spdk/autotest.sh@25 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/nvmf/common.sh 00:03:30.904 11:14:14 -- nvmf/common.sh@7 -- # uname -s 00:03:30.904 11:14:14 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:03:30.904 11:14:14 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:03:30.904 11:14:14 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:03:30.904 11:14:14 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:03:30.904 11:14:14 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:03:30.904 11:14:14 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:03:30.904 11:14:14 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:03:30.904 11:14:14 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:03:30.904 11:14:14 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:03:30.904 11:14:14 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:03:30.904 11:14:14 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:809e2efd-7f71-e711-906e-0017a4403562 00:03:30.904 11:14:14 -- nvmf/common.sh@18 -- # NVME_HOSTID=809e2efd-7f71-e711-906e-0017a4403562 00:03:30.904 11:14:14 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:03:30.904 11:14:14 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:03:30.904 11:14:14 -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:03:30.904 11:14:14 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:03:30.904 11:14:14 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/common.sh 00:03:30.904 11:14:14 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:03:30.904 11:14:14 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:03:30.904 11:14:14 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:03:30.904 11:14:14 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:03:30.904 11:14:14 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:03:30.904 11:14:14 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:03:30.904 11:14:14 -- paths/export.sh@5 -- # export PATH 00:03:30.904 11:14:14 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:03:30.904 11:14:14 -- nvmf/common.sh@47 -- # : 0 00:03:30.904 11:14:14 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:03:30.904 11:14:14 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:03:30.904 11:14:14 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:03:30.904 11:14:14 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:03:30.904 11:14:14 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:03:30.904 11:14:14 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:03:30.904 11:14:14 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:03:30.904 11:14:14 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:03:30.904 11:14:14 -- spdk/autotest.sh@27 -- # '[' 0 -ne 0 ']' 00:03:30.904 11:14:14 -- spdk/autotest.sh@32 -- # uname -s 00:03:30.904 11:14:14 -- spdk/autotest.sh@32 -- # '[' Linux = Linux ']' 00:03:30.904 11:14:14 -- spdk/autotest.sh@33 -- # old_core_pattern='|/usr/lib/systemd/systemd-coredump %P %u %g %s %t %c %h' 00:03:30.904 11:14:14 -- spdk/autotest.sh@34 -- # mkdir -p /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/coredumps 00:03:30.904 11:14:14 -- spdk/autotest.sh@39 -- # echo '|/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/core-collector.sh %P %s %t' 00:03:30.904 11:14:14 -- spdk/autotest.sh@40 -- # echo /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/coredumps 00:03:30.904 11:14:14 -- spdk/autotest.sh@44 -- # modprobe nbd 00:03:30.904 11:14:14 -- spdk/autotest.sh@46 -- # type -P udevadm 00:03:31.162 11:14:14 -- spdk/autotest.sh@46 -- # udevadm=/usr/sbin/udevadm 00:03:31.162 11:14:14 -- spdk/autotest.sh@48 -- # udevadm_pid=772201 00:03:31.162 11:14:14 -- spdk/autotest.sh@47 -- # /usr/sbin/udevadm monitor --property 00:03:31.162 11:14:14 -- spdk/autotest.sh@53 -- # start_monitor_resources 00:03:31.162 11:14:14 -- pm/common@17 -- # local monitor 00:03:31.162 11:14:14 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:03:31.162 11:14:14 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:03:31.162 11:14:14 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:03:31.162 11:14:14 -- pm/common@21 -- # date +%s 00:03:31.162 11:14:14 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:03:31.162 11:14:14 -- pm/common@21 -- # date +%s 00:03:31.162 11:14:14 -- pm/common@25 -- # sleep 1 00:03:31.162 11:14:14 -- pm/common@21 -- # date +%s 00:03:31.162 11:14:14 -- pm/common@21 -- # date +%s 00:03:31.162 11:14:14 -- pm/common@21 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm/collect-cpu-load -d /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1721034854 00:03:31.162 11:14:14 -- pm/common@21 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm/collect-vmstat -d /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1721034854 00:03:31.162 11:14:14 -- pm/common@21 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm/collect-cpu-temp -d /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1721034854 00:03:31.162 11:14:14 -- pm/common@21 -- # sudo -E /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm/collect-bmc-pm -d /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1721034854 00:03:31.162 Redirecting to /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/monitor.autotest.sh.1721034854_collect-vmstat.pm.log 00:03:31.162 Redirecting to /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/monitor.autotest.sh.1721034854_collect-cpu-load.pm.log 00:03:31.162 Redirecting to /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/monitor.autotest.sh.1721034854_collect-cpu-temp.pm.log 00:03:31.162 Redirecting to /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/monitor.autotest.sh.1721034854_collect-bmc-pm.bmc.pm.log 00:03:32.097 11:14:15 -- spdk/autotest.sh@55 -- # trap 'autotest_cleanup || :; exit 1' SIGINT SIGTERM EXIT 00:03:32.097 11:14:15 -- spdk/autotest.sh@57 -- # timing_enter autotest 00:03:32.097 11:14:15 -- common/autotest_common.sh@722 -- # xtrace_disable 00:03:32.097 11:14:15 -- common/autotest_common.sh@10 -- # set +x 00:03:32.097 11:14:15 -- spdk/autotest.sh@59 -- # create_test_list 00:03:32.097 11:14:15 -- common/autotest_common.sh@746 -- # xtrace_disable 00:03:32.097 11:14:15 -- common/autotest_common.sh@10 -- # set +x 00:03:32.097 11:14:15 -- spdk/autotest.sh@61 -- # dirname /var/jenkins/workspace/crypto-phy-autotest/spdk/autotest.sh 00:03:32.097 11:14:15 -- spdk/autotest.sh@61 -- # readlink -f /var/jenkins/workspace/crypto-phy-autotest/spdk 00:03:32.097 11:14:15 -- spdk/autotest.sh@61 -- # src=/var/jenkins/workspace/crypto-phy-autotest/spdk 00:03:32.097 11:14:15 -- spdk/autotest.sh@62 -- # out=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:03:32.097 11:14:15 -- spdk/autotest.sh@63 -- # cd /var/jenkins/workspace/crypto-phy-autotest/spdk 00:03:32.097 11:14:15 -- spdk/autotest.sh@65 -- # freebsd_update_contigmem_mod 00:03:32.097 11:14:15 -- common/autotest_common.sh@1455 -- # uname 00:03:32.097 11:14:15 -- common/autotest_common.sh@1455 -- # '[' Linux = FreeBSD ']' 00:03:32.097 11:14:15 -- spdk/autotest.sh@66 -- # freebsd_set_maxsock_buf 00:03:32.097 11:14:15 -- common/autotest_common.sh@1475 -- # uname 00:03:32.097 11:14:15 -- common/autotest_common.sh@1475 -- # [[ Linux = FreeBSD ]] 00:03:32.097 11:14:15 -- spdk/autotest.sh@71 -- # grep CC_TYPE mk/cc.mk 00:03:32.097 11:14:15 -- spdk/autotest.sh@71 -- # CC_TYPE=CC_TYPE=gcc 00:03:32.097 11:14:15 -- spdk/autotest.sh@72 -- # hash lcov 00:03:32.097 11:14:15 -- spdk/autotest.sh@72 -- # [[ CC_TYPE=gcc == *\c\l\a\n\g* ]] 00:03:32.097 11:14:15 -- spdk/autotest.sh@80 -- # export 'LCOV_OPTS= 00:03:32.097 --rc lcov_branch_coverage=1 00:03:32.097 --rc lcov_function_coverage=1 00:03:32.097 --rc genhtml_branch_coverage=1 00:03:32.097 --rc genhtml_function_coverage=1 00:03:32.097 --rc genhtml_legend=1 00:03:32.097 --rc geninfo_all_blocks=1 00:03:32.097 ' 00:03:32.097 11:14:15 -- spdk/autotest.sh@80 -- # LCOV_OPTS=' 00:03:32.097 --rc lcov_branch_coverage=1 00:03:32.097 --rc lcov_function_coverage=1 00:03:32.097 --rc genhtml_branch_coverage=1 00:03:32.097 --rc genhtml_function_coverage=1 00:03:32.097 --rc genhtml_legend=1 00:03:32.097 --rc geninfo_all_blocks=1 00:03:32.097 ' 00:03:32.097 11:14:15 -- spdk/autotest.sh@81 -- # export 'LCOV=lcov 00:03:32.097 --rc lcov_branch_coverage=1 00:03:32.097 --rc lcov_function_coverage=1 00:03:32.097 --rc genhtml_branch_coverage=1 00:03:32.097 --rc genhtml_function_coverage=1 00:03:32.097 --rc genhtml_legend=1 00:03:32.097 --rc geninfo_all_blocks=1 00:03:32.097 --no-external' 00:03:32.097 11:14:15 -- spdk/autotest.sh@81 -- # LCOV='lcov 00:03:32.097 --rc lcov_branch_coverage=1 00:03:32.097 --rc lcov_function_coverage=1 00:03:32.097 --rc genhtml_branch_coverage=1 00:03:32.097 --rc genhtml_function_coverage=1 00:03:32.097 --rc genhtml_legend=1 00:03:32.097 --rc geninfo_all_blocks=1 00:03:32.097 --no-external' 00:03:32.097 11:14:15 -- spdk/autotest.sh@83 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -v 00:03:32.097 lcov: LCOV version 1.14 00:03:32.097 11:14:15 -- spdk/autotest.sh@85 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -c -i -t Baseline -d /var/jenkins/workspace/crypto-phy-autotest/spdk -o /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/cov_base.info 00:03:50.182 /var/jenkins/workspace/crypto-phy-autotest/spdk/lib/nvme/nvme_stubs.gcno:no functions found 00:03:50.182 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/lib/nvme/nvme_stubs.gcno 00:04:02.386 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/accel.gcno:no functions found 00:04:02.386 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/accel.gcno 00:04:02.386 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/accel_module.gcno:no functions found 00:04:02.386 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/accel_module.gcno 00:04:02.386 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/barrier.gcno:no functions found 00:04:02.386 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/barrier.gcno 00:04:02.386 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/assert.gcno:no functions found 00:04:02.386 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/assert.gcno 00:04:02.386 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/base64.gcno:no functions found 00:04:02.386 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/base64.gcno 00:04:02.386 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/bdev_zone.gcno:no functions found 00:04:02.386 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/bdev_zone.gcno 00:04:02.386 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/blobfs.gcno:no functions found 00:04:02.386 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/blobfs.gcno 00:04:02.386 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/bdev.gcno:no functions found 00:04:02.386 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/bdev.gcno 00:04:02.386 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/bdev_module.gcno:no functions found 00:04:02.386 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/bdev_module.gcno 00:04:02.386 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/bit_array.gcno:no functions found 00:04:02.386 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/bit_array.gcno 00:04:02.386 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/bit_pool.gcno:no functions found 00:04:02.386 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/bit_pool.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/blob.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/blob.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/config.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/config.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/crc32.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/crc32.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/blobfs_bdev.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/blobfs_bdev.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/conf.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/conf.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/crc64.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/crc64.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/blob_bdev.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/blob_bdev.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/crc16.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/crc16.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/cpuset.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/cpuset.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/endian.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/endian.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/dif.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/dif.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/env_dpdk.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/env_dpdk.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/env.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/env.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/fd.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/fd.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/event.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/event.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/dma.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/dma.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/ftl.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/ftl.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/file.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/file.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/fd_group.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/fd_group.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/histogram_data.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/histogram_data.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/hexlify.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/hexlify.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/idxd_spec.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/idxd_spec.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/idxd.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/idxd.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/ioat_spec.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/ioat_spec.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/init.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/init.gcno 00:04:02.644 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/ioat.gcno:no functions found 00:04:02.644 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/ioat.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/jsonrpc.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/jsonrpc.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/json.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/json.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/keyring.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/keyring.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/iscsi_spec.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/iscsi_spec.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/gpt_spec.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/gpt_spec.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/keyring_module.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/keyring_module.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/likely.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/likely.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/log.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/log.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/lvol.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/lvol.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/memory.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/memory.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/mmio.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/mmio.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nbd.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nbd.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/notify.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/notify.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvme.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvme.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvme_intel.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvme_intel.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvme_ocssd.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvme_ocssd.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvme_spec.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvme_spec.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvme_ocssd_spec.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvme_ocssd_spec.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvme_zns.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvme_zns.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvmf_cmd.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvmf_cmd.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvmf_fc_spec.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvmf_fc_spec.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvmf.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvmf.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvmf_spec.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvmf_spec.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/opal.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/opal.gcno 00:04:02.903 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvmf_transport.gcno:no functions found 00:04:02.903 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/nvmf_transport.gcno 00:04:03.162 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/opal_spec.gcno:no functions found 00:04:03.162 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/opal_spec.gcno 00:04:03.162 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/pipe.gcno:no functions found 00:04:03.162 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/pipe.gcno 00:04:03.162 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/pci_ids.gcno:no functions found 00:04:03.162 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/pci_ids.gcno 00:04:03.162 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/queue.gcno:no functions found 00:04:03.162 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/queue.gcno 00:04:03.162 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/reduce.gcno:no functions found 00:04:03.162 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/reduce.gcno 00:04:03.162 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/scheduler.gcno:no functions found 00:04:03.162 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/scheduler.gcno 00:04:03.162 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/rpc.gcno:no functions found 00:04:03.162 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/rpc.gcno 00:04:03.162 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/scsi.gcno:no functions found 00:04:03.162 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/scsi.gcno 00:04:03.162 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/scsi_spec.gcno:no functions found 00:04:03.162 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/scsi_spec.gcno 00:04:03.162 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/sock.gcno:no functions found 00:04:03.162 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/sock.gcno 00:04:03.162 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/stdinc.gcno:no functions found 00:04:03.162 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/stdinc.gcno 00:04:03.162 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/string.gcno:no functions found 00:04:03.162 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/string.gcno 00:04:03.162 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/thread.gcno:no functions found 00:04:03.162 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/thread.gcno 00:04:03.162 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/tree.gcno:no functions found 00:04:03.162 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/tree.gcno 00:04:03.162 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/trace.gcno:no functions found 00:04:03.162 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/trace.gcno 00:04:03.162 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/trace_parser.gcno:no functions found 00:04:03.162 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/trace_parser.gcno 00:04:03.162 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/ublk.gcno:no functions found 00:04:03.162 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/ublk.gcno 00:04:03.162 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/util.gcno:no functions found 00:04:03.162 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/util.gcno 00:04:03.162 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/uuid.gcno:no functions found 00:04:03.163 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/uuid.gcno 00:04:03.163 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/version.gcno:no functions found 00:04:03.163 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/version.gcno 00:04:03.163 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/vfio_user_pci.gcno:no functions found 00:04:03.163 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/vfio_user_pci.gcno 00:04:03.163 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/vfio_user_spec.gcno:no functions found 00:04:03.163 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/vfio_user_spec.gcno 00:04:03.163 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/vhost.gcno:no functions found 00:04:03.163 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/vhost.gcno 00:04:03.163 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/vmd.gcno:no functions found 00:04:03.163 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/vmd.gcno 00:04:03.163 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/xor.gcno:no functions found 00:04:03.163 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/xor.gcno 00:04:03.163 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/zipf.gcno:no functions found 00:04:03.163 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/crypto-phy-autotest/spdk/test/cpp_headers/zipf.gcno 00:04:07.340 11:14:50 -- spdk/autotest.sh@89 -- # timing_enter pre_cleanup 00:04:07.340 11:14:50 -- common/autotest_common.sh@722 -- # xtrace_disable 00:04:07.340 11:14:50 -- common/autotest_common.sh@10 -- # set +x 00:04:07.340 11:14:50 -- spdk/autotest.sh@91 -- # rm -f 00:04:07.340 11:14:50 -- spdk/autotest.sh@94 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh reset 00:04:10.616 0000:d7:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:d7:05.5 00:04:10.616 0000:85:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:85:05.5 00:04:10.616 0000:5e:00.0 (8086 0b60): Already using the nvme driver 00:04:10.616 0000:00:04.7 (8086 2021): Already using the ioatdma driver 00:04:10.616 0000:00:04.6 (8086 2021): Already using the ioatdma driver 00:04:10.616 0000:00:04.5 (8086 2021): Already using the ioatdma driver 00:04:10.616 0000:00:04.4 (8086 2021): Already using the ioatdma driver 00:04:10.616 0000:00:04.3 (8086 2021): Already using the ioatdma driver 00:04:10.616 0000:00:04.2 (8086 2021): Already using the ioatdma driver 00:04:10.616 0000:00:04.1 (8086 2021): Already using the ioatdma driver 00:04:10.616 0000:00:04.0 (8086 2021): Already using the ioatdma driver 00:04:10.873 0000:80:04.7 (8086 2021): Already using the ioatdma driver 00:04:10.873 0000:80:04.6 (8086 2021): Already using the ioatdma driver 00:04:10.873 0000:80:04.5 (8086 2021): Already using the ioatdma driver 00:04:10.873 0000:80:04.4 (8086 2021): Already using the ioatdma driver 00:04:10.873 0000:80:04.3 (8086 2021): Already using the ioatdma driver 00:04:10.873 0000:80:04.2 (8086 2021): Already using the ioatdma driver 00:04:10.873 0000:80:04.1 (8086 2021): Already using the ioatdma driver 00:04:10.873 0000:80:04.0 (8086 2021): Already using the ioatdma driver 00:04:10.873 11:14:54 -- spdk/autotest.sh@96 -- # get_zoned_devs 00:04:10.873 11:14:54 -- common/autotest_common.sh@1669 -- # zoned_devs=() 00:04:10.873 11:14:54 -- common/autotest_common.sh@1669 -- # local -gA zoned_devs 00:04:10.873 11:14:54 -- common/autotest_common.sh@1670 -- # local nvme bdf 00:04:10.873 11:14:54 -- common/autotest_common.sh@1672 -- # for nvme in /sys/block/nvme* 00:04:10.873 11:14:54 -- common/autotest_common.sh@1673 -- # is_block_zoned nvme0n1 00:04:10.873 11:14:54 -- common/autotest_common.sh@1662 -- # local device=nvme0n1 00:04:10.873 11:14:54 -- common/autotest_common.sh@1664 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:04:10.873 11:14:54 -- common/autotest_common.sh@1665 -- # [[ none != none ]] 00:04:10.873 11:14:54 -- spdk/autotest.sh@98 -- # (( 0 > 0 )) 00:04:10.873 11:14:54 -- spdk/autotest.sh@110 -- # for dev in /dev/nvme*n!(*p*) 00:04:11.130 11:14:54 -- spdk/autotest.sh@112 -- # [[ -z '' ]] 00:04:11.130 11:14:54 -- spdk/autotest.sh@113 -- # block_in_use /dev/nvme0n1 00:04:11.130 11:14:54 -- scripts/common.sh@378 -- # local block=/dev/nvme0n1 pt 00:04:11.130 11:14:54 -- scripts/common.sh@387 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/spdk-gpt.py /dev/nvme0n1 00:04:11.130 No valid GPT data, bailing 00:04:11.130 11:14:54 -- scripts/common.sh@391 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:04:11.130 11:14:54 -- scripts/common.sh@391 -- # pt= 00:04:11.130 11:14:54 -- scripts/common.sh@392 -- # return 1 00:04:11.130 11:14:54 -- spdk/autotest.sh@114 -- # dd if=/dev/zero of=/dev/nvme0n1 bs=1M count=1 00:04:11.130 1+0 records in 00:04:11.130 1+0 records out 00:04:11.130 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.004828 s, 217 MB/s 00:04:11.130 11:14:54 -- spdk/autotest.sh@118 -- # sync 00:04:11.130 11:14:54 -- spdk/autotest.sh@120 -- # xtrace_disable_per_cmd reap_spdk_processes 00:04:11.130 11:14:54 -- common/autotest_common.sh@22 -- # eval 'reap_spdk_processes 12> /dev/null' 00:04:11.130 11:14:54 -- common/autotest_common.sh@22 -- # reap_spdk_processes 00:04:16.391 11:14:59 -- spdk/autotest.sh@124 -- # uname -s 00:04:16.391 11:14:59 -- spdk/autotest.sh@124 -- # '[' Linux = Linux ']' 00:04:16.391 11:14:59 -- spdk/autotest.sh@125 -- # run_test setup.sh /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/test-setup.sh 00:04:16.391 11:14:59 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:16.391 11:14:59 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:16.391 11:14:59 -- common/autotest_common.sh@10 -- # set +x 00:04:16.391 ************************************ 00:04:16.391 START TEST setup.sh 00:04:16.391 ************************************ 00:04:16.391 11:14:59 setup.sh -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/test-setup.sh 00:04:16.391 * Looking for test storage... 00:04:16.391 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup 00:04:16.391 11:14:59 setup.sh -- setup/test-setup.sh@10 -- # uname -s 00:04:16.391 11:14:59 setup.sh -- setup/test-setup.sh@10 -- # [[ Linux == Linux ]] 00:04:16.391 11:14:59 setup.sh -- setup/test-setup.sh@12 -- # run_test acl /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/acl.sh 00:04:16.392 11:14:59 setup.sh -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:16.392 11:14:59 setup.sh -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:16.392 11:14:59 setup.sh -- common/autotest_common.sh@10 -- # set +x 00:04:16.392 ************************************ 00:04:16.392 START TEST acl 00:04:16.392 ************************************ 00:04:16.392 11:14:59 setup.sh.acl -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/acl.sh 00:04:16.392 * Looking for test storage... 00:04:16.392 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup 00:04:16.392 11:14:59 setup.sh.acl -- setup/acl.sh@10 -- # get_zoned_devs 00:04:16.392 11:14:59 setup.sh.acl -- common/autotest_common.sh@1669 -- # zoned_devs=() 00:04:16.392 11:14:59 setup.sh.acl -- common/autotest_common.sh@1669 -- # local -gA zoned_devs 00:04:16.392 11:14:59 setup.sh.acl -- common/autotest_common.sh@1670 -- # local nvme bdf 00:04:16.392 11:14:59 setup.sh.acl -- common/autotest_common.sh@1672 -- # for nvme in /sys/block/nvme* 00:04:16.392 11:14:59 setup.sh.acl -- common/autotest_common.sh@1673 -- # is_block_zoned nvme0n1 00:04:16.392 11:14:59 setup.sh.acl -- common/autotest_common.sh@1662 -- # local device=nvme0n1 00:04:16.392 11:14:59 setup.sh.acl -- common/autotest_common.sh@1664 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:04:16.392 11:14:59 setup.sh.acl -- common/autotest_common.sh@1665 -- # [[ none != none ]] 00:04:16.392 11:14:59 setup.sh.acl -- setup/acl.sh@12 -- # devs=() 00:04:16.392 11:14:59 setup.sh.acl -- setup/acl.sh@12 -- # declare -a devs 00:04:16.392 11:14:59 setup.sh.acl -- setup/acl.sh@13 -- # drivers=() 00:04:16.392 11:14:59 setup.sh.acl -- setup/acl.sh@13 -- # declare -A drivers 00:04:16.392 11:14:59 setup.sh.acl -- setup/acl.sh@51 -- # setup reset 00:04:16.392 11:14:59 setup.sh.acl -- setup/common.sh@9 -- # [[ reset == output ]] 00:04:16.392 11:14:59 setup.sh.acl -- setup/common.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh reset 00:04:20.580 11:15:03 setup.sh.acl -- setup/acl.sh@52 -- # collect_setup_devs 00:04:20.580 11:15:03 setup.sh.acl -- setup/acl.sh@16 -- # local dev driver 00:04:20.580 11:15:03 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:20.580 11:15:03 setup.sh.acl -- setup/acl.sh@15 -- # setup output status 00:04:20.580 11:15:03 setup.sh.acl -- setup/common.sh@9 -- # [[ output == output ]] 00:04:20.580 11:15:03 setup.sh.acl -- setup/common.sh@10 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh status 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ (8086 == *:*:*.* ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # continue 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ (8086 == *:*:*.* ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # continue 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:23.944 Hugepages 00:04:23.944 node hugesize free / total 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 1048576kB == *:*:*.* ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # continue 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 2048kB == *:*:*.* ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # continue 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 1048576kB == *:*:*.* ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # continue 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:23.944 00:04:23.944 Type BDF Vendor Device NUMA Driver Device Block devices 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 2048kB == *:*:*.* ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # continue 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:00:04.0 == *:*:*.* ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:00:04.1 == *:*:*.* ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:00:04.2 == *:*:*.* ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:00:04.3 == *:*:*.* ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:00:04.4 == *:*:*.* ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:00:04.5 == *:*:*.* ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:00:04.6 == *:*:*.* ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:00:04.7 == *:*:*.* ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:04:23.944 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:5e:00.0 == *:*:*.* ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # [[ nvme == nvme ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@21 -- # [[ '' == *\0\0\0\0\:\5\e\:\0\0\.\0* ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@22 -- # devs+=("$dev") 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@22 -- # drivers["$dev"]=nvme 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:80:04.0 == *:*:*.* ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:80:04.1 == *:*:*.* ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:80:04.2 == *:*:*.* ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:80:04.3 == *:*:*.* ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:80:04.4 == *:*:*.* ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:80:04.5 == *:*:*.* ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:80:04.6 == *:*:*.* ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:80:04.7 == *:*:*.* ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:85:05.5 == *:*:*.* ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # [[ vfio-pci == nvme ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:d7:05.5 == *:*:*.* ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # [[ vfio-pci == nvme ]] 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@24 -- # (( 1 > 0 )) 00:04:24.203 11:15:07 setup.sh.acl -- setup/acl.sh@54 -- # run_test denied denied 00:04:24.203 11:15:07 setup.sh.acl -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:24.203 11:15:07 setup.sh.acl -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:24.203 11:15:07 setup.sh.acl -- common/autotest_common.sh@10 -- # set +x 00:04:24.203 ************************************ 00:04:24.203 START TEST denied 00:04:24.203 ************************************ 00:04:24.203 11:15:07 setup.sh.acl.denied -- common/autotest_common.sh@1123 -- # denied 00:04:24.203 11:15:07 setup.sh.acl.denied -- setup/acl.sh@38 -- # PCI_BLOCKED=' 0000:5e:00.0' 00:04:24.203 11:15:07 setup.sh.acl.denied -- setup/acl.sh@38 -- # setup output config 00:04:24.203 11:15:07 setup.sh.acl.denied -- setup/acl.sh@39 -- # grep 'Skipping denied controller at 0000:5e:00.0' 00:04:24.203 11:15:07 setup.sh.acl.denied -- setup/common.sh@9 -- # [[ output == output ]] 00:04:24.203 11:15:07 setup.sh.acl.denied -- setup/common.sh@10 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh config 00:04:28.391 0000:5e:00.0 (8086 0b60): Skipping denied controller at 0000:5e:00.0 00:04:28.391 11:15:11 setup.sh.acl.denied -- setup/acl.sh@40 -- # verify 0000:5e:00.0 00:04:28.391 11:15:11 setup.sh.acl.denied -- setup/acl.sh@28 -- # local dev driver 00:04:28.391 11:15:11 setup.sh.acl.denied -- setup/acl.sh@30 -- # for dev in "$@" 00:04:28.391 11:15:11 setup.sh.acl.denied -- setup/acl.sh@31 -- # [[ -e /sys/bus/pci/devices/0000:5e:00.0 ]] 00:04:28.391 11:15:11 setup.sh.acl.denied -- setup/acl.sh@32 -- # readlink -f /sys/bus/pci/devices/0000:5e:00.0/driver 00:04:28.391 11:15:11 setup.sh.acl.denied -- setup/acl.sh@32 -- # driver=/sys/bus/pci/drivers/nvme 00:04:28.391 11:15:11 setup.sh.acl.denied -- setup/acl.sh@33 -- # [[ nvme == \n\v\m\e ]] 00:04:28.391 11:15:11 setup.sh.acl.denied -- setup/acl.sh@41 -- # setup reset 00:04:28.391 11:15:11 setup.sh.acl.denied -- setup/common.sh@9 -- # [[ reset == output ]] 00:04:28.391 11:15:11 setup.sh.acl.denied -- setup/common.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh reset 00:04:33.658 00:04:33.658 real 0m9.280s 00:04:33.658 user 0m3.105s 00:04:33.658 sys 0m5.497s 00:04:33.658 11:15:16 setup.sh.acl.denied -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:33.658 11:15:16 setup.sh.acl.denied -- common/autotest_common.sh@10 -- # set +x 00:04:33.658 ************************************ 00:04:33.658 END TEST denied 00:04:33.658 ************************************ 00:04:33.658 11:15:17 setup.sh.acl -- common/autotest_common.sh@1142 -- # return 0 00:04:33.658 11:15:17 setup.sh.acl -- setup/acl.sh@55 -- # run_test allowed allowed 00:04:33.658 11:15:17 setup.sh.acl -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:33.658 11:15:17 setup.sh.acl -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:33.658 11:15:17 setup.sh.acl -- common/autotest_common.sh@10 -- # set +x 00:04:33.658 ************************************ 00:04:33.658 START TEST allowed 00:04:33.658 ************************************ 00:04:33.658 11:15:17 setup.sh.acl.allowed -- common/autotest_common.sh@1123 -- # allowed 00:04:33.658 11:15:17 setup.sh.acl.allowed -- setup/acl.sh@45 -- # PCI_ALLOWED=0000:5e:00.0 00:04:33.658 11:15:17 setup.sh.acl.allowed -- setup/acl.sh@45 -- # setup output config 00:04:33.658 11:15:17 setup.sh.acl.allowed -- setup/acl.sh@46 -- # grep -E '0000:5e:00.0 .*: nvme -> .*' 00:04:33.658 11:15:17 setup.sh.acl.allowed -- setup/common.sh@9 -- # [[ output == output ]] 00:04:33.658 11:15:17 setup.sh.acl.allowed -- setup/common.sh@10 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh config 00:04:40.222 0000:5e:00.0 (8086 0b60): nvme -> vfio-pci 00:04:40.222 11:15:23 setup.sh.acl.allowed -- setup/acl.sh@47 -- # verify 00:04:40.222 11:15:23 setup.sh.acl.allowed -- setup/acl.sh@28 -- # local dev driver 00:04:40.222 11:15:23 setup.sh.acl.allowed -- setup/acl.sh@48 -- # setup reset 00:04:40.222 11:15:23 setup.sh.acl.allowed -- setup/common.sh@9 -- # [[ reset == output ]] 00:04:40.222 11:15:23 setup.sh.acl.allowed -- setup/common.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh reset 00:04:44.408 00:04:44.408 real 0m10.284s 00:04:44.408 user 0m2.539s 00:04:44.408 sys 0m5.199s 00:04:44.408 11:15:27 setup.sh.acl.allowed -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:44.408 11:15:27 setup.sh.acl.allowed -- common/autotest_common.sh@10 -- # set +x 00:04:44.408 ************************************ 00:04:44.408 END TEST allowed 00:04:44.408 ************************************ 00:04:44.408 11:15:27 setup.sh.acl -- common/autotest_common.sh@1142 -- # return 0 00:04:44.408 00:04:44.408 real 0m27.786s 00:04:44.408 user 0m8.569s 00:04:44.408 sys 0m16.296s 00:04:44.408 11:15:27 setup.sh.acl -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:44.408 11:15:27 setup.sh.acl -- common/autotest_common.sh@10 -- # set +x 00:04:44.408 ************************************ 00:04:44.408 END TEST acl 00:04:44.408 ************************************ 00:04:44.408 11:15:27 setup.sh -- common/autotest_common.sh@1142 -- # return 0 00:04:44.408 11:15:27 setup.sh -- setup/test-setup.sh@13 -- # run_test hugepages /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/hugepages.sh 00:04:44.408 11:15:27 setup.sh -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:44.408 11:15:27 setup.sh -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:44.408 11:15:27 setup.sh -- common/autotest_common.sh@10 -- # set +x 00:04:44.408 ************************************ 00:04:44.408 START TEST hugepages 00:04:44.408 ************************************ 00:04:44.408 11:15:27 setup.sh.hugepages -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/hugepages.sh 00:04:44.408 * Looking for test storage... 00:04:44.408 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup 00:04:44.408 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@10 -- # nodes_sys=() 00:04:44.408 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@10 -- # declare -a nodes_sys 00:04:44.408 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@12 -- # declare -i default_hugepages=0 00:04:44.408 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@13 -- # declare -i no_nodes=0 00:04:44.408 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@14 -- # declare -i nr_hugepages=0 00:04:44.408 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@16 -- # get_meminfo Hugepagesize 00:04:44.408 11:15:27 setup.sh.hugepages -- setup/common.sh@17 -- # local get=Hugepagesize 00:04:44.408 11:15:27 setup.sh.hugepages -- setup/common.sh@18 -- # local node= 00:04:44.408 11:15:27 setup.sh.hugepages -- setup/common.sh@19 -- # local var val 00:04:44.408 11:15:27 setup.sh.hugepages -- setup/common.sh@20 -- # local mem_f mem 00:04:44.408 11:15:27 setup.sh.hugepages -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:04:44.408 11:15:27 setup.sh.hugepages -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:04:44.408 11:15:27 setup.sh.hugepages -- setup/common.sh@25 -- # [[ -n '' ]] 00:04:44.408 11:15:27 setup.sh.hugepages -- setup/common.sh@28 -- # mapfile -t mem 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 76729920 kB' 'MemAvailable: 80029008 kB' 'Buffers: 12176 kB' 'Cached: 9444056 kB' 'SwapCached: 0 kB' 'Active: 6500252 kB' 'Inactive: 3456260 kB' 'Active(anon): 6106668 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 503552 kB' 'Mapped: 168120 kB' 'Shmem: 5606388 kB' 'KReclaimable: 205772 kB' 'Slab: 528812 kB' 'SReclaimable: 205772 kB' 'SUnreclaim: 323040 kB' 'KernelStack: 16096 kB' 'PageTables: 8092 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 52438188 kB' 'Committed_AS: 7529188 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 200840 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 2048' 'HugePages_Free: 2048' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 4194304 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.409 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Hugepagesize == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@33 -- # echo 2048 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/common.sh@33 -- # return 0 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@16 -- # default_hugepages=2048 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@17 -- # default_huge_nr=/sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@18 -- # global_huge_nr=/proc/sys/vm/nr_hugepages 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@21 -- # unset -v HUGE_EVEN_ALLOC 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@22 -- # unset -v HUGEMEM 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@23 -- # unset -v HUGENODE 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@24 -- # unset -v NRHUGE 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@207 -- # get_nodes 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@27 -- # local node 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=1024 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=1024 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@32 -- # no_nodes=2 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@208 -- # clear_hp 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@37 -- # local node hp 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@39 -- # for node in "${!nodes_sys[@]}" 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@41 -- # echo 0 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@41 -- # echo 0 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@39 -- # for node in "${!nodes_sys[@]}" 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@41 -- # echo 0 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@41 -- # echo 0 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@45 -- # export CLEAR_HUGE=yes 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@45 -- # CLEAR_HUGE=yes 00:04:44.410 11:15:27 setup.sh.hugepages -- setup/hugepages.sh@210 -- # run_test default_setup default_setup 00:04:44.410 11:15:27 setup.sh.hugepages -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:44.410 11:15:27 setup.sh.hugepages -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:44.410 11:15:27 setup.sh.hugepages -- common/autotest_common.sh@10 -- # set +x 00:04:44.410 ************************************ 00:04:44.410 START TEST default_setup 00:04:44.410 ************************************ 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- common/autotest_common.sh@1123 -- # default_setup 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/hugepages.sh@136 -- # get_test_nr_hugepages 2097152 0 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/hugepages.sh@49 -- # local size=2097152 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/hugepages.sh@50 -- # (( 2 > 1 )) 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/hugepages.sh@51 -- # shift 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/hugepages.sh@52 -- # node_ids=('0') 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/hugepages.sh@52 -- # local node_ids 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/hugepages.sh@57 -- # nr_hugepages=1024 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 0 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/hugepages.sh@62 -- # user_nodes=('0') 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/hugepages.sh@62 -- # local user_nodes 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/hugepages.sh@64 -- # local _nr_hugepages=1024 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/hugepages.sh@67 -- # nodes_test=() 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/hugepages.sh@67 -- # local -g nodes_test 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/hugepages.sh@69 -- # (( 1 > 0 )) 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/hugepages.sh@70 -- # for _no_nodes in "${user_nodes[@]}" 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/hugepages.sh@71 -- # nodes_test[_no_nodes]=1024 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/hugepages.sh@73 -- # return 0 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/hugepages.sh@137 -- # setup output 00:04:44.410 11:15:27 setup.sh.hugepages.default_setup -- setup/common.sh@9 -- # [[ output == output ]] 00:04:44.411 11:15:27 setup.sh.hugepages.default_setup -- setup/common.sh@10 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh 00:04:47.691 0000:d7:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:d7:05.5 00:04:47.691 0000:85:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:85:05.5 00:04:47.691 0000:00:04.7 (8086 2021): ioatdma -> vfio-pci 00:04:47.691 0000:00:04.6 (8086 2021): ioatdma -> vfio-pci 00:04:47.691 0000:00:04.5 (8086 2021): ioatdma -> vfio-pci 00:04:47.691 0000:00:04.4 (8086 2021): ioatdma -> vfio-pci 00:04:47.691 0000:00:04.3 (8086 2021): ioatdma -> vfio-pci 00:04:47.691 0000:00:04.2 (8086 2021): ioatdma -> vfio-pci 00:04:47.691 0000:00:04.1 (8086 2021): ioatdma -> vfio-pci 00:04:47.691 0000:00:04.0 (8086 2021): ioatdma -> vfio-pci 00:04:47.692 0000:80:04.7 (8086 2021): ioatdma -> vfio-pci 00:04:47.692 0000:80:04.6 (8086 2021): ioatdma -> vfio-pci 00:04:47.692 0000:80:04.5 (8086 2021): ioatdma -> vfio-pci 00:04:47.950 0000:80:04.4 (8086 2021): ioatdma -> vfio-pci 00:04:47.950 0000:80:04.3 (8086 2021): ioatdma -> vfio-pci 00:04:47.950 0000:80:04.2 (8086 2021): ioatdma -> vfio-pci 00:04:47.950 0000:80:04.1 (8086 2021): ioatdma -> vfio-pci 00:04:47.950 0000:80:04.0 (8086 2021): ioatdma -> vfio-pci 00:04:50.513 0000:5e:00.0 (8086 0b60): nvme -> vfio-pci 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/hugepages.sh@138 -- # verify_nr_hugepages 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/hugepages.sh@89 -- # local node 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/hugepages.sh@90 -- # local sorted_t 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/hugepages.sh@91 -- # local sorted_s 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/hugepages.sh@92 -- # local surp 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/hugepages.sh@93 -- # local resv 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/hugepages.sh@94 -- # local anon 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@17 -- # local get=AnonHugePages 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@18 -- # local node= 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@19 -- # local var val 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@20 -- # local mem_f mem 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@25 -- # [[ -n '' ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@28 -- # mapfile -t mem 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78871632 kB' 'MemAvailable: 82170432 kB' 'Buffers: 12176 kB' 'Cached: 9444168 kB' 'SwapCached: 0 kB' 'Active: 6518196 kB' 'Inactive: 3456260 kB' 'Active(anon): 6124612 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 521496 kB' 'Mapped: 168224 kB' 'Shmem: 5606500 kB' 'KReclaimable: 205196 kB' 'Slab: 526328 kB' 'SReclaimable: 205196 kB' 'SUnreclaim: 321132 kB' 'KernelStack: 16784 kB' 'PageTables: 9444 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53486764 kB' 'Committed_AS: 7546376 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 201080 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.513 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@33 -- # echo 0 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@33 -- # return 0 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/hugepages.sh@97 -- # anon=0 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@17 -- # local get=HugePages_Surp 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@18 -- # local node= 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@19 -- # local var val 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@20 -- # local mem_f mem 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@25 -- # [[ -n '' ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@28 -- # mapfile -t mem 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78882212 kB' 'MemAvailable: 82181012 kB' 'Buffers: 12176 kB' 'Cached: 9444180 kB' 'SwapCached: 0 kB' 'Active: 6517200 kB' 'Inactive: 3456260 kB' 'Active(anon): 6123616 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 520372 kB' 'Mapped: 168240 kB' 'Shmem: 5606512 kB' 'KReclaimable: 205196 kB' 'Slab: 526428 kB' 'SReclaimable: 205196 kB' 'SUnreclaim: 321232 kB' 'KernelStack: 16208 kB' 'PageTables: 8072 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53486764 kB' 'Committed_AS: 7545276 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 200856 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.514 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.515 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@33 -- # echo 0 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@33 -- # return 0 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/hugepages.sh@99 -- # surp=0 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@18 -- # local node= 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@19 -- # local var val 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@20 -- # local mem_f mem 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@25 -- # [[ -n '' ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@28 -- # mapfile -t mem 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78883004 kB' 'MemAvailable: 82181804 kB' 'Buffers: 12176 kB' 'Cached: 9444196 kB' 'SwapCached: 0 kB' 'Active: 6517132 kB' 'Inactive: 3456260 kB' 'Active(anon): 6123548 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 520312 kB' 'Mapped: 168240 kB' 'Shmem: 5606528 kB' 'KReclaimable: 205196 kB' 'Slab: 526428 kB' 'SReclaimable: 205196 kB' 'SUnreclaim: 321232 kB' 'KernelStack: 16176 kB' 'PageTables: 8364 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53486764 kB' 'Committed_AS: 7546784 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 200856 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.516 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.517 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@33 -- # echo 0 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/common.sh@33 -- # return 0 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/hugepages.sh@100 -- # resv=0 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/hugepages.sh@102 -- # echo nr_hugepages=1024 00:04:50.518 nr_hugepages=1024 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:04:50.518 resv_hugepages=0 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:04:50.518 surplus_hugepages=0 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:04:50.518 anon_hugepages=0 00:04:50.518 11:15:33 setup.sh.hugepages.default_setup -- setup/hugepages.sh@107 -- # (( 1024 == nr_hugepages + surp + resv )) 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/hugepages.sh@109 -- # (( 1024 == nr_hugepages )) 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@17 -- # local get=HugePages_Total 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@18 -- # local node= 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@19 -- # local var val 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@20 -- # local mem_f mem 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@25 -- # [[ -n '' ]] 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@28 -- # mapfile -t mem 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78881084 kB' 'MemAvailable: 82179884 kB' 'Buffers: 12176 kB' 'Cached: 9444220 kB' 'SwapCached: 0 kB' 'Active: 6517600 kB' 'Inactive: 3456260 kB' 'Active(anon): 6124016 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 520720 kB' 'Mapped: 168240 kB' 'Shmem: 5606552 kB' 'KReclaimable: 205196 kB' 'Slab: 526428 kB' 'SReclaimable: 205196 kB' 'SUnreclaim: 321232 kB' 'KernelStack: 16400 kB' 'PageTables: 8452 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53486764 kB' 'Committed_AS: 7546808 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 200968 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.518 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.519 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@33 -- # echo 1024 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@33 -- # return 0 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/hugepages.sh@110 -- # (( 1024 == nr_hugepages + surp + resv )) 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/hugepages.sh@112 -- # get_nodes 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/hugepages.sh@27 -- # local node 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=1024 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=0 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/hugepages.sh@32 -- # no_nodes=2 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@17 -- # local get=HugePages_Surp 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@18 -- # local node=0 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@19 -- # local var val 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@20 -- # local mem_f mem 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@28 -- # mapfile -t mem 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 48116940 kB' 'MemFree: 36570232 kB' 'MemUsed: 11546708 kB' 'SwapCached: 0 kB' 'Active: 5341852 kB' 'Inactive: 3372048 kB' 'Active(anon): 5183956 kB' 'Inactive(anon): 0 kB' 'Active(file): 157896 kB' 'Inactive(file): 3372048 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8448428 kB' 'Mapped: 75832 kB' 'AnonPages: 268660 kB' 'Shmem: 4918484 kB' 'KernelStack: 8984 kB' 'PageTables: 4288 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 125360 kB' 'Slab: 329224 kB' 'SReclaimable: 125360 kB' 'SUnreclaim: 203864 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Surp: 0' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.520 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@33 -- # echo 0 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/common.sh@33 -- # return 0 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/hugepages.sh@128 -- # echo 'node0=1024 expecting 1024' 00:04:50.521 node0=1024 expecting 1024 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- setup/hugepages.sh@130 -- # [[ 1024 == \1\0\2\4 ]] 00:04:50.521 00:04:50.521 real 0m6.429s 00:04:50.521 user 0m1.604s 00:04:50.521 sys 0m2.565s 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:50.521 11:15:34 setup.sh.hugepages.default_setup -- common/autotest_common.sh@10 -- # set +x 00:04:50.521 ************************************ 00:04:50.521 END TEST default_setup 00:04:50.521 ************************************ 00:04:50.780 11:15:34 setup.sh.hugepages -- common/autotest_common.sh@1142 -- # return 0 00:04:50.780 11:15:34 setup.sh.hugepages -- setup/hugepages.sh@211 -- # run_test per_node_1G_alloc per_node_1G_alloc 00:04:50.780 11:15:34 setup.sh.hugepages -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:50.780 11:15:34 setup.sh.hugepages -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:50.780 11:15:34 setup.sh.hugepages -- common/autotest_common.sh@10 -- # set +x 00:04:50.780 ************************************ 00:04:50.780 START TEST per_node_1G_alloc 00:04:50.780 ************************************ 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- common/autotest_common.sh@1123 -- # per_node_1G_alloc 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@143 -- # local IFS=, 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@145 -- # get_test_nr_hugepages 1048576 0 1 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@49 -- # local size=1048576 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@50 -- # (( 3 > 1 )) 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@51 -- # shift 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@52 -- # node_ids=('0' '1') 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@52 -- # local node_ids 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@57 -- # nr_hugepages=512 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 0 1 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@62 -- # user_nodes=('0' '1') 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@62 -- # local user_nodes 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@64 -- # local _nr_hugepages=512 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@67 -- # nodes_test=() 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@67 -- # local -g nodes_test 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@69 -- # (( 2 > 0 )) 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@70 -- # for _no_nodes in "${user_nodes[@]}" 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@71 -- # nodes_test[_no_nodes]=512 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@70 -- # for _no_nodes in "${user_nodes[@]}" 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@71 -- # nodes_test[_no_nodes]=512 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@73 -- # return 0 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@146 -- # NRHUGE=512 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@146 -- # HUGENODE=0,1 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@146 -- # setup output 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@9 -- # [[ output == output ]] 00:04:50.780 11:15:34 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@10 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh 00:04:54.970 0000:d7:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:d7:05.5 00:04:54.970 0000:85:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:85:05.5 00:04:54.970 0000:00:04.7 (8086 2021): Already using the vfio-pci driver 00:04:54.970 0000:5e:00.0 (8086 0b60): Already using the vfio-pci driver 00:04:54.970 0000:00:04.6 (8086 2021): Already using the vfio-pci driver 00:04:54.970 0000:00:04.5 (8086 2021): Already using the vfio-pci driver 00:04:54.970 0000:00:04.4 (8086 2021): Already using the vfio-pci driver 00:04:54.970 0000:00:04.3 (8086 2021): Already using the vfio-pci driver 00:04:54.970 0000:00:04.2 (8086 2021): Already using the vfio-pci driver 00:04:54.970 0000:00:04.1 (8086 2021): Already using the vfio-pci driver 00:04:54.970 0000:00:04.0 (8086 2021): Already using the vfio-pci driver 00:04:54.970 0000:80:04.7 (8086 2021): Already using the vfio-pci driver 00:04:54.970 0000:80:04.6 (8086 2021): Already using the vfio-pci driver 00:04:54.970 0000:80:04.5 (8086 2021): Already using the vfio-pci driver 00:04:54.970 0000:80:04.4 (8086 2021): Already using the vfio-pci driver 00:04:54.970 0000:80:04.3 (8086 2021): Already using the vfio-pci driver 00:04:54.970 0000:80:04.2 (8086 2021): Already using the vfio-pci driver 00:04:54.970 0000:80:04.1 (8086 2021): Already using the vfio-pci driver 00:04:54.970 0000:80:04.0 (8086 2021): Already using the vfio-pci driver 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@147 -- # nr_hugepages=1024 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@147 -- # verify_nr_hugepages 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@89 -- # local node 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@90 -- # local sorted_t 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@91 -- # local sorted_s 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@92 -- # local surp 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@93 -- # local resv 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@94 -- # local anon 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@17 -- # local get=AnonHugePages 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@18 -- # local node= 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@19 -- # local var val 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78890092 kB' 'MemAvailable: 82188892 kB' 'Buffers: 12176 kB' 'Cached: 9444312 kB' 'SwapCached: 0 kB' 'Active: 6517024 kB' 'Inactive: 3456260 kB' 'Active(anon): 6123440 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 520004 kB' 'Mapped: 167280 kB' 'Shmem: 5606644 kB' 'KReclaimable: 205196 kB' 'Slab: 526684 kB' 'SReclaimable: 205196 kB' 'SUnreclaim: 321488 kB' 'KernelStack: 16288 kB' 'PageTables: 7924 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53486764 kB' 'Committed_AS: 7542148 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 201144 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # echo 0 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # return 0 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@97 -- # anon=0 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@18 -- # local node= 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@19 -- # local var val 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:04:54.970 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78891392 kB' 'MemAvailable: 82190192 kB' 'Buffers: 12176 kB' 'Cached: 9444316 kB' 'SwapCached: 0 kB' 'Active: 6515812 kB' 'Inactive: 3456260 kB' 'Active(anon): 6122228 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 518776 kB' 'Mapped: 167200 kB' 'Shmem: 5606648 kB' 'KReclaimable: 205196 kB' 'Slab: 526756 kB' 'SReclaimable: 205196 kB' 'SUnreclaim: 321560 kB' 'KernelStack: 15984 kB' 'PageTables: 7468 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53486764 kB' 'Committed_AS: 7542168 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 200904 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # echo 0 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # return 0 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@99 -- # surp=0 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@18 -- # local node= 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@19 -- # local var val 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.971 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78888968 kB' 'MemAvailable: 82187768 kB' 'Buffers: 12176 kB' 'Cached: 9444332 kB' 'SwapCached: 0 kB' 'Active: 6516400 kB' 'Inactive: 3456260 kB' 'Active(anon): 6122816 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 519324 kB' 'Mapped: 167208 kB' 'Shmem: 5606664 kB' 'KReclaimable: 205196 kB' 'Slab: 526724 kB' 'SReclaimable: 205196 kB' 'SUnreclaim: 321528 kB' 'KernelStack: 15984 kB' 'PageTables: 7836 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53486764 kB' 'Committed_AS: 7540704 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 200984 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # echo 0 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # return 0 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@100 -- # resv=0 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@102 -- # echo nr_hugepages=1024 00:04:54.972 nr_hugepages=1024 00:04:54.972 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:04:54.972 resv_hugepages=0 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:04:54.973 surplus_hugepages=0 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:04:54.973 anon_hugepages=0 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@107 -- # (( 1024 == nr_hugepages + surp + resv )) 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@109 -- # (( 1024 == nr_hugepages )) 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@17 -- # local get=HugePages_Total 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@18 -- # local node= 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@19 -- # local var val 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78887504 kB' 'MemAvailable: 82186304 kB' 'Buffers: 12176 kB' 'Cached: 9444332 kB' 'SwapCached: 0 kB' 'Active: 6516308 kB' 'Inactive: 3456260 kB' 'Active(anon): 6122724 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 519220 kB' 'Mapped: 167200 kB' 'Shmem: 5606664 kB' 'KReclaimable: 205196 kB' 'Slab: 526724 kB' 'SReclaimable: 205196 kB' 'SUnreclaim: 321528 kB' 'KernelStack: 16096 kB' 'PageTables: 7968 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53486764 kB' 'Committed_AS: 7542212 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 200984 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:37 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # echo 1024 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # return 0 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@110 -- # (( 1024 == nr_hugepages + surp + resv )) 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@112 -- # get_nodes 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@27 -- # local node 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=512 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=512 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@32 -- # no_nodes=2 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@18 -- # local node=0 00:04:54.973 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@19 -- # local var val 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 48116940 kB' 'MemFree: 37615116 kB' 'MemUsed: 10501824 kB' 'SwapCached: 0 kB' 'Active: 5340752 kB' 'Inactive: 3372048 kB' 'Active(anon): 5182856 kB' 'Inactive(anon): 0 kB' 'Active(file): 157896 kB' 'Inactive(file): 3372048 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8448464 kB' 'Mapped: 75604 kB' 'AnonPages: 267492 kB' 'Shmem: 4918520 kB' 'KernelStack: 9160 kB' 'PageTables: 4556 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 125360 kB' 'Slab: 329340 kB' 'SReclaimable: 125360 kB' 'SUnreclaim: 203980 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 512' 'HugePages_Free: 512' 'HugePages_Surp: 0' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # echo 0 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # return 0 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 1 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@18 -- # local node=1 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@19 -- # local var val 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node1/meminfo ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node1/meminfo 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 44176532 kB' 'MemFree: 41271236 kB' 'MemUsed: 2905296 kB' 'SwapCached: 0 kB' 'Active: 1176184 kB' 'Inactive: 84212 kB' 'Active(anon): 940496 kB' 'Inactive(anon): 0 kB' 'Active(file): 235688 kB' 'Inactive(file): 84212 kB' 'Unevictable: 0 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 1008112 kB' 'Mapped: 91596 kB' 'AnonPages: 252376 kB' 'Shmem: 688212 kB' 'KernelStack: 7128 kB' 'PageTables: 3732 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 79836 kB' 'Slab: 197548 kB' 'SReclaimable: 79836 kB' 'SUnreclaim: 117712 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 512' 'HugePages_Free: 512' 'HugePages_Surp: 0' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.974 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # echo 0 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # return 0 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@128 -- # echo 'node0=512 expecting 512' 00:04:54.975 node0=512 expecting 512 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@128 -- # echo 'node1=512 expecting 512' 00:04:54.975 node1=512 expecting 512 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@130 -- # [[ 512 == \5\1\2 ]] 00:04:54.975 00:04:54.975 real 0m3.922s 00:04:54.975 user 0m1.505s 00:04:54.975 sys 0m2.514s 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:54.975 11:15:38 setup.sh.hugepages.per_node_1G_alloc -- common/autotest_common.sh@10 -- # set +x 00:04:54.975 ************************************ 00:04:54.975 END TEST per_node_1G_alloc 00:04:54.975 ************************************ 00:04:54.975 11:15:38 setup.sh.hugepages -- common/autotest_common.sh@1142 -- # return 0 00:04:54.975 11:15:38 setup.sh.hugepages -- setup/hugepages.sh@212 -- # run_test even_2G_alloc even_2G_alloc 00:04:54.975 11:15:38 setup.sh.hugepages -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:54.975 11:15:38 setup.sh.hugepages -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:54.975 11:15:38 setup.sh.hugepages -- common/autotest_common.sh@10 -- # set +x 00:04:54.975 ************************************ 00:04:54.975 START TEST even_2G_alloc 00:04:54.975 ************************************ 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- common/autotest_common.sh@1123 -- # even_2G_alloc 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@152 -- # get_test_nr_hugepages 2097152 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@49 -- # local size=2097152 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@50 -- # (( 1 > 1 )) 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@57 -- # nr_hugepages=1024 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@62 -- # user_nodes=() 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@62 -- # local user_nodes 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@64 -- # local _nr_hugepages=1024 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@67 -- # nodes_test=() 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@67 -- # local -g nodes_test 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@69 -- # (( 0 > 0 )) 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@74 -- # (( 0 > 0 )) 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@82 -- # nodes_test[_no_nodes - 1]=512 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@83 -- # : 512 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@84 -- # : 1 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@82 -- # nodes_test[_no_nodes - 1]=512 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@83 -- # : 0 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@84 -- # : 0 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@153 -- # NRHUGE=1024 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@153 -- # HUGE_EVEN_ALLOC=yes 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@153 -- # setup output 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@9 -- # [[ output == output ]] 00:04:54.975 11:15:38 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@10 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh 00:04:58.281 0000:d7:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:d7:05.5 00:04:58.281 0000:85:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:85:05.5 00:04:58.281 0000:00:04.7 (8086 2021): Already using the vfio-pci driver 00:04:58.281 0000:5e:00.0 (8086 0b60): Already using the vfio-pci driver 00:04:58.281 0000:00:04.6 (8086 2021): Already using the vfio-pci driver 00:04:58.281 0000:00:04.5 (8086 2021): Already using the vfio-pci driver 00:04:58.281 0000:00:04.4 (8086 2021): Already using the vfio-pci driver 00:04:58.281 0000:00:04.3 (8086 2021): Already using the vfio-pci driver 00:04:58.282 0000:00:04.2 (8086 2021): Already using the vfio-pci driver 00:04:58.282 0000:00:04.1 (8086 2021): Already using the vfio-pci driver 00:04:58.282 0000:00:04.0 (8086 2021): Already using the vfio-pci driver 00:04:58.282 0000:80:04.7 (8086 2021): Already using the vfio-pci driver 00:04:58.282 0000:80:04.6 (8086 2021): Already using the vfio-pci driver 00:04:58.282 0000:80:04.5 (8086 2021): Already using the vfio-pci driver 00:04:58.282 0000:80:04.4 (8086 2021): Already using the vfio-pci driver 00:04:58.282 0000:80:04.3 (8086 2021): Already using the vfio-pci driver 00:04:58.282 0000:80:04.2 (8086 2021): Already using the vfio-pci driver 00:04:58.282 0000:80:04.1 (8086 2021): Already using the vfio-pci driver 00:04:58.282 0000:80:04.0 (8086 2021): Already using the vfio-pci driver 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@154 -- # verify_nr_hugepages 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@89 -- # local node 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@90 -- # local sorted_t 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@91 -- # local sorted_s 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@92 -- # local surp 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@93 -- # local resv 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@94 -- # local anon 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@17 -- # local get=AnonHugePages 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@18 -- # local node= 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@19 -- # local var val 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78902468 kB' 'MemAvailable: 82201268 kB' 'Buffers: 12176 kB' 'Cached: 9444464 kB' 'SwapCached: 0 kB' 'Active: 6518320 kB' 'Inactive: 3456260 kB' 'Active(anon): 6124736 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 520684 kB' 'Mapped: 167364 kB' 'Shmem: 5606796 kB' 'KReclaimable: 205196 kB' 'Slab: 526660 kB' 'SReclaimable: 205196 kB' 'SUnreclaim: 321464 kB' 'KernelStack: 16112 kB' 'PageTables: 7924 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53486764 kB' 'Committed_AS: 7540084 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 200952 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.282 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # echo 0 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # return 0 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@97 -- # anon=0 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@18 -- # local node= 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@19 -- # local var val 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78902244 kB' 'MemAvailable: 82201044 kB' 'Buffers: 12176 kB' 'Cached: 9444468 kB' 'SwapCached: 0 kB' 'Active: 6517292 kB' 'Inactive: 3456260 kB' 'Active(anon): 6123708 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 520148 kB' 'Mapped: 167196 kB' 'Shmem: 5606800 kB' 'KReclaimable: 205196 kB' 'Slab: 526632 kB' 'SReclaimable: 205196 kB' 'SUnreclaim: 321436 kB' 'KernelStack: 16080 kB' 'PageTables: 7800 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53486764 kB' 'Committed_AS: 7540100 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 200968 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.283 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.284 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # echo 0 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # return 0 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@99 -- # surp=0 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@18 -- # local node= 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@19 -- # local var val 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78902244 kB' 'MemAvailable: 82201044 kB' 'Buffers: 12176 kB' 'Cached: 9444468 kB' 'SwapCached: 0 kB' 'Active: 6517328 kB' 'Inactive: 3456260 kB' 'Active(anon): 6123744 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 520180 kB' 'Mapped: 167196 kB' 'Shmem: 5606800 kB' 'KReclaimable: 205196 kB' 'Slab: 526632 kB' 'SReclaimable: 205196 kB' 'SUnreclaim: 321436 kB' 'KernelStack: 16096 kB' 'PageTables: 7852 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53486764 kB' 'Committed_AS: 7540120 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 200968 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.285 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.286 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.286 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.286 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.286 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.286 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.286 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.286 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.286 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.286 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.286 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.286 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.286 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.286 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.286 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.286 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.286 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.286 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.286 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.287 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # echo 0 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # return 0 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@100 -- # resv=0 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@102 -- # echo nr_hugepages=1024 00:04:58.288 nr_hugepages=1024 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:04:58.288 resv_hugepages=0 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:04:58.288 surplus_hugepages=0 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:04:58.288 anon_hugepages=0 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@107 -- # (( 1024 == nr_hugepages + surp + resv )) 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@109 -- # (( 1024 == nr_hugepages )) 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@17 -- # local get=HugePages_Total 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@18 -- # local node= 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@19 -- # local var val 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78902496 kB' 'MemAvailable: 82201296 kB' 'Buffers: 12176 kB' 'Cached: 9444508 kB' 'SwapCached: 0 kB' 'Active: 6517224 kB' 'Inactive: 3456260 kB' 'Active(anon): 6123640 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 520088 kB' 'Mapped: 167196 kB' 'Shmem: 5606840 kB' 'KReclaimable: 205196 kB' 'Slab: 526632 kB' 'SReclaimable: 205196 kB' 'SUnreclaim: 321436 kB' 'KernelStack: 16080 kB' 'PageTables: 7824 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53486764 kB' 'Committed_AS: 7561308 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 200952 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.288 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # echo 1024 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # return 0 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@110 -- # (( 1024 == nr_hugepages + surp + resv )) 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@112 -- # get_nodes 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@27 -- # local node 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=512 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=512 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@32 -- # no_nodes=2 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@18 -- # local node=0 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@19 -- # local var val 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:04:58.289 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 48116940 kB' 'MemFree: 37600372 kB' 'MemUsed: 10516568 kB' 'SwapCached: 0 kB' 'Active: 5340676 kB' 'Inactive: 3372048 kB' 'Active(anon): 5182780 kB' 'Inactive(anon): 0 kB' 'Active(file): 157896 kB' 'Inactive(file): 3372048 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8448468 kB' 'Mapped: 75600 kB' 'AnonPages: 267376 kB' 'Shmem: 4918524 kB' 'KernelStack: 8936 kB' 'PageTables: 4024 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 125360 kB' 'Slab: 329252 kB' 'SReclaimable: 125360 kB' 'SUnreclaim: 203892 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 512' 'HugePages_Free: 512' 'HugePages_Surp: 0' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.290 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # echo 0 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # return 0 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 1 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@18 -- # local node=1 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@19 -- # local var val 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node1/meminfo ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node1/meminfo 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 44176532 kB' 'MemFree: 41303588 kB' 'MemUsed: 2872944 kB' 'SwapCached: 0 kB' 'Active: 1176972 kB' 'Inactive: 84212 kB' 'Active(anon): 941284 kB' 'Inactive(anon): 0 kB' 'Active(file): 235688 kB' 'Inactive(file): 84212 kB' 'Unevictable: 0 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 1008260 kB' 'Mapped: 91596 kB' 'AnonPages: 253028 kB' 'Shmem: 688360 kB' 'KernelStack: 7096 kB' 'PageTables: 3656 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 79836 kB' 'Slab: 197380 kB' 'SReclaimable: 79836 kB' 'SUnreclaim: 117544 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 512' 'HugePages_Free: 512' 'HugePages_Surp: 0' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.291 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # echo 0 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # return 0 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@128 -- # echo 'node0=512 expecting 512' 00:04:58.292 node0=512 expecting 512 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@128 -- # echo 'node1=512 expecting 512' 00:04:58.292 node1=512 expecting 512 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@130 -- # [[ 512 == \5\1\2 ]] 00:04:58.292 00:04:58.292 real 0m3.668s 00:04:58.292 user 0m1.406s 00:04:58.292 sys 0m2.342s 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:58.292 11:15:41 setup.sh.hugepages.even_2G_alloc -- common/autotest_common.sh@10 -- # set +x 00:04:58.292 ************************************ 00:04:58.292 END TEST even_2G_alloc 00:04:58.292 ************************************ 00:04:58.549 11:15:41 setup.sh.hugepages -- common/autotest_common.sh@1142 -- # return 0 00:04:58.549 11:15:41 setup.sh.hugepages -- setup/hugepages.sh@213 -- # run_test odd_alloc odd_alloc 00:04:58.549 11:15:41 setup.sh.hugepages -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:58.549 11:15:41 setup.sh.hugepages -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:58.549 11:15:41 setup.sh.hugepages -- common/autotest_common.sh@10 -- # set +x 00:04:58.549 ************************************ 00:04:58.549 START TEST odd_alloc 00:04:58.549 ************************************ 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- common/autotest_common.sh@1123 -- # odd_alloc 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@159 -- # get_test_nr_hugepages 2098176 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@49 -- # local size=2098176 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@50 -- # (( 1 > 1 )) 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@57 -- # nr_hugepages=1025 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@62 -- # user_nodes=() 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@62 -- # local user_nodes 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@64 -- # local _nr_hugepages=1025 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@67 -- # nodes_test=() 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@67 -- # local -g nodes_test 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@69 -- # (( 0 > 0 )) 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@74 -- # (( 0 > 0 )) 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@82 -- # nodes_test[_no_nodes - 1]=512 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@83 -- # : 513 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@84 -- # : 1 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@82 -- # nodes_test[_no_nodes - 1]=513 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@83 -- # : 0 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@84 -- # : 0 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@160 -- # HUGEMEM=2049 00:04:58.549 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@160 -- # HUGE_EVEN_ALLOC=yes 00:04:58.550 11:15:41 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@160 -- # setup output 00:04:58.550 11:15:41 setup.sh.hugepages.odd_alloc -- setup/common.sh@9 -- # [[ output == output ]] 00:04:58.550 11:15:41 setup.sh.hugepages.odd_alloc -- setup/common.sh@10 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh 00:05:01.828 0000:d7:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:d7:05.5 00:05:01.828 0000:85:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:85:05.5 00:05:01.828 0000:00:04.7 (8086 2021): Already using the vfio-pci driver 00:05:01.828 0000:5e:00.0 (8086 0b60): Already using the vfio-pci driver 00:05:01.828 0000:00:04.6 (8086 2021): Already using the vfio-pci driver 00:05:01.828 0000:00:04.5 (8086 2021): Already using the vfio-pci driver 00:05:01.828 0000:00:04.4 (8086 2021): Already using the vfio-pci driver 00:05:01.828 0000:00:04.3 (8086 2021): Already using the vfio-pci driver 00:05:01.828 0000:00:04.2 (8086 2021): Already using the vfio-pci driver 00:05:01.828 0000:00:04.1 (8086 2021): Already using the vfio-pci driver 00:05:01.828 0000:00:04.0 (8086 2021): Already using the vfio-pci driver 00:05:01.828 0000:80:04.7 (8086 2021): Already using the vfio-pci driver 00:05:01.828 0000:80:04.6 (8086 2021): Already using the vfio-pci driver 00:05:01.828 0000:80:04.5 (8086 2021): Already using the vfio-pci driver 00:05:01.828 0000:80:04.4 (8086 2021): Already using the vfio-pci driver 00:05:01.828 0000:80:04.3 (8086 2021): Already using the vfio-pci driver 00:05:02.091 0000:80:04.2 (8086 2021): Already using the vfio-pci driver 00:05:02.091 0000:80:04.1 (8086 2021): Already using the vfio-pci driver 00:05:02.091 0000:80:04.0 (8086 2021): Already using the vfio-pci driver 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@161 -- # verify_nr_hugepages 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@89 -- # local node 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@90 -- # local sorted_t 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@91 -- # local sorted_s 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@92 -- # local surp 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@93 -- # local resv 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@94 -- # local anon 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@17 -- # local get=AnonHugePages 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@18 -- # local node= 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@19 -- # local var val 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78885528 kB' 'MemAvailable: 82184328 kB' 'Buffers: 12176 kB' 'Cached: 9444624 kB' 'SwapCached: 0 kB' 'Active: 6517628 kB' 'Inactive: 3456260 kB' 'Active(anon): 6124044 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 520440 kB' 'Mapped: 167364 kB' 'Shmem: 5606956 kB' 'KReclaimable: 205196 kB' 'Slab: 526892 kB' 'SReclaimable: 205196 kB' 'SUnreclaim: 321696 kB' 'KernelStack: 16112 kB' 'PageTables: 7932 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53485740 kB' 'Committed_AS: 7540952 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 200872 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1025' 'HugePages_Free: 1025' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2099200 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.091 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.092 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # echo 0 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # return 0 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@97 -- # anon=0 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@18 -- # local node= 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@19 -- # local var val 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78886536 kB' 'MemAvailable: 82185336 kB' 'Buffers: 12176 kB' 'Cached: 9444624 kB' 'SwapCached: 0 kB' 'Active: 6518144 kB' 'Inactive: 3456260 kB' 'Active(anon): 6124560 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 520968 kB' 'Mapped: 167364 kB' 'Shmem: 5606956 kB' 'KReclaimable: 205196 kB' 'Slab: 526892 kB' 'SReclaimable: 205196 kB' 'SUnreclaim: 321696 kB' 'KernelStack: 16096 kB' 'PageTables: 7812 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53485740 kB' 'Committed_AS: 7540968 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 200824 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1025' 'HugePages_Free: 1025' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2099200 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.093 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # echo 0 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # return 0 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@99 -- # surp=0 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@18 -- # local node= 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@19 -- # local var val 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.094 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78886888 kB' 'MemAvailable: 82185688 kB' 'Buffers: 12176 kB' 'Cached: 9444644 kB' 'SwapCached: 0 kB' 'Active: 6517868 kB' 'Inactive: 3456260 kB' 'Active(anon): 6124284 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 520636 kB' 'Mapped: 167212 kB' 'Shmem: 5606976 kB' 'KReclaimable: 205196 kB' 'Slab: 526916 kB' 'SReclaimable: 205196 kB' 'SUnreclaim: 321720 kB' 'KernelStack: 16096 kB' 'PageTables: 7812 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53485740 kB' 'Committed_AS: 7540988 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 200840 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1025' 'HugePages_Free: 1025' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2099200 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.095 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # echo 0 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # return 0 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@100 -- # resv=0 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@102 -- # echo nr_hugepages=1025 00:05:02.096 nr_hugepages=1025 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:05:02.096 resv_hugepages=0 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:05:02.096 surplus_hugepages=0 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:05:02.096 anon_hugepages=0 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@107 -- # (( 1025 == nr_hugepages + surp + resv )) 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@109 -- # (( 1025 == nr_hugepages )) 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@17 -- # local get=HugePages_Total 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@18 -- # local node= 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@19 -- # local var val 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78887344 kB' 'MemAvailable: 82186144 kB' 'Buffers: 12176 kB' 'Cached: 9444684 kB' 'SwapCached: 0 kB' 'Active: 6517576 kB' 'Inactive: 3456260 kB' 'Active(anon): 6123992 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 520260 kB' 'Mapped: 167212 kB' 'Shmem: 5607016 kB' 'KReclaimable: 205196 kB' 'Slab: 526916 kB' 'SReclaimable: 205196 kB' 'SUnreclaim: 321720 kB' 'KernelStack: 16080 kB' 'PageTables: 7760 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53485740 kB' 'Committed_AS: 7541008 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 200856 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1025' 'HugePages_Free: 1025' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2099200 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.096 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.097 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # echo 1025 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # return 0 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@110 -- # (( 1025 == nr_hugepages + surp + resv )) 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@112 -- # get_nodes 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@27 -- # local node 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=512 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=513 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@32 -- # no_nodes=2 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@18 -- # local node=0 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@19 -- # local var val 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 48116940 kB' 'MemFree: 37589432 kB' 'MemUsed: 10527508 kB' 'SwapCached: 0 kB' 'Active: 5340660 kB' 'Inactive: 3372048 kB' 'Active(anon): 5182764 kB' 'Inactive(anon): 0 kB' 'Active(file): 157896 kB' 'Inactive(file): 3372048 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8448496 kB' 'Mapped: 75612 kB' 'AnonPages: 267412 kB' 'Shmem: 4918552 kB' 'KernelStack: 8952 kB' 'PageTables: 4028 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 125360 kB' 'Slab: 329500 kB' 'SReclaimable: 125360 kB' 'SUnreclaim: 204140 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 512' 'HugePages_Free: 512' 'HugePages_Surp: 0' 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.098 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.360 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # echo 0 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # return 0 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 1 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@18 -- # local node=1 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@19 -- # local var val 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node1/meminfo ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node1/meminfo 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 44176532 kB' 'MemFree: 41298416 kB' 'MemUsed: 2878116 kB' 'SwapCached: 0 kB' 'Active: 1177572 kB' 'Inactive: 84212 kB' 'Active(anon): 941884 kB' 'Inactive(anon): 0 kB' 'Active(file): 235688 kB' 'Inactive(file): 84212 kB' 'Unevictable: 0 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 1008384 kB' 'Mapped: 91600 kB' 'AnonPages: 253492 kB' 'Shmem: 688484 kB' 'KernelStack: 7144 kB' 'PageTables: 3784 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 79836 kB' 'Slab: 197416 kB' 'SReclaimable: 79836 kB' 'SUnreclaim: 117580 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 513' 'HugePages_Free: 513' 'HugePages_Surp: 0' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.361 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # echo 0 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # return 0 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@128 -- # echo 'node0=512 expecting 513' 00:05:02.362 node0=512 expecting 513 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@128 -- # echo 'node1=513 expecting 512' 00:05:02.362 node1=513 expecting 512 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@130 -- # [[ 512 513 == \5\1\2\ \5\1\3 ]] 00:05:02.362 00:05:02.362 real 0m3.812s 00:05:02.362 user 0m1.496s 00:05:02.362 sys 0m2.420s 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:02.362 11:15:45 setup.sh.hugepages.odd_alloc -- common/autotest_common.sh@10 -- # set +x 00:05:02.362 ************************************ 00:05:02.362 END TEST odd_alloc 00:05:02.362 ************************************ 00:05:02.362 11:15:45 setup.sh.hugepages -- common/autotest_common.sh@1142 -- # return 0 00:05:02.362 11:15:45 setup.sh.hugepages -- setup/hugepages.sh@214 -- # run_test custom_alloc custom_alloc 00:05:02.362 11:15:45 setup.sh.hugepages -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:05:02.362 11:15:45 setup.sh.hugepages -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:02.362 11:15:45 setup.sh.hugepages -- common/autotest_common.sh@10 -- # set +x 00:05:02.362 ************************************ 00:05:02.362 START TEST custom_alloc 00:05:02.362 ************************************ 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- common/autotest_common.sh@1123 -- # custom_alloc 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@167 -- # local IFS=, 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@169 -- # local node 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@170 -- # nodes_hp=() 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@170 -- # local nodes_hp 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@172 -- # local nr_hugepages=0 _nr_hugepages=0 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@174 -- # get_test_nr_hugepages 1048576 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@49 -- # local size=1048576 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@50 -- # (( 1 > 1 )) 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@57 -- # nr_hugepages=512 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@62 -- # user_nodes=() 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@62 -- # local user_nodes 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@64 -- # local _nr_hugepages=512 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@67 -- # nodes_test=() 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@67 -- # local -g nodes_test 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@69 -- # (( 0 > 0 )) 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@74 -- # (( 0 > 0 )) 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@82 -- # nodes_test[_no_nodes - 1]=256 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@83 -- # : 256 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@84 -- # : 1 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@82 -- # nodes_test[_no_nodes - 1]=256 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@83 -- # : 0 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@84 -- # : 0 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@175 -- # nodes_hp[0]=512 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@176 -- # (( 2 > 1 )) 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@177 -- # get_test_nr_hugepages 2097152 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@49 -- # local size=2097152 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@50 -- # (( 1 > 1 )) 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@57 -- # nr_hugepages=1024 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@62 -- # user_nodes=() 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@62 -- # local user_nodes 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@64 -- # local _nr_hugepages=1024 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:05:02.362 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@67 -- # nodes_test=() 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@67 -- # local -g nodes_test 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@69 -- # (( 0 > 0 )) 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@74 -- # (( 1 > 0 )) 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@75 -- # for _no_nodes in "${!nodes_hp[@]}" 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@76 -- # nodes_test[_no_nodes]=512 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@78 -- # return 0 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@178 -- # nodes_hp[1]=1024 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@181 -- # for node in "${!nodes_hp[@]}" 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@182 -- # HUGENODE+=("nodes_hp[$node]=${nodes_hp[node]}") 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@183 -- # (( _nr_hugepages += nodes_hp[node] )) 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@181 -- # for node in "${!nodes_hp[@]}" 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@182 -- # HUGENODE+=("nodes_hp[$node]=${nodes_hp[node]}") 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@183 -- # (( _nr_hugepages += nodes_hp[node] )) 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@186 -- # get_test_nr_hugepages_per_node 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@62 -- # user_nodes=() 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@62 -- # local user_nodes 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@64 -- # local _nr_hugepages=1024 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@67 -- # nodes_test=() 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@67 -- # local -g nodes_test 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@69 -- # (( 0 > 0 )) 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@74 -- # (( 2 > 0 )) 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@75 -- # for _no_nodes in "${!nodes_hp[@]}" 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@76 -- # nodes_test[_no_nodes]=512 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@75 -- # for _no_nodes in "${!nodes_hp[@]}" 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@76 -- # nodes_test[_no_nodes]=1024 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@78 -- # return 0 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@187 -- # HUGENODE='nodes_hp[0]=512,nodes_hp[1]=1024' 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@187 -- # setup output 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/common.sh@9 -- # [[ output == output ]] 00:05:02.363 11:15:45 setup.sh.hugepages.custom_alloc -- setup/common.sh@10 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh 00:05:05.647 0000:d7:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:d7:05.5 00:05:05.647 0000:85:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:85:05.5 00:05:05.647 0000:00:04.7 (8086 2021): Already using the vfio-pci driver 00:05:05.647 0000:5e:00.0 (8086 0b60): Already using the vfio-pci driver 00:05:05.647 0000:00:04.6 (8086 2021): Already using the vfio-pci driver 00:05:05.647 0000:00:04.5 (8086 2021): Already using the vfio-pci driver 00:05:05.647 0000:00:04.4 (8086 2021): Already using the vfio-pci driver 00:05:05.647 0000:00:04.3 (8086 2021): Already using the vfio-pci driver 00:05:05.647 0000:00:04.2 (8086 2021): Already using the vfio-pci driver 00:05:05.647 0000:00:04.1 (8086 2021): Already using the vfio-pci driver 00:05:05.647 0000:00:04.0 (8086 2021): Already using the vfio-pci driver 00:05:05.647 0000:80:04.7 (8086 2021): Already using the vfio-pci driver 00:05:05.647 0000:80:04.6 (8086 2021): Already using the vfio-pci driver 00:05:05.647 0000:80:04.5 (8086 2021): Already using the vfio-pci driver 00:05:05.647 0000:80:04.4 (8086 2021): Already using the vfio-pci driver 00:05:05.647 0000:80:04.3 (8086 2021): Already using the vfio-pci driver 00:05:05.647 0000:80:04.2 (8086 2021): Already using the vfio-pci driver 00:05:05.647 0000:80:04.1 (8086 2021): Already using the vfio-pci driver 00:05:05.647 0000:80:04.0 (8086 2021): Already using the vfio-pci driver 00:05:05.911 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@188 -- # nr_hugepages=1536 00:05:05.911 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@188 -- # verify_nr_hugepages 00:05:05.911 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@89 -- # local node 00:05:05.911 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@90 -- # local sorted_t 00:05:05.911 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@91 -- # local sorted_s 00:05:05.911 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@92 -- # local surp 00:05:05.911 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@93 -- # local resv 00:05:05.911 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@94 -- # local anon 00:05:05.911 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:05:05.911 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:05:05.911 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@17 -- # local get=AnonHugePages 00:05:05.911 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@18 -- # local node= 00:05:05.911 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@19 -- # local var val 00:05:05.911 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:05.911 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:05.911 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:05:05.911 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:05:05.911 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 77820264 kB' 'MemAvailable: 81119048 kB' 'Buffers: 12176 kB' 'Cached: 9444772 kB' 'SwapCached: 0 kB' 'Active: 6518716 kB' 'Inactive: 3456260 kB' 'Active(anon): 6125132 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 521264 kB' 'Mapped: 167228 kB' 'Shmem: 5607104 kB' 'KReclaimable: 205164 kB' 'Slab: 526924 kB' 'SReclaimable: 205164 kB' 'SUnreclaim: 321760 kB' 'KernelStack: 16128 kB' 'PageTables: 7852 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 52962476 kB' 'Committed_AS: 7541164 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 201000 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1536' 'HugePages_Free: 1536' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 3145728 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.912 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # echo 0 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # return 0 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@97 -- # anon=0 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@18 -- # local node= 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@19 -- # local var val 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 77821000 kB' 'MemAvailable: 81119784 kB' 'Buffers: 12176 kB' 'Cached: 9444776 kB' 'SwapCached: 0 kB' 'Active: 6518776 kB' 'Inactive: 3456260 kB' 'Active(anon): 6125192 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 521324 kB' 'Mapped: 167212 kB' 'Shmem: 5607108 kB' 'KReclaimable: 205164 kB' 'Slab: 526900 kB' 'SReclaimable: 205164 kB' 'SUnreclaim: 321736 kB' 'KernelStack: 16128 kB' 'PageTables: 7828 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 52962476 kB' 'Committed_AS: 7541184 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 200968 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1536' 'HugePages_Free: 1536' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 3145728 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.913 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.914 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # echo 0 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # return 0 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@99 -- # surp=0 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@18 -- # local node= 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@19 -- # local var val 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 77824368 kB' 'MemAvailable: 81123152 kB' 'Buffers: 12176 kB' 'Cached: 9444792 kB' 'SwapCached: 0 kB' 'Active: 6518400 kB' 'Inactive: 3456260 kB' 'Active(anon): 6124816 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 520940 kB' 'Mapped: 167212 kB' 'Shmem: 5607124 kB' 'KReclaimable: 205164 kB' 'Slab: 526960 kB' 'SReclaimable: 205164 kB' 'SUnreclaim: 321796 kB' 'KernelStack: 16112 kB' 'PageTables: 7812 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 52962476 kB' 'Committed_AS: 7541204 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 200952 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1536' 'HugePages_Free: 1536' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 3145728 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.915 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.916 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # echo 0 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # return 0 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@100 -- # resv=0 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@102 -- # echo nr_hugepages=1536 00:05:05.917 nr_hugepages=1536 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:05:05.917 resv_hugepages=0 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:05:05.917 surplus_hugepages=0 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:05:05.917 anon_hugepages=0 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@107 -- # (( 1536 == nr_hugepages + surp + resv )) 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@109 -- # (( 1536 == nr_hugepages )) 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@17 -- # local get=HugePages_Total 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@18 -- # local node= 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@19 -- # local var val 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 77824592 kB' 'MemAvailable: 81123376 kB' 'Buffers: 12176 kB' 'Cached: 9444832 kB' 'SwapCached: 0 kB' 'Active: 6518064 kB' 'Inactive: 3456260 kB' 'Active(anon): 6124480 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 520548 kB' 'Mapped: 167212 kB' 'Shmem: 5607164 kB' 'KReclaimable: 205164 kB' 'Slab: 526928 kB' 'SReclaimable: 205164 kB' 'SUnreclaim: 321764 kB' 'KernelStack: 16096 kB' 'PageTables: 7760 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 52962476 kB' 'Committed_AS: 7541228 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 200952 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1536' 'HugePages_Free: 1536' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 3145728 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.917 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # echo 1536 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # return 0 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@110 -- # (( 1536 == nr_hugepages + surp + resv )) 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@112 -- # get_nodes 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@27 -- # local node 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=512 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=1024 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@32 -- # no_nodes=2 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@18 -- # local node=0 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@19 -- # local var val 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:05.918 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 48116940 kB' 'MemFree: 37582544 kB' 'MemUsed: 10534396 kB' 'SwapCached: 0 kB' 'Active: 5341404 kB' 'Inactive: 3372048 kB' 'Active(anon): 5183508 kB' 'Inactive(anon): 0 kB' 'Active(file): 157896 kB' 'Inactive(file): 3372048 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8448588 kB' 'Mapped: 75612 kB' 'AnonPages: 268108 kB' 'Shmem: 4918644 kB' 'KernelStack: 9000 kB' 'PageTables: 4216 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 125360 kB' 'Slab: 329776 kB' 'SReclaimable: 125360 kB' 'SUnreclaim: 204416 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 512' 'HugePages_Free: 512' 'HugePages_Surp: 0' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.919 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # echo 0 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # return 0 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 1 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@18 -- # local node=1 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@19 -- # local var val 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node1/meminfo ]] 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node1/meminfo 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 44176532 kB' 'MemFree: 40241180 kB' 'MemUsed: 3935352 kB' 'SwapCached: 0 kB' 'Active: 1177140 kB' 'Inactive: 84212 kB' 'Active(anon): 941452 kB' 'Inactive(anon): 0 kB' 'Active(file): 235688 kB' 'Inactive(file): 84212 kB' 'Unevictable: 0 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 1008436 kB' 'Mapped: 91600 kB' 'AnonPages: 252960 kB' 'Shmem: 688536 kB' 'KernelStack: 7128 kB' 'PageTables: 3640 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 79804 kB' 'Slab: 197152 kB' 'SReclaimable: 79804 kB' 'SUnreclaim: 117348 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Surp: 0' 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:05.920 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.180 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # echo 0 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # return 0 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@128 -- # echo 'node0=512 expecting 512' 00:05:06.181 node0=512 expecting 512 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@128 -- # echo 'node1=1024 expecting 1024' 00:05:06.181 node1=1024 expecting 1024 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@130 -- # [[ 512,1024 == \5\1\2\,\1\0\2\4 ]] 00:05:06.181 00:05:06.181 real 0m3.721s 00:05:06.181 user 0m1.457s 00:05:06.181 sys 0m2.331s 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:06.181 11:15:49 setup.sh.hugepages.custom_alloc -- common/autotest_common.sh@10 -- # set +x 00:05:06.181 ************************************ 00:05:06.181 END TEST custom_alloc 00:05:06.181 ************************************ 00:05:06.181 11:15:49 setup.sh.hugepages -- common/autotest_common.sh@1142 -- # return 0 00:05:06.181 11:15:49 setup.sh.hugepages -- setup/hugepages.sh@215 -- # run_test no_shrink_alloc no_shrink_alloc 00:05:06.181 11:15:49 setup.sh.hugepages -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:05:06.181 11:15:49 setup.sh.hugepages -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:06.181 11:15:49 setup.sh.hugepages -- common/autotest_common.sh@10 -- # set +x 00:05:06.181 ************************************ 00:05:06.181 START TEST no_shrink_alloc 00:05:06.181 ************************************ 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- common/autotest_common.sh@1123 -- # no_shrink_alloc 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@195 -- # get_test_nr_hugepages 2097152 0 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@49 -- # local size=2097152 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@50 -- # (( 2 > 1 )) 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@51 -- # shift 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@52 -- # node_ids=('0') 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@52 -- # local node_ids 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@57 -- # nr_hugepages=1024 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 0 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@62 -- # user_nodes=('0') 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@62 -- # local user_nodes 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@64 -- # local _nr_hugepages=1024 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@67 -- # nodes_test=() 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@67 -- # local -g nodes_test 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@69 -- # (( 1 > 0 )) 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@70 -- # for _no_nodes in "${user_nodes[@]}" 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@71 -- # nodes_test[_no_nodes]=1024 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@73 -- # return 0 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@198 -- # setup output 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@9 -- # [[ output == output ]] 00:05:06.181 11:15:49 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@10 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh 00:05:09.477 0000:d7:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:d7:05.5 00:05:09.477 0000:85:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:85:05.5 00:05:09.477 0000:00:04.7 (8086 2021): Already using the vfio-pci driver 00:05:09.477 0000:5e:00.0 (8086 0b60): Already using the vfio-pci driver 00:05:09.477 0000:00:04.6 (8086 2021): Already using the vfio-pci driver 00:05:09.477 0000:00:04.5 (8086 2021): Already using the vfio-pci driver 00:05:09.477 0000:00:04.4 (8086 2021): Already using the vfio-pci driver 00:05:09.477 0000:00:04.3 (8086 2021): Already using the vfio-pci driver 00:05:09.477 0000:00:04.2 (8086 2021): Already using the vfio-pci driver 00:05:09.477 0000:00:04.1 (8086 2021): Already using the vfio-pci driver 00:05:09.477 0000:00:04.0 (8086 2021): Already using the vfio-pci driver 00:05:09.477 0000:80:04.7 (8086 2021): Already using the vfio-pci driver 00:05:09.477 0000:80:04.6 (8086 2021): Already using the vfio-pci driver 00:05:09.477 0000:80:04.5 (8086 2021): Already using the vfio-pci driver 00:05:09.477 0000:80:04.4 (8086 2021): Already using the vfio-pci driver 00:05:09.477 0000:80:04.3 (8086 2021): Already using the vfio-pci driver 00:05:09.477 0000:80:04.2 (8086 2021): Already using the vfio-pci driver 00:05:09.477 0000:80:04.1 (8086 2021): Already using the vfio-pci driver 00:05:09.477 0000:80:04.0 (8086 2021): Already using the vfio-pci driver 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@199 -- # verify_nr_hugepages 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@89 -- # local node 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@90 -- # local sorted_t 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@91 -- # local sorted_s 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@92 -- # local surp 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@93 -- # local resv 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@94 -- # local anon 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@17 -- # local get=AnonHugePages 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@18 -- # local node= 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@19 -- # local var val 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78880484 kB' 'MemAvailable: 82179236 kB' 'Buffers: 12176 kB' 'Cached: 9444912 kB' 'SwapCached: 0 kB' 'Active: 6521772 kB' 'Inactive: 3456260 kB' 'Active(anon): 6128188 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 523712 kB' 'Mapped: 167260 kB' 'Shmem: 5607244 kB' 'KReclaimable: 205100 kB' 'Slab: 527176 kB' 'SReclaimable: 205100 kB' 'SUnreclaim: 322076 kB' 'KernelStack: 16336 kB' 'PageTables: 8268 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53486764 kB' 'Committed_AS: 7541692 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 201112 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.477 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.478 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # echo 0 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # return 0 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@97 -- # anon=0 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@18 -- # local node= 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@19 -- # local var val 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78881372 kB' 'MemAvailable: 82180124 kB' 'Buffers: 12176 kB' 'Cached: 9444912 kB' 'SwapCached: 0 kB' 'Active: 6521720 kB' 'Inactive: 3456260 kB' 'Active(anon): 6128136 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 523612 kB' 'Mapped: 167220 kB' 'Shmem: 5607244 kB' 'KReclaimable: 205100 kB' 'Slab: 527128 kB' 'SReclaimable: 205100 kB' 'SUnreclaim: 322028 kB' 'KernelStack: 16112 kB' 'PageTables: 7680 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53486764 kB' 'Committed_AS: 7541712 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 201048 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.479 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.480 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # echo 0 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # return 0 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@99 -- # surp=0 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@18 -- # local node= 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@19 -- # local var val 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78881980 kB' 'MemAvailable: 82180732 kB' 'Buffers: 12176 kB' 'Cached: 9444936 kB' 'SwapCached: 0 kB' 'Active: 6521060 kB' 'Inactive: 3456260 kB' 'Active(anon): 6127476 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 523472 kB' 'Mapped: 167224 kB' 'Shmem: 5607268 kB' 'KReclaimable: 205100 kB' 'Slab: 527160 kB' 'SReclaimable: 205100 kB' 'SUnreclaim: 322060 kB' 'KernelStack: 16224 kB' 'PageTables: 8052 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53486764 kB' 'Committed_AS: 7541732 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 201032 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.481 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.482 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # echo 0 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # return 0 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@100 -- # resv=0 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@102 -- # echo nr_hugepages=1024 00:05:09.483 nr_hugepages=1024 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:05:09.483 resv_hugepages=0 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:05:09.483 surplus_hugepages=0 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:05:09.483 anon_hugepages=0 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@107 -- # (( 1024 == nr_hugepages + surp + resv )) 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@109 -- # (( 1024 == nr_hugepages )) 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@17 -- # local get=HugePages_Total 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@18 -- # local node= 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@19 -- # local var val 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78881696 kB' 'MemAvailable: 82180448 kB' 'Buffers: 12176 kB' 'Cached: 9444956 kB' 'SwapCached: 0 kB' 'Active: 6521864 kB' 'Inactive: 3456260 kB' 'Active(anon): 6128280 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 524276 kB' 'Mapped: 167728 kB' 'Shmem: 5607288 kB' 'KReclaimable: 205100 kB' 'Slab: 527160 kB' 'SReclaimable: 205100 kB' 'SUnreclaim: 322060 kB' 'KernelStack: 16224 kB' 'PageTables: 8052 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53486764 kB' 'Committed_AS: 7542980 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 201032 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.483 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.484 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # echo 1024 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # return 0 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@110 -- # (( 1024 == nr_hugepages + surp + resv )) 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@112 -- # get_nodes 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@27 -- # local node 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=1024 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=0 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@32 -- # no_nodes=2 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:05:09.485 11:15:52 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@18 -- # local node=0 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@19 -- # local var val 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 48116940 kB' 'MemFree: 36531784 kB' 'MemUsed: 11585156 kB' 'SwapCached: 0 kB' 'Active: 5345288 kB' 'Inactive: 3372048 kB' 'Active(anon): 5187392 kB' 'Inactive(anon): 0 kB' 'Active(file): 157896 kB' 'Inactive(file): 3372048 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8448676 kB' 'Mapped: 76128 kB' 'AnonPages: 271864 kB' 'Shmem: 4918732 kB' 'KernelStack: 8952 kB' 'PageTables: 4024 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 125360 kB' 'Slab: 330032 kB' 'SReclaimable: 125360 kB' 'SUnreclaim: 204672 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Surp: 0' 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.485 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.486 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.487 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.487 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.487 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.487 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:09.487 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:09.487 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:09.487 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:09.487 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # echo 0 00:05:09.487 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # return 0 00:05:09.487 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:05:09.487 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:05:09.487 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:05:09.487 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:05:09.487 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@128 -- # echo 'node0=1024 expecting 1024' 00:05:09.487 node0=1024 expecting 1024 00:05:09.487 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@130 -- # [[ 1024 == \1\0\2\4 ]] 00:05:09.487 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@202 -- # CLEAR_HUGE=no 00:05:09.487 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@202 -- # NRHUGE=512 00:05:09.487 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@202 -- # setup output 00:05:09.487 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@9 -- # [[ output == output ]] 00:05:09.487 11:15:53 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@10 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh 00:05:13.708 0000:d7:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:d7:05.5 00:05:13.708 0000:85:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:85:05.5 00:05:13.708 0000:00:04.7 (8086 2021): Already using the vfio-pci driver 00:05:13.708 0000:5e:00.0 (8086 0b60): Already using the vfio-pci driver 00:05:13.708 0000:00:04.6 (8086 2021): Already using the vfio-pci driver 00:05:13.708 0000:00:04.5 (8086 2021): Already using the vfio-pci driver 00:05:13.708 0000:00:04.4 (8086 2021): Already using the vfio-pci driver 00:05:13.708 0000:00:04.3 (8086 2021): Already using the vfio-pci driver 00:05:13.708 0000:00:04.2 (8086 2021): Already using the vfio-pci driver 00:05:13.708 0000:00:04.1 (8086 2021): Already using the vfio-pci driver 00:05:13.708 0000:00:04.0 (8086 2021): Already using the vfio-pci driver 00:05:13.708 0000:80:04.7 (8086 2021): Already using the vfio-pci driver 00:05:13.708 0000:80:04.6 (8086 2021): Already using the vfio-pci driver 00:05:13.708 0000:80:04.5 (8086 2021): Already using the vfio-pci driver 00:05:13.708 0000:80:04.4 (8086 2021): Already using the vfio-pci driver 00:05:13.708 0000:80:04.3 (8086 2021): Already using the vfio-pci driver 00:05:13.708 0000:80:04.2 (8086 2021): Already using the vfio-pci driver 00:05:13.708 0000:80:04.1 (8086 2021): Already using the vfio-pci driver 00:05:13.708 0000:80:04.0 (8086 2021): Already using the vfio-pci driver 00:05:13.708 INFO: Requested 512 hugepages but 1024 already allocated on node0 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@204 -- # verify_nr_hugepages 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@89 -- # local node 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@90 -- # local sorted_t 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@91 -- # local sorted_s 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@92 -- # local surp 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@93 -- # local resv 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@94 -- # local anon 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@17 -- # local get=AnonHugePages 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@18 -- # local node= 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@19 -- # local var val 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.708 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78873952 kB' 'MemAvailable: 82172704 kB' 'Buffers: 12176 kB' 'Cached: 9445060 kB' 'SwapCached: 0 kB' 'Active: 6522560 kB' 'Inactive: 3456260 kB' 'Active(anon): 6128976 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 524392 kB' 'Mapped: 167256 kB' 'Shmem: 5607392 kB' 'KReclaimable: 205100 kB' 'Slab: 527480 kB' 'SReclaimable: 205100 kB' 'SUnreclaim: 322380 kB' 'KernelStack: 16384 kB' 'PageTables: 8324 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53486764 kB' 'Committed_AS: 7545204 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 201160 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.709 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # echo 0 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # return 0 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@97 -- # anon=0 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@18 -- # local node= 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@19 -- # local var val 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78876436 kB' 'MemAvailable: 82175188 kB' 'Buffers: 12176 kB' 'Cached: 9445060 kB' 'SwapCached: 0 kB' 'Active: 6522360 kB' 'Inactive: 3456260 kB' 'Active(anon): 6128776 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 524132 kB' 'Mapped: 167256 kB' 'Shmem: 5607392 kB' 'KReclaimable: 205100 kB' 'Slab: 527400 kB' 'SReclaimable: 205100 kB' 'SUnreclaim: 322300 kB' 'KernelStack: 16240 kB' 'PageTables: 7976 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53486764 kB' 'Committed_AS: 7545220 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 201032 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.710 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.711 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # echo 0 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # return 0 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@99 -- # surp=0 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@18 -- # local node= 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@19 -- # local var val 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78883572 kB' 'MemAvailable: 82182324 kB' 'Buffers: 12176 kB' 'Cached: 9445064 kB' 'SwapCached: 0 kB' 'Active: 6522276 kB' 'Inactive: 3456260 kB' 'Active(anon): 6128692 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 524464 kB' 'Mapped: 167180 kB' 'Shmem: 5607396 kB' 'KReclaimable: 205100 kB' 'Slab: 527424 kB' 'SReclaimable: 205100 kB' 'SUnreclaim: 322324 kB' 'KernelStack: 16288 kB' 'PageTables: 7956 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53486764 kB' 'Committed_AS: 7545244 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 201064 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.712 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.713 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # echo 0 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # return 0 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@100 -- # resv=0 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@102 -- # echo nr_hugepages=1024 00:05:13.714 nr_hugepages=1024 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:05:13.714 resv_hugepages=0 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:05:13.714 surplus_hugepages=0 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:05:13.714 anon_hugepages=0 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@107 -- # (( 1024 == nr_hugepages + surp + resv )) 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@109 -- # (( 1024 == nr_hugepages )) 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@17 -- # local get=HugePages_Total 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@18 -- # local node= 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@19 -- # local var val 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 92293472 kB' 'MemFree: 78891920 kB' 'MemAvailable: 82190672 kB' 'Buffers: 12176 kB' 'Cached: 9445064 kB' 'SwapCached: 0 kB' 'Active: 6523424 kB' 'Inactive: 3456260 kB' 'Active(anon): 6129840 kB' 'Inactive(anon): 0 kB' 'Active(file): 393584 kB' 'Inactive(file): 3456260 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 525636 kB' 'Mapped: 167180 kB' 'Shmem: 5607396 kB' 'KReclaimable: 205100 kB' 'Slab: 527360 kB' 'SReclaimable: 205100 kB' 'SUnreclaim: 322260 kB' 'KernelStack: 16688 kB' 'PageTables: 9140 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 53486764 kB' 'Committed_AS: 7545264 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 201160 kB' 'VmallocChunk: 0 kB' 'Percpu: 55040 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 820644 kB' 'DirectMap2M: 13535232 kB' 'DirectMap1G: 87031808 kB' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.714 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # echo 1024 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # return 0 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@110 -- # (( 1024 == nr_hugepages + surp + resv )) 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@112 -- # get_nodes 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@27 -- # local node 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=1024 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=0 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@32 -- # no_nodes=2 00:05:13.715 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@18 -- # local node=0 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@19 -- # local var val 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@20 -- # local mem_f mem 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 48116940 kB' 'MemFree: 36542580 kB' 'MemUsed: 11574360 kB' 'SwapCached: 0 kB' 'Active: 5342792 kB' 'Inactive: 3372048 kB' 'Active(anon): 5184896 kB' 'Inactive(anon): 0 kB' 'Active(file): 157896 kB' 'Inactive(file): 3372048 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8448736 kB' 'Mapped: 75568 kB' 'AnonPages: 269228 kB' 'Shmem: 4918792 kB' 'KernelStack: 8984 kB' 'PageTables: 4172 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 125360 kB' 'Slab: 329968 kB' 'SReclaimable: 125360 kB' 'SUnreclaim: 204608 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Surp: 0' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.716 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # echo 0 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # return 0 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@128 -- # echo 'node0=1024 expecting 1024' 00:05:13.717 node0=1024 expecting 1024 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@130 -- # [[ 1024 == \1\0\2\4 ]] 00:05:13.717 00:05:13.717 real 0m7.294s 00:05:13.717 user 0m2.632s 00:05:13.717 sys 0m4.788s 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:13.717 11:15:56 setup.sh.hugepages.no_shrink_alloc -- common/autotest_common.sh@10 -- # set +x 00:05:13.717 ************************************ 00:05:13.717 END TEST no_shrink_alloc 00:05:13.717 ************************************ 00:05:13.717 11:15:56 setup.sh.hugepages -- common/autotest_common.sh@1142 -- # return 0 00:05:13.717 11:15:56 setup.sh.hugepages -- setup/hugepages.sh@217 -- # clear_hp 00:05:13.717 11:15:56 setup.sh.hugepages -- setup/hugepages.sh@37 -- # local node hp 00:05:13.717 11:15:56 setup.sh.hugepages -- setup/hugepages.sh@39 -- # for node in "${!nodes_sys[@]}" 00:05:13.717 11:15:56 setup.sh.hugepages -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:05:13.717 11:15:56 setup.sh.hugepages -- setup/hugepages.sh@41 -- # echo 0 00:05:13.717 11:15:56 setup.sh.hugepages -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:05:13.717 11:15:56 setup.sh.hugepages -- setup/hugepages.sh@41 -- # echo 0 00:05:13.717 11:15:56 setup.sh.hugepages -- setup/hugepages.sh@39 -- # for node in "${!nodes_sys[@]}" 00:05:13.717 11:15:56 setup.sh.hugepages -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:05:13.717 11:15:56 setup.sh.hugepages -- setup/hugepages.sh@41 -- # echo 0 00:05:13.717 11:15:56 setup.sh.hugepages -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:05:13.717 11:15:56 setup.sh.hugepages -- setup/hugepages.sh@41 -- # echo 0 00:05:13.717 11:15:56 setup.sh.hugepages -- setup/hugepages.sh@45 -- # export CLEAR_HUGE=yes 00:05:13.717 11:15:56 setup.sh.hugepages -- setup/hugepages.sh@45 -- # CLEAR_HUGE=yes 00:05:13.717 00:05:13.717 real 0m29.472s 00:05:13.717 user 0m10.323s 00:05:13.717 sys 0m17.415s 00:05:13.717 11:15:56 setup.sh.hugepages -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:13.717 11:15:56 setup.sh.hugepages -- common/autotest_common.sh@10 -- # set +x 00:05:13.717 ************************************ 00:05:13.717 END TEST hugepages 00:05:13.717 ************************************ 00:05:13.717 11:15:56 setup.sh -- common/autotest_common.sh@1142 -- # return 0 00:05:13.717 11:15:56 setup.sh -- setup/test-setup.sh@14 -- # run_test driver /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/driver.sh 00:05:13.717 11:15:56 setup.sh -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:05:13.717 11:15:56 setup.sh -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:13.717 11:15:56 setup.sh -- common/autotest_common.sh@10 -- # set +x 00:05:13.717 ************************************ 00:05:13.717 START TEST driver 00:05:13.717 ************************************ 00:05:13.717 11:15:57 setup.sh.driver -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/driver.sh 00:05:13.717 * Looking for test storage... 00:05:13.717 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup 00:05:13.717 11:15:57 setup.sh.driver -- setup/driver.sh@68 -- # setup reset 00:05:13.717 11:15:57 setup.sh.driver -- setup/common.sh@9 -- # [[ reset == output ]] 00:05:13.717 11:15:57 setup.sh.driver -- setup/common.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh reset 00:05:18.987 11:16:02 setup.sh.driver -- setup/driver.sh@69 -- # run_test guess_driver guess_driver 00:05:18.987 11:16:02 setup.sh.driver -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:05:18.987 11:16:02 setup.sh.driver -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:18.987 11:16:02 setup.sh.driver -- common/autotest_common.sh@10 -- # set +x 00:05:18.987 ************************************ 00:05:18.987 START TEST guess_driver 00:05:18.987 ************************************ 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- common/autotest_common.sh@1123 -- # guess_driver 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@46 -- # local driver setup_driver marker 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@47 -- # local fail=0 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@49 -- # pick_driver 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@36 -- # vfio 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@21 -- # local iommu_grups 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@22 -- # local unsafe_vfio 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@24 -- # [[ -e /sys/module/vfio/parameters/enable_unsafe_noiommu_mode ]] 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@25 -- # unsafe_vfio=N 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@27 -- # iommu_groups=(/sys/kernel/iommu_groups/*) 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@29 -- # (( 216 > 0 )) 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@30 -- # is_driver vfio_pci 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@14 -- # mod vfio_pci 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@12 -- # dep vfio_pci 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@11 -- # modprobe --show-depends vfio_pci 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@12 -- # [[ insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/virt/lib/irqbypass.ko.xz 00:05:18.987 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/iommu/iommufd/iommufd.ko.xz 00:05:18.987 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/vfio/vfio.ko.xz 00:05:18.987 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/iommu/iommufd/iommufd.ko.xz 00:05:18.987 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/vfio/vfio.ko.xz 00:05:18.987 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/vfio/vfio_iommu_type1.ko.xz 00:05:18.987 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/vfio/pci/vfio-pci-core.ko.xz 00:05:18.987 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/vfio/pci/vfio-pci.ko.xz == *\.\k\o* ]] 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@30 -- # return 0 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@37 -- # echo vfio-pci 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@49 -- # driver=vfio-pci 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@51 -- # [[ vfio-pci == \N\o\ \v\a\l\i\d\ \d\r\i\v\e\r\ \f\o\u\n\d ]] 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@56 -- # echo 'Looking for driver=vfio-pci' 00:05:18.987 Looking for driver=vfio-pci 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/driver.sh@45 -- # setup output config 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/common.sh@9 -- # [[ output == output ]] 00:05:18.987 11:16:02 setup.sh.driver.guess_driver -- setup/common.sh@10 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh config 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ not == \-\> ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # continue 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ not == \-\> ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # continue 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:05:22.272 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:05:22.273 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:05:22.273 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:05:22.273 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:05:22.273 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:05:22.273 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:05:22.273 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:05:22.273 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:05:22.273 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:05:22.273 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:05:22.273 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:05:22.273 11:16:05 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:05:24.804 11:16:08 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:05:24.804 11:16:08 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:05:24.804 11:16:08 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:05:24.804 11:16:08 setup.sh.driver.guess_driver -- setup/driver.sh@64 -- # (( fail == 0 )) 00:05:24.804 11:16:08 setup.sh.driver.guess_driver -- setup/driver.sh@65 -- # setup reset 00:05:24.804 11:16:08 setup.sh.driver.guess_driver -- setup/common.sh@9 -- # [[ reset == output ]] 00:05:24.804 11:16:08 setup.sh.driver.guess_driver -- setup/common.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh reset 00:05:30.069 00:05:30.069 real 0m10.947s 00:05:30.069 user 0m2.596s 00:05:30.069 sys 0m5.168s 00:05:30.069 11:16:13 setup.sh.driver.guess_driver -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:30.069 11:16:13 setup.sh.driver.guess_driver -- common/autotest_common.sh@10 -- # set +x 00:05:30.069 ************************************ 00:05:30.069 END TEST guess_driver 00:05:30.069 ************************************ 00:05:30.069 11:16:13 setup.sh.driver -- common/autotest_common.sh@1142 -- # return 0 00:05:30.069 00:05:30.069 real 0m16.048s 00:05:30.069 user 0m4.057s 00:05:30.069 sys 0m8.029s 00:05:30.069 11:16:13 setup.sh.driver -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:30.069 11:16:13 setup.sh.driver -- common/autotest_common.sh@10 -- # set +x 00:05:30.069 ************************************ 00:05:30.069 END TEST driver 00:05:30.069 ************************************ 00:05:30.069 11:16:13 setup.sh -- common/autotest_common.sh@1142 -- # return 0 00:05:30.069 11:16:13 setup.sh -- setup/test-setup.sh@15 -- # run_test devices /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/devices.sh 00:05:30.069 11:16:13 setup.sh -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:05:30.069 11:16:13 setup.sh -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:30.069 11:16:13 setup.sh -- common/autotest_common.sh@10 -- # set +x 00:05:30.069 ************************************ 00:05:30.069 START TEST devices 00:05:30.069 ************************************ 00:05:30.069 11:16:13 setup.sh.devices -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/devices.sh 00:05:30.069 * Looking for test storage... 00:05:30.069 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup 00:05:30.069 11:16:13 setup.sh.devices -- setup/devices.sh@190 -- # trap cleanup EXIT 00:05:30.069 11:16:13 setup.sh.devices -- setup/devices.sh@192 -- # setup reset 00:05:30.069 11:16:13 setup.sh.devices -- setup/common.sh@9 -- # [[ reset == output ]] 00:05:30.069 11:16:13 setup.sh.devices -- setup/common.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh reset 00:05:34.259 11:16:17 setup.sh.devices -- setup/devices.sh@194 -- # get_zoned_devs 00:05:34.259 11:16:17 setup.sh.devices -- common/autotest_common.sh@1669 -- # zoned_devs=() 00:05:34.259 11:16:17 setup.sh.devices -- common/autotest_common.sh@1669 -- # local -gA zoned_devs 00:05:34.259 11:16:17 setup.sh.devices -- common/autotest_common.sh@1670 -- # local nvme bdf 00:05:34.259 11:16:17 setup.sh.devices -- common/autotest_common.sh@1672 -- # for nvme in /sys/block/nvme* 00:05:34.259 11:16:17 setup.sh.devices -- common/autotest_common.sh@1673 -- # is_block_zoned nvme0n1 00:05:34.259 11:16:17 setup.sh.devices -- common/autotest_common.sh@1662 -- # local device=nvme0n1 00:05:34.259 11:16:17 setup.sh.devices -- common/autotest_common.sh@1664 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:05:34.259 11:16:17 setup.sh.devices -- common/autotest_common.sh@1665 -- # [[ none != none ]] 00:05:34.259 11:16:17 setup.sh.devices -- setup/devices.sh@196 -- # blocks=() 00:05:34.259 11:16:17 setup.sh.devices -- setup/devices.sh@196 -- # declare -a blocks 00:05:34.259 11:16:17 setup.sh.devices -- setup/devices.sh@197 -- # blocks_to_pci=() 00:05:34.259 11:16:17 setup.sh.devices -- setup/devices.sh@197 -- # declare -A blocks_to_pci 00:05:34.259 11:16:17 setup.sh.devices -- setup/devices.sh@198 -- # min_disk_size=3221225472 00:05:34.259 11:16:17 setup.sh.devices -- setup/devices.sh@200 -- # for block in "/sys/block/nvme"!(*c*) 00:05:34.259 11:16:17 setup.sh.devices -- setup/devices.sh@201 -- # ctrl=nvme0n1 00:05:34.259 11:16:17 setup.sh.devices -- setup/devices.sh@201 -- # ctrl=nvme0 00:05:34.259 11:16:17 setup.sh.devices -- setup/devices.sh@202 -- # pci=0000:5e:00.0 00:05:34.259 11:16:17 setup.sh.devices -- setup/devices.sh@203 -- # [[ '' == *\0\0\0\0\:\5\e\:\0\0\.\0* ]] 00:05:34.259 11:16:17 setup.sh.devices -- setup/devices.sh@204 -- # block_in_use nvme0n1 00:05:34.259 11:16:17 setup.sh.devices -- scripts/common.sh@378 -- # local block=nvme0n1 pt 00:05:34.259 11:16:17 setup.sh.devices -- scripts/common.sh@387 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:05:34.259 No valid GPT data, bailing 00:05:34.259 11:16:17 setup.sh.devices -- scripts/common.sh@391 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:05:34.259 11:16:17 setup.sh.devices -- scripts/common.sh@391 -- # pt= 00:05:34.259 11:16:17 setup.sh.devices -- scripts/common.sh@392 -- # return 1 00:05:34.259 11:16:17 setup.sh.devices -- setup/devices.sh@204 -- # sec_size_to_bytes nvme0n1 00:05:34.259 11:16:17 setup.sh.devices -- setup/common.sh@76 -- # local dev=nvme0n1 00:05:34.259 11:16:17 setup.sh.devices -- setup/common.sh@78 -- # [[ -e /sys/block/nvme0n1 ]] 00:05:34.259 11:16:17 setup.sh.devices -- setup/common.sh@80 -- # echo 7681501126656 00:05:34.259 11:16:17 setup.sh.devices -- setup/devices.sh@204 -- # (( 7681501126656 >= min_disk_size )) 00:05:34.259 11:16:17 setup.sh.devices -- setup/devices.sh@205 -- # blocks+=("${block##*/}") 00:05:34.259 11:16:17 setup.sh.devices -- setup/devices.sh@206 -- # blocks_to_pci["${block##*/}"]=0000:5e:00.0 00:05:34.259 11:16:17 setup.sh.devices -- setup/devices.sh@209 -- # (( 1 > 0 )) 00:05:34.259 11:16:17 setup.sh.devices -- setup/devices.sh@211 -- # declare -r test_disk=nvme0n1 00:05:34.259 11:16:17 setup.sh.devices -- setup/devices.sh@213 -- # run_test nvme_mount nvme_mount 00:05:34.259 11:16:17 setup.sh.devices -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:05:34.259 11:16:17 setup.sh.devices -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:34.259 11:16:17 setup.sh.devices -- common/autotest_common.sh@10 -- # set +x 00:05:34.259 ************************************ 00:05:34.259 START TEST nvme_mount 00:05:34.259 ************************************ 00:05:34.259 11:16:17 setup.sh.devices.nvme_mount -- common/autotest_common.sh@1123 -- # nvme_mount 00:05:34.259 11:16:17 setup.sh.devices.nvme_mount -- setup/devices.sh@95 -- # nvme_disk=nvme0n1 00:05:34.259 11:16:17 setup.sh.devices.nvme_mount -- setup/devices.sh@96 -- # nvme_disk_p=nvme0n1p1 00:05:34.259 11:16:17 setup.sh.devices.nvme_mount -- setup/devices.sh@97 -- # nvme_mount=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount 00:05:34.260 11:16:17 setup.sh.devices.nvme_mount -- setup/devices.sh@98 -- # nvme_dummy_test_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:05:34.260 11:16:17 setup.sh.devices.nvme_mount -- setup/devices.sh@101 -- # partition_drive nvme0n1 1 00:05:34.260 11:16:17 setup.sh.devices.nvme_mount -- setup/common.sh@39 -- # local disk=nvme0n1 00:05:34.260 11:16:17 setup.sh.devices.nvme_mount -- setup/common.sh@40 -- # local part_no=1 00:05:34.260 11:16:17 setup.sh.devices.nvme_mount -- setup/common.sh@41 -- # local size=1073741824 00:05:34.260 11:16:17 setup.sh.devices.nvme_mount -- setup/common.sh@43 -- # local part part_start=0 part_end=0 00:05:34.260 11:16:17 setup.sh.devices.nvme_mount -- setup/common.sh@44 -- # parts=() 00:05:34.260 11:16:17 setup.sh.devices.nvme_mount -- setup/common.sh@44 -- # local parts 00:05:34.260 11:16:17 setup.sh.devices.nvme_mount -- setup/common.sh@46 -- # (( part = 1 )) 00:05:34.260 11:16:17 setup.sh.devices.nvme_mount -- setup/common.sh@46 -- # (( part <= part_no )) 00:05:34.260 11:16:17 setup.sh.devices.nvme_mount -- setup/common.sh@47 -- # parts+=("${disk}p$part") 00:05:34.260 11:16:17 setup.sh.devices.nvme_mount -- setup/common.sh@46 -- # (( part++ )) 00:05:34.260 11:16:17 setup.sh.devices.nvme_mount -- setup/common.sh@46 -- # (( part <= part_no )) 00:05:34.260 11:16:17 setup.sh.devices.nvme_mount -- setup/common.sh@51 -- # (( size /= 512 )) 00:05:34.260 11:16:17 setup.sh.devices.nvme_mount -- setup/common.sh@56 -- # sgdisk /dev/nvme0n1 --zap-all 00:05:34.260 11:16:17 setup.sh.devices.nvme_mount -- setup/common.sh@53 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/sync_dev_uevents.sh block/partition nvme0n1p1 00:05:35.199 Creating new GPT entries in memory. 00:05:35.200 GPT data structures destroyed! You may now partition the disk using fdisk or 00:05:35.200 other utilities. 00:05:35.200 11:16:18 setup.sh.devices.nvme_mount -- setup/common.sh@57 -- # (( part = 1 )) 00:05:35.200 11:16:18 setup.sh.devices.nvme_mount -- setup/common.sh@57 -- # (( part <= part_no )) 00:05:35.200 11:16:18 setup.sh.devices.nvme_mount -- setup/common.sh@58 -- # (( part_start = part_start == 0 ? 2048 : part_end + 1 )) 00:05:35.200 11:16:18 setup.sh.devices.nvme_mount -- setup/common.sh@59 -- # (( part_end = part_start + size - 1 )) 00:05:35.200 11:16:18 setup.sh.devices.nvme_mount -- setup/common.sh@60 -- # flock /dev/nvme0n1 sgdisk /dev/nvme0n1 --new=1:2048:2099199 00:05:36.133 Creating new GPT entries in memory. 00:05:36.133 The operation has completed successfully. 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/common.sh@57 -- # (( part++ )) 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/common.sh@57 -- # (( part <= part_no )) 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/common.sh@62 -- # wait 805667 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/devices.sh@102 -- # mkfs /dev/nvme0n1p1 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/common.sh@66 -- # local dev=/dev/nvme0n1p1 mount=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount size= 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/common.sh@68 -- # mkdir -p /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/common.sh@70 -- # [[ -e /dev/nvme0n1p1 ]] 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/common.sh@71 -- # mkfs.ext4 -qF /dev/nvme0n1p1 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/common.sh@72 -- # mount /dev/nvme0n1p1 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/devices.sh@105 -- # verify 0000:5e:00.0 nvme0n1:nvme0n1p1 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/devices.sh@48 -- # local dev=0000:5e:00.0 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/devices.sh@49 -- # local mounts=nvme0n1:nvme0n1p1 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/devices.sh@50 -- # local mount_point=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/devices.sh@51 -- # local test_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/devices.sh@53 -- # local found=0 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/devices.sh@55 -- # [[ -n /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount/test_nvme ]] 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/devices.sh@56 -- # : 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/devices.sh@59 -- # local pci status 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/devices.sh@47 -- # PCI_ALLOWED=0000:5e:00.0 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/devices.sh@47 -- # setup output config 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/common.sh@9 -- # [[ output == output ]] 00:05:36.133 11:16:19 setup.sh.devices.nvme_mount -- setup/common.sh@10 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh config 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:5e:00.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ Active devices: mount@nvme0n1:nvme0n1p1, so not binding PCI dev == *\A\c\t\i\v\e\ \d\e\v\i\c\e\s\:\ *\n\v\m\e\0\n\1\:\n\v\m\e\0\n\1\p\1* ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@63 -- # found=1 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.7 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.6 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.4 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.3 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.2 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.1 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:d7:05.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:d7:05.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:85:05.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:85:05.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.7 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.6 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.4 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.3 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.2 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.1 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@66 -- # (( found == 1 )) 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@68 -- # [[ -n /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@71 -- # mountpoint -q /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@73 -- # [[ -e /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount/test_nvme ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@74 -- # rm /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@110 -- # cleanup_nvme 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@20 -- # mountpoint -q /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@21 -- # umount /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@24 -- # [[ -b /dev/nvme0n1p1 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@25 -- # wipefs --all /dev/nvme0n1p1 00:05:40.350 /dev/nvme0n1p1: 2 bytes were erased at offset 0x00000438 (ext4): 53 ef 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@27 -- # [[ -b /dev/nvme0n1 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@28 -- # wipefs --all /dev/nvme0n1 00:05:40.350 /dev/nvme0n1: 8 bytes were erased at offset 0x00000200 (gpt): 45 46 49 20 50 41 52 54 00:05:40.350 /dev/nvme0n1: 8 bytes were erased at offset 0x6fc7d255e00 (gpt): 45 46 49 20 50 41 52 54 00:05:40.350 /dev/nvme0n1: 2 bytes were erased at offset 0x000001fe (PMBR): 55 aa 00:05:40.350 /dev/nvme0n1: calling ioctl to re-read partition table: Success 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@113 -- # mkfs /dev/nvme0n1 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount 1024M 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/common.sh@66 -- # local dev=/dev/nvme0n1 mount=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount size=1024M 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/common.sh@68 -- # mkdir -p /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/common.sh@70 -- # [[ -e /dev/nvme0n1 ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/common.sh@71 -- # mkfs.ext4 -qF /dev/nvme0n1 1024M 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/common.sh@72 -- # mount /dev/nvme0n1 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@116 -- # verify 0000:5e:00.0 nvme0n1:nvme0n1 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@48 -- # local dev=0000:5e:00.0 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@49 -- # local mounts=nvme0n1:nvme0n1 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@50 -- # local mount_point=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@51 -- # local test_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@53 -- # local found=0 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@55 -- # [[ -n /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount/test_nvme ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@56 -- # : 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@59 -- # local pci status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@47 -- # PCI_ALLOWED=0000:5e:00.0 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/devices.sh@47 -- # setup output config 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/common.sh@9 -- # [[ output == output ]] 00:05:40.350 11:16:23 setup.sh.devices.nvme_mount -- setup/common.sh@10 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh config 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:5e:00.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ Active devices: mount@nvme0n1:nvme0n1, so not binding PCI dev == *\A\c\t\i\v\e\ \d\e\v\i\c\e\s\:\ *\n\v\m\e\0\n\1\:\n\v\m\e\0\n\1* ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@63 -- # found=1 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.7 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.6 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.4 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.3 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.2 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.1 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:d7:05.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:d7:05.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:85:05.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:85:05.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.7 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.6 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.4 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.3 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.2 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.1 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:43.634 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.893 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@66 -- # (( found == 1 )) 00:05:43.893 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@68 -- # [[ -n /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount ]] 00:05:43.893 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@71 -- # mountpoint -q /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount 00:05:43.893 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@73 -- # [[ -e /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount/test_nvme ]] 00:05:43.893 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@74 -- # rm /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:05:43.893 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@123 -- # umount /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount 00:05:43.893 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@125 -- # verify 0000:5e:00.0 data@nvme0n1 '' '' 00:05:43.893 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@48 -- # local dev=0000:5e:00.0 00:05:43.893 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@49 -- # local mounts=data@nvme0n1 00:05:43.893 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@50 -- # local mount_point= 00:05:43.893 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@51 -- # local test_file= 00:05:43.893 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@53 -- # local found=0 00:05:43.893 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@55 -- # [[ -n '' ]] 00:05:43.893 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@59 -- # local pci status 00:05:43.893 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:43.893 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@47 -- # PCI_ALLOWED=0000:5e:00.0 00:05:43.893 11:16:27 setup.sh.devices.nvme_mount -- setup/devices.sh@47 -- # setup output config 00:05:43.893 11:16:27 setup.sh.devices.nvme_mount -- setup/common.sh@9 -- # [[ output == output ]] 00:05:43.893 11:16:27 setup.sh.devices.nvme_mount -- setup/common.sh@10 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh config 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:5e:00.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ Active devices: data@nvme0n1, so not binding PCI dev == *\A\c\t\i\v\e\ \d\e\v\i\c\e\s\:\ *\d\a\t\a\@\n\v\m\e\0\n\1* ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@63 -- # found=1 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.7 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.6 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.4 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.3 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.2 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.1 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:d7:05.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:d7:05.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:85:05.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:85:05.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.7 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.6 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.4 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.3 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.2 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.1 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@66 -- # (( found == 1 )) 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@68 -- # [[ -n '' ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@68 -- # return 0 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@128 -- # cleanup_nvme 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@20 -- # mountpoint -q /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@24 -- # [[ -b /dev/nvme0n1p1 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@27 -- # [[ -b /dev/nvme0n1 ]] 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- setup/devices.sh@28 -- # wipefs --all /dev/nvme0n1 00:05:48.076 /dev/nvme0n1: 2 bytes were erased at offset 0x00000438 (ext4): 53 ef 00:05:48.076 00:05:48.076 real 0m13.479s 00:05:48.076 user 0m3.977s 00:05:48.076 sys 0m7.452s 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:48.076 11:16:30 setup.sh.devices.nvme_mount -- common/autotest_common.sh@10 -- # set +x 00:05:48.076 ************************************ 00:05:48.076 END TEST nvme_mount 00:05:48.076 ************************************ 00:05:48.076 11:16:31 setup.sh.devices -- common/autotest_common.sh@1142 -- # return 0 00:05:48.076 11:16:31 setup.sh.devices -- setup/devices.sh@214 -- # run_test dm_mount dm_mount 00:05:48.076 11:16:31 setup.sh.devices -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:05:48.076 11:16:31 setup.sh.devices -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:48.076 11:16:31 setup.sh.devices -- common/autotest_common.sh@10 -- # set +x 00:05:48.076 ************************************ 00:05:48.076 START TEST dm_mount 00:05:48.076 ************************************ 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- common/autotest_common.sh@1123 -- # dm_mount 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/devices.sh@144 -- # pv=nvme0n1 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/devices.sh@145 -- # pv0=nvme0n1p1 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/devices.sh@146 -- # pv1=nvme0n1p2 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/devices.sh@148 -- # partition_drive nvme0n1 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/common.sh@39 -- # local disk=nvme0n1 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/common.sh@40 -- # local part_no=2 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/common.sh@41 -- # local size=1073741824 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/common.sh@43 -- # local part part_start=0 part_end=0 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/common.sh@44 -- # parts=() 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/common.sh@44 -- # local parts 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/common.sh@46 -- # (( part = 1 )) 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/common.sh@46 -- # (( part <= part_no )) 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/common.sh@47 -- # parts+=("${disk}p$part") 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/common.sh@46 -- # (( part++ )) 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/common.sh@46 -- # (( part <= part_no )) 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/common.sh@47 -- # parts+=("${disk}p$part") 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/common.sh@46 -- # (( part++ )) 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/common.sh@46 -- # (( part <= part_no )) 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/common.sh@51 -- # (( size /= 512 )) 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/common.sh@56 -- # sgdisk /dev/nvme0n1 --zap-all 00:05:48.076 11:16:31 setup.sh.devices.dm_mount -- setup/common.sh@53 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/sync_dev_uevents.sh block/partition nvme0n1p1 nvme0n1p2 00:05:48.643 Creating new GPT entries in memory. 00:05:48.643 GPT data structures destroyed! You may now partition the disk using fdisk or 00:05:48.643 other utilities. 00:05:48.643 11:16:32 setup.sh.devices.dm_mount -- setup/common.sh@57 -- # (( part = 1 )) 00:05:48.643 11:16:32 setup.sh.devices.dm_mount -- setup/common.sh@57 -- # (( part <= part_no )) 00:05:48.643 11:16:32 setup.sh.devices.dm_mount -- setup/common.sh@58 -- # (( part_start = part_start == 0 ? 2048 : part_end + 1 )) 00:05:48.643 11:16:32 setup.sh.devices.dm_mount -- setup/common.sh@59 -- # (( part_end = part_start + size - 1 )) 00:05:48.643 11:16:32 setup.sh.devices.dm_mount -- setup/common.sh@60 -- # flock /dev/nvme0n1 sgdisk /dev/nvme0n1 --new=1:2048:2099199 00:05:49.576 Creating new GPT entries in memory. 00:05:49.576 The operation has completed successfully. 00:05:49.576 11:16:33 setup.sh.devices.dm_mount -- setup/common.sh@57 -- # (( part++ )) 00:05:49.576 11:16:33 setup.sh.devices.dm_mount -- setup/common.sh@57 -- # (( part <= part_no )) 00:05:49.576 11:16:33 setup.sh.devices.dm_mount -- setup/common.sh@58 -- # (( part_start = part_start == 0 ? 2048 : part_end + 1 )) 00:05:49.576 11:16:33 setup.sh.devices.dm_mount -- setup/common.sh@59 -- # (( part_end = part_start + size - 1 )) 00:05:49.576 11:16:33 setup.sh.devices.dm_mount -- setup/common.sh@60 -- # flock /dev/nvme0n1 sgdisk /dev/nvme0n1 --new=2:2099200:4196351 00:05:50.946 The operation has completed successfully. 00:05:50.946 11:16:34 setup.sh.devices.dm_mount -- setup/common.sh@57 -- # (( part++ )) 00:05:50.946 11:16:34 setup.sh.devices.dm_mount -- setup/common.sh@57 -- # (( part <= part_no )) 00:05:50.946 11:16:34 setup.sh.devices.dm_mount -- setup/common.sh@62 -- # wait 809933 00:05:50.946 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@150 -- # dm_name=nvme_dm_test 00:05:50.946 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@151 -- # dm_mount=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/dm_mount 00:05:50.946 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@152 -- # dm_dummy_test_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/dm_mount/test_dm 00:05:50.946 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@155 -- # dmsetup create nvme_dm_test 00:05:50.946 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@160 -- # for t in {1..5} 00:05:50.946 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@161 -- # [[ -e /dev/mapper/nvme_dm_test ]] 00:05:50.946 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@161 -- # break 00:05:50.946 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@164 -- # [[ -e /dev/mapper/nvme_dm_test ]] 00:05:50.946 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@165 -- # readlink -f /dev/mapper/nvme_dm_test 00:05:50.946 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@165 -- # dm=/dev/dm-0 00:05:50.946 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@166 -- # dm=dm-0 00:05:50.946 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@168 -- # [[ -e /sys/class/block/nvme0n1p1/holders/dm-0 ]] 00:05:50.946 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@169 -- # [[ -e /sys/class/block/nvme0n1p2/holders/dm-0 ]] 00:05:50.946 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@171 -- # mkfs /dev/mapper/nvme_dm_test /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/dm_mount 00:05:50.946 11:16:34 setup.sh.devices.dm_mount -- setup/common.sh@66 -- # local dev=/dev/mapper/nvme_dm_test mount=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/dm_mount size= 00:05:50.947 11:16:34 setup.sh.devices.dm_mount -- setup/common.sh@68 -- # mkdir -p /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/dm_mount 00:05:50.947 11:16:34 setup.sh.devices.dm_mount -- setup/common.sh@70 -- # [[ -e /dev/mapper/nvme_dm_test ]] 00:05:50.947 11:16:34 setup.sh.devices.dm_mount -- setup/common.sh@71 -- # mkfs.ext4 -qF /dev/mapper/nvme_dm_test 00:05:50.947 11:16:34 setup.sh.devices.dm_mount -- setup/common.sh@72 -- # mount /dev/mapper/nvme_dm_test /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/dm_mount 00:05:50.947 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@174 -- # verify 0000:5e:00.0 nvme0n1:nvme_dm_test /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/dm_mount /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/dm_mount/test_dm 00:05:50.947 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@48 -- # local dev=0000:5e:00.0 00:05:50.947 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@49 -- # local mounts=nvme0n1:nvme_dm_test 00:05:50.947 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@50 -- # local mount_point=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/dm_mount 00:05:50.947 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@51 -- # local test_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/dm_mount/test_dm 00:05:50.947 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@53 -- # local found=0 00:05:50.947 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@55 -- # [[ -n /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/dm_mount/test_dm ]] 00:05:50.947 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@56 -- # : 00:05:50.947 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@59 -- # local pci status 00:05:50.947 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:50.947 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@47 -- # PCI_ALLOWED=0000:5e:00.0 00:05:50.947 11:16:34 setup.sh.devices.dm_mount -- setup/devices.sh@47 -- # setup output config 00:05:50.947 11:16:34 setup.sh.devices.dm_mount -- setup/common.sh@9 -- # [[ output == output ]] 00:05:50.947 11:16:34 setup.sh.devices.dm_mount -- setup/common.sh@10 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh config 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:5e:00.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ Active devices: holder@nvme0n1p1:dm-0,holder@nvme0n1p2:dm-0,mount@nvme0n1:nvme_dm_test, so not binding PCI dev == *\A\c\t\i\v\e\ \d\e\v\i\c\e\s\:\ *\n\v\m\e\0\n\1\:\n\v\m\e\_\d\m\_\t\e\s\t* ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@63 -- # found=1 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.7 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.6 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.4 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.3 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.2 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.1 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:d7:05.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:d7:05.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:85:05.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:85:05.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.7 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.6 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.4 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.3 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.2 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.1 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@66 -- # (( found == 1 )) 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@68 -- # [[ -n /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/dm_mount ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@71 -- # mountpoint -q /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/dm_mount 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@73 -- # [[ -e /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/dm_mount/test_dm ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@74 -- # rm /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/dm_mount/test_dm 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@182 -- # umount /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/dm_mount 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@184 -- # verify 0000:5e:00.0 holder@nvme0n1p1:dm-0,holder@nvme0n1p2:dm-0 '' '' 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@48 -- # local dev=0000:5e:00.0 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@49 -- # local mounts=holder@nvme0n1p1:dm-0,holder@nvme0n1p2:dm-0 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@50 -- # local mount_point= 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@51 -- # local test_file= 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@53 -- # local found=0 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@55 -- # [[ -n '' ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@59 -- # local pci status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@47 -- # PCI_ALLOWED=0000:5e:00.0 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/devices.sh@47 -- # setup output config 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/common.sh@9 -- # [[ output == output ]] 00:05:54.222 11:16:37 setup.sh.devices.dm_mount -- setup/common.sh@10 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh config 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:5e:00.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ Active devices: holder@nvme0n1p1:dm-0,holder@nvme0n1p2:dm-0, so not binding PCI dev == *\A\c\t\i\v\e\ \d\e\v\i\c\e\s\:\ *\h\o\l\d\e\r\@\n\v\m\e\0\n\1\p\1\:\d\m\-\0\,\h\o\l\d\e\r\@\n\v\m\e\0\n\1\p\2\:\d\m\-\0* ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@63 -- # found=1 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.7 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.6 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.4 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.3 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.2 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.1 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:d7:05.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:d7:05.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:85:05.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:85:05.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.7 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.6 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.4 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.3 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.2 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.1 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@66 -- # (( found == 1 )) 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@68 -- # [[ -n '' ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@68 -- # return 0 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@187 -- # cleanup_dm 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@33 -- # mountpoint -q /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/dm_mount 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@36 -- # [[ -L /dev/mapper/nvme_dm_test ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@37 -- # dmsetup remove --force nvme_dm_test 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@39 -- # [[ -b /dev/nvme0n1p1 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@40 -- # wipefs --all /dev/nvme0n1p1 00:05:58.403 /dev/nvme0n1p1: 2 bytes were erased at offset 0x00000438 (ext4): 53 ef 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@42 -- # [[ -b /dev/nvme0n1p2 ]] 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- setup/devices.sh@43 -- # wipefs --all /dev/nvme0n1p2 00:05:58.403 00:05:58.403 real 0m10.280s 00:05:58.403 user 0m2.514s 00:05:58.403 sys 0m4.770s 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:58.403 11:16:41 setup.sh.devices.dm_mount -- common/autotest_common.sh@10 -- # set +x 00:05:58.403 ************************************ 00:05:58.403 END TEST dm_mount 00:05:58.403 ************************************ 00:05:58.403 11:16:41 setup.sh.devices -- common/autotest_common.sh@1142 -- # return 0 00:05:58.403 11:16:41 setup.sh.devices -- setup/devices.sh@1 -- # cleanup 00:05:58.403 11:16:41 setup.sh.devices -- setup/devices.sh@11 -- # cleanup_nvme 00:05:58.403 11:16:41 setup.sh.devices -- setup/devices.sh@20 -- # mountpoint -q /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/nvme_mount 00:05:58.403 11:16:41 setup.sh.devices -- setup/devices.sh@24 -- # [[ -b /dev/nvme0n1p1 ]] 00:05:58.403 11:16:41 setup.sh.devices -- setup/devices.sh@25 -- # wipefs --all /dev/nvme0n1p1 00:05:58.403 11:16:41 setup.sh.devices -- setup/devices.sh@27 -- # [[ -b /dev/nvme0n1 ]] 00:05:58.403 11:16:41 setup.sh.devices -- setup/devices.sh@28 -- # wipefs --all /dev/nvme0n1 00:05:58.403 /dev/nvme0n1: 8 bytes were erased at offset 0x00000200 (gpt): 45 46 49 20 50 41 52 54 00:05:58.403 /dev/nvme0n1: 8 bytes were erased at offset 0x6fc7d255e00 (gpt): 45 46 49 20 50 41 52 54 00:05:58.403 /dev/nvme0n1: 2 bytes were erased at offset 0x000001fe (PMBR): 55 aa 00:05:58.403 /dev/nvme0n1: calling ioctl to re-read partition table: Success 00:05:58.403 11:16:41 setup.sh.devices -- setup/devices.sh@12 -- # cleanup_dm 00:05:58.403 11:16:41 setup.sh.devices -- setup/devices.sh@33 -- # mountpoint -q /var/jenkins/workspace/crypto-phy-autotest/spdk/test/setup/dm_mount 00:05:58.403 11:16:41 setup.sh.devices -- setup/devices.sh@36 -- # [[ -L /dev/mapper/nvme_dm_test ]] 00:05:58.403 11:16:41 setup.sh.devices -- setup/devices.sh@39 -- # [[ -b /dev/nvme0n1p1 ]] 00:05:58.403 11:16:41 setup.sh.devices -- setup/devices.sh@42 -- # [[ -b /dev/nvme0n1p2 ]] 00:05:58.403 11:16:41 setup.sh.devices -- setup/devices.sh@14 -- # [[ -b /dev/nvme0n1 ]] 00:05:58.403 11:16:41 setup.sh.devices -- setup/devices.sh@15 -- # wipefs --all /dev/nvme0n1 00:05:58.403 00:05:58.403 real 0m28.500s 00:05:58.403 user 0m8.176s 00:05:58.403 sys 0m15.206s 00:05:58.403 11:16:41 setup.sh.devices -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:58.403 11:16:41 setup.sh.devices -- common/autotest_common.sh@10 -- # set +x 00:05:58.403 ************************************ 00:05:58.403 END TEST devices 00:05:58.403 ************************************ 00:05:58.403 11:16:41 setup.sh -- common/autotest_common.sh@1142 -- # return 0 00:05:58.403 00:05:58.403 real 1m42.238s 00:05:58.403 user 0m31.276s 00:05:58.403 sys 0m57.263s 00:05:58.403 11:16:41 setup.sh -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:58.403 11:16:41 setup.sh -- common/autotest_common.sh@10 -- # set +x 00:05:58.403 ************************************ 00:05:58.403 END TEST setup.sh 00:05:58.403 ************************************ 00:05:58.403 11:16:41 -- common/autotest_common.sh@1142 -- # return 0 00:05:58.403 11:16:41 -- spdk/autotest.sh@128 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh status 00:06:01.704 0000:d7:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:d7:05.5 00:06:01.704 0000:85:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:85:05.5 00:06:01.704 Hugepages 00:06:01.704 node hugesize free / total 00:06:01.704 node0 1048576kB 0 / 0 00:06:01.704 node0 2048kB 1024 / 1024 00:06:01.704 node1 1048576kB 0 / 0 00:06:01.704 node1 2048kB 1024 / 1024 00:06:01.704 00:06:01.704 Type BDF Vendor Device NUMA Driver Device Block devices 00:06:01.704 I/OAT 0000:00:04.0 8086 2021 0 ioatdma - - 00:06:01.704 I/OAT 0000:00:04.1 8086 2021 0 ioatdma - - 00:06:01.964 I/OAT 0000:00:04.2 8086 2021 0 ioatdma - - 00:06:01.964 I/OAT 0000:00:04.3 8086 2021 0 ioatdma - - 00:06:01.964 I/OAT 0000:00:04.4 8086 2021 0 ioatdma - - 00:06:01.964 I/OAT 0000:00:04.5 8086 2021 0 ioatdma - - 00:06:01.964 I/OAT 0000:00:04.6 8086 2021 0 ioatdma - - 00:06:01.964 I/OAT 0000:00:04.7 8086 2021 0 ioatdma - - 00:06:01.964 NVMe 0000:5e:00.0 8086 0b60 0 nvme nvme0 nvme0n1 00:06:01.964 I/OAT 0000:80:04.0 8086 2021 1 ioatdma - - 00:06:01.964 I/OAT 0000:80:04.1 8086 2021 1 ioatdma - - 00:06:01.964 I/OAT 0000:80:04.2 8086 2021 1 ioatdma - - 00:06:01.964 I/OAT 0000:80:04.3 8086 2021 1 ioatdma - - 00:06:01.965 I/OAT 0000:80:04.4 8086 2021 1 ioatdma - - 00:06:01.965 I/OAT 0000:80:04.5 8086 2021 1 ioatdma - - 00:06:01.965 I/OAT 0000:80:04.6 8086 2021 1 ioatdma - - 00:06:01.965 I/OAT 0000:80:04.7 8086 2021 1 ioatdma - - 00:06:01.965 VMD 0000:85:05.5 8086 201d 1 vfio-pci - - 00:06:01.965 VMD 0000:d7:05.5 8086 201d 1 vfio-pci - - 00:06:01.965 11:16:45 -- spdk/autotest.sh@130 -- # uname -s 00:06:01.965 11:16:45 -- spdk/autotest.sh@130 -- # [[ Linux == Linux ]] 00:06:01.965 11:16:45 -- spdk/autotest.sh@132 -- # nvme_namespace_revert 00:06:01.965 11:16:45 -- common/autotest_common.sh@1531 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh 00:06:06.224 0000:d7:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:d7:05.5 00:06:06.224 0000:85:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:85:05.5 00:06:06.224 0000:00:04.7 (8086 2021): ioatdma -> vfio-pci 00:06:06.224 0000:00:04.6 (8086 2021): ioatdma -> vfio-pci 00:06:06.225 0000:00:04.5 (8086 2021): ioatdma -> vfio-pci 00:06:06.225 0000:00:04.4 (8086 2021): ioatdma -> vfio-pci 00:06:06.225 0000:00:04.3 (8086 2021): ioatdma -> vfio-pci 00:06:06.225 0000:00:04.2 (8086 2021): ioatdma -> vfio-pci 00:06:06.225 0000:00:04.1 (8086 2021): ioatdma -> vfio-pci 00:06:06.225 0000:00:04.0 (8086 2021): ioatdma -> vfio-pci 00:06:06.225 0000:80:04.7 (8086 2021): ioatdma -> vfio-pci 00:06:06.225 0000:80:04.6 (8086 2021): ioatdma -> vfio-pci 00:06:06.225 0000:80:04.5 (8086 2021): ioatdma -> vfio-pci 00:06:06.225 0000:80:04.4 (8086 2021): ioatdma -> vfio-pci 00:06:06.225 0000:80:04.3 (8086 2021): ioatdma -> vfio-pci 00:06:06.225 0000:80:04.2 (8086 2021): ioatdma -> vfio-pci 00:06:06.225 0000:80:04.1 (8086 2021): ioatdma -> vfio-pci 00:06:06.225 0000:80:04.0 (8086 2021): ioatdma -> vfio-pci 00:06:08.746 0000:5e:00.0 (8086 0b60): nvme -> vfio-pci 00:06:08.746 11:16:52 -- common/autotest_common.sh@1532 -- # sleep 1 00:06:09.677 11:16:53 -- common/autotest_common.sh@1533 -- # bdfs=() 00:06:09.677 11:16:53 -- common/autotest_common.sh@1533 -- # local bdfs 00:06:09.677 11:16:53 -- common/autotest_common.sh@1534 -- # bdfs=($(get_nvme_bdfs)) 00:06:09.677 11:16:53 -- common/autotest_common.sh@1534 -- # get_nvme_bdfs 00:06:09.677 11:16:53 -- common/autotest_common.sh@1513 -- # bdfs=() 00:06:09.677 11:16:53 -- common/autotest_common.sh@1513 -- # local bdfs 00:06:09.677 11:16:53 -- common/autotest_common.sh@1514 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:06:09.677 11:16:53 -- common/autotest_common.sh@1514 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/gen_nvme.sh 00:06:09.677 11:16:53 -- common/autotest_common.sh@1514 -- # jq -r '.config[].params.traddr' 00:06:09.677 11:16:53 -- common/autotest_common.sh@1515 -- # (( 1 == 0 )) 00:06:09.677 11:16:53 -- common/autotest_common.sh@1519 -- # printf '%s\n' 0000:5e:00.0 00:06:09.677 11:16:53 -- common/autotest_common.sh@1536 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh reset 00:06:12.947 0000:d7:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:d7:05.5 00:06:12.947 0000:85:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:85:05.5 00:06:12.947 Waiting for block devices as requested 00:06:12.947 0000:5e:00.0 (8086 0b60): vfio-pci -> nvme 00:06:12.947 0000:00:04.7 (8086 2021): vfio-pci -> ioatdma 00:06:13.203 0000:00:04.6 (8086 2021): vfio-pci -> ioatdma 00:06:13.204 0000:00:04.5 (8086 2021): vfio-pci -> ioatdma 00:06:13.204 0000:00:04.4 (8086 2021): vfio-pci -> ioatdma 00:06:13.460 0000:00:04.3 (8086 2021): vfio-pci -> ioatdma 00:06:13.460 0000:00:04.2 (8086 2021): vfio-pci -> ioatdma 00:06:13.460 0000:00:04.1 (8086 2021): vfio-pci -> ioatdma 00:06:13.718 0000:00:04.0 (8086 2021): vfio-pci -> ioatdma 00:06:13.718 0000:80:04.7 (8086 2021): vfio-pci -> ioatdma 00:06:13.718 0000:80:04.6 (8086 2021): vfio-pci -> ioatdma 00:06:13.974 0000:80:04.5 (8086 2021): vfio-pci -> ioatdma 00:06:13.974 0000:80:04.4 (8086 2021): vfio-pci -> ioatdma 00:06:13.974 0000:80:04.3 (8086 2021): vfio-pci -> ioatdma 00:06:14.231 0000:80:04.2 (8086 2021): vfio-pci -> ioatdma 00:06:14.232 0000:80:04.1 (8086 2021): vfio-pci -> ioatdma 00:06:14.232 0000:80:04.0 (8086 2021): vfio-pci -> ioatdma 00:06:14.489 11:16:57 -- common/autotest_common.sh@1538 -- # for bdf in "${bdfs[@]}" 00:06:14.489 11:16:57 -- common/autotest_common.sh@1539 -- # get_nvme_ctrlr_from_bdf 0000:5e:00.0 00:06:14.489 11:16:57 -- common/autotest_common.sh@1502 -- # readlink -f /sys/class/nvme/nvme0 00:06:14.489 11:16:57 -- common/autotest_common.sh@1502 -- # grep 0000:5e:00.0/nvme/nvme 00:06:14.489 11:16:57 -- common/autotest_common.sh@1502 -- # bdf_sysfs_path=/sys/devices/pci0000:5d/0000:5d:02.0/0000:5e:00.0/nvme/nvme0 00:06:14.489 11:16:57 -- common/autotest_common.sh@1503 -- # [[ -z /sys/devices/pci0000:5d/0000:5d:02.0/0000:5e:00.0/nvme/nvme0 ]] 00:06:14.489 11:16:57 -- common/autotest_common.sh@1507 -- # basename /sys/devices/pci0000:5d/0000:5d:02.0/0000:5e:00.0/nvme/nvme0 00:06:14.489 11:16:57 -- common/autotest_common.sh@1507 -- # printf '%s\n' nvme0 00:06:14.489 11:16:57 -- common/autotest_common.sh@1539 -- # nvme_ctrlr=/dev/nvme0 00:06:14.490 11:16:57 -- common/autotest_common.sh@1540 -- # [[ -z /dev/nvme0 ]] 00:06:14.490 11:16:57 -- common/autotest_common.sh@1545 -- # nvme id-ctrl /dev/nvme0 00:06:14.490 11:16:57 -- common/autotest_common.sh@1545 -- # cut -d: -f2 00:06:14.490 11:16:57 -- common/autotest_common.sh@1545 -- # grep oacs 00:06:14.490 11:16:57 -- common/autotest_common.sh@1545 -- # oacs=' 0x3f' 00:06:14.490 11:16:57 -- common/autotest_common.sh@1546 -- # oacs_ns_manage=8 00:06:14.490 11:16:57 -- common/autotest_common.sh@1548 -- # [[ 8 -ne 0 ]] 00:06:14.490 11:16:57 -- common/autotest_common.sh@1554 -- # nvme id-ctrl /dev/nvme0 00:06:14.490 11:16:57 -- common/autotest_common.sh@1554 -- # grep unvmcap 00:06:14.490 11:16:57 -- common/autotest_common.sh@1554 -- # cut -d: -f2 00:06:14.490 11:16:57 -- common/autotest_common.sh@1554 -- # unvmcap=' 0' 00:06:14.490 11:16:57 -- common/autotest_common.sh@1555 -- # [[ 0 -eq 0 ]] 00:06:14.490 11:16:57 -- common/autotest_common.sh@1557 -- # continue 00:06:14.490 11:16:57 -- spdk/autotest.sh@135 -- # timing_exit pre_cleanup 00:06:14.490 11:16:57 -- common/autotest_common.sh@728 -- # xtrace_disable 00:06:14.490 11:16:57 -- common/autotest_common.sh@10 -- # set +x 00:06:14.490 11:16:57 -- spdk/autotest.sh@138 -- # timing_enter afterboot 00:06:14.490 11:16:57 -- common/autotest_common.sh@722 -- # xtrace_disable 00:06:14.490 11:16:57 -- common/autotest_common.sh@10 -- # set +x 00:06:14.490 11:16:57 -- spdk/autotest.sh@139 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/setup.sh 00:06:17.764 0000:d7:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:d7:05.5 00:06:17.764 0000:85:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:85:05.5 00:06:17.764 0000:00:04.7 (8086 2021): ioatdma -> vfio-pci 00:06:17.764 0000:00:04.6 (8086 2021): ioatdma -> vfio-pci 00:06:17.764 0000:00:04.5 (8086 2021): ioatdma -> vfio-pci 00:06:17.764 0000:00:04.4 (8086 2021): ioatdma -> vfio-pci 00:06:17.764 0000:00:04.3 (8086 2021): ioatdma -> vfio-pci 00:06:17.764 0000:00:04.2 (8086 2021): ioatdma -> vfio-pci 00:06:17.764 0000:00:04.1 (8086 2021): ioatdma -> vfio-pci 00:06:17.764 0000:00:04.0 (8086 2021): ioatdma -> vfio-pci 00:06:17.764 0000:80:04.7 (8086 2021): ioatdma -> vfio-pci 00:06:17.764 0000:80:04.6 (8086 2021): ioatdma -> vfio-pci 00:06:18.021 0000:80:04.5 (8086 2021): ioatdma -> vfio-pci 00:06:18.021 0000:80:04.4 (8086 2021): ioatdma -> vfio-pci 00:06:18.021 0000:80:04.3 (8086 2021): ioatdma -> vfio-pci 00:06:18.021 0000:80:04.2 (8086 2021): ioatdma -> vfio-pci 00:06:18.021 0000:80:04.1 (8086 2021): ioatdma -> vfio-pci 00:06:18.021 0000:80:04.0 (8086 2021): ioatdma -> vfio-pci 00:06:20.545 0000:5e:00.0 (8086 0b60): nvme -> vfio-pci 00:06:20.545 11:17:03 -- spdk/autotest.sh@140 -- # timing_exit afterboot 00:06:20.545 11:17:03 -- common/autotest_common.sh@728 -- # xtrace_disable 00:06:20.545 11:17:03 -- common/autotest_common.sh@10 -- # set +x 00:06:20.545 11:17:04 -- spdk/autotest.sh@144 -- # opal_revert_cleanup 00:06:20.545 11:17:04 -- common/autotest_common.sh@1591 -- # mapfile -t bdfs 00:06:20.545 11:17:04 -- common/autotest_common.sh@1591 -- # get_nvme_bdfs_by_id 0x0a54 00:06:20.545 11:17:04 -- common/autotest_common.sh@1577 -- # bdfs=() 00:06:20.545 11:17:04 -- common/autotest_common.sh@1577 -- # local bdfs 00:06:20.545 11:17:04 -- common/autotest_common.sh@1579 -- # get_nvme_bdfs 00:06:20.545 11:17:04 -- common/autotest_common.sh@1513 -- # bdfs=() 00:06:20.545 11:17:04 -- common/autotest_common.sh@1513 -- # local bdfs 00:06:20.545 11:17:04 -- common/autotest_common.sh@1514 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:06:20.545 11:17:04 -- common/autotest_common.sh@1514 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/gen_nvme.sh 00:06:20.545 11:17:04 -- common/autotest_common.sh@1514 -- # jq -r '.config[].params.traddr' 00:06:20.545 11:17:04 -- common/autotest_common.sh@1515 -- # (( 1 == 0 )) 00:06:20.545 11:17:04 -- common/autotest_common.sh@1519 -- # printf '%s\n' 0000:5e:00.0 00:06:20.545 11:17:04 -- common/autotest_common.sh@1579 -- # for bdf in $(get_nvme_bdfs) 00:06:20.545 11:17:04 -- common/autotest_common.sh@1580 -- # cat /sys/bus/pci/devices/0000:5e:00.0/device 00:06:20.545 11:17:04 -- common/autotest_common.sh@1580 -- # device=0x0b60 00:06:20.545 11:17:04 -- common/autotest_common.sh@1581 -- # [[ 0x0b60 == \0\x\0\a\5\4 ]] 00:06:20.545 11:17:04 -- common/autotest_common.sh@1586 -- # printf '%s\n' 00:06:20.545 11:17:04 -- common/autotest_common.sh@1592 -- # [[ -z '' ]] 00:06:20.545 11:17:04 -- common/autotest_common.sh@1593 -- # return 0 00:06:20.545 11:17:04 -- spdk/autotest.sh@150 -- # '[' 0 -eq 1 ']' 00:06:20.545 11:17:04 -- spdk/autotest.sh@154 -- # '[' 1 -eq 1 ']' 00:06:20.545 11:17:04 -- spdk/autotest.sh@155 -- # [[ 1 -eq 1 ]] 00:06:20.545 11:17:04 -- spdk/autotest.sh@156 -- # [[ 0 -eq 1 ]] 00:06:20.545 11:17:04 -- spdk/autotest.sh@159 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/qat_setup.sh 00:06:21.475 Restarting all devices. 00:06:25.697 lstat() error: No such file or directory 00:06:25.697 QAT Error: No GENERAL section found 00:06:25.697 Failed to configure qat_dev0 00:06:25.697 lstat() error: No such file or directory 00:06:25.697 QAT Error: No GENERAL section found 00:06:25.697 Failed to configure qat_dev1 00:06:25.697 lstat() error: No such file or directory 00:06:25.697 QAT Error: No GENERAL section found 00:06:25.697 Failed to configure qat_dev2 00:06:25.697 enable sriov 00:06:25.697 Checking status of all devices. 00:06:25.697 There is 3 QAT acceleration device(s) in the system: 00:06:25.697 qat_dev0 - type: c6xx, inst_id: 0, node_id: 0, bsf: 0000:3d:00.0, #accel: 5 #engines: 10 state: down 00:06:25.697 qat_dev1 - type: c6xx, inst_id: 1, node_id: 0, bsf: 0000:3f:00.0, #accel: 5 #engines: 10 state: down 00:06:25.697 qat_dev2 - type: c6xx, inst_id: 2, node_id: 1, bsf: 0000:da:00.0, #accel: 5 #engines: 10 state: down 00:06:26.282 0000:3d:00.0 set to 16 VFs 00:06:27.653 0000:3f:00.0 set to 16 VFs 00:06:29.026 0000:da:00.0 set to 16 VFs 00:06:32.308 Properly configured the qat device with driver uio_pci_generic. 00:06:32.308 11:17:15 -- spdk/autotest.sh@162 -- # timing_enter lib 00:06:32.308 11:17:15 -- common/autotest_common.sh@722 -- # xtrace_disable 00:06:32.308 11:17:15 -- common/autotest_common.sh@10 -- # set +x 00:06:32.308 11:17:15 -- spdk/autotest.sh@164 -- # [[ 0 -eq 1 ]] 00:06:32.308 11:17:15 -- spdk/autotest.sh@168 -- # run_test env /var/jenkins/workspace/crypto-phy-autotest/spdk/test/env/env.sh 00:06:32.308 11:17:15 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:32.308 11:17:15 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:32.308 11:17:15 -- common/autotest_common.sh@10 -- # set +x 00:06:32.308 ************************************ 00:06:32.308 START TEST env 00:06:32.308 ************************************ 00:06:32.308 11:17:15 env -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/env/env.sh 00:06:32.308 * Looking for test storage... 00:06:32.308 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/env 00:06:32.308 11:17:15 env -- env/env.sh@10 -- # run_test env_memory /var/jenkins/workspace/crypto-phy-autotest/spdk/test/env/memory/memory_ut 00:06:32.308 11:17:15 env -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:32.308 11:17:15 env -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:32.308 11:17:15 env -- common/autotest_common.sh@10 -- # set +x 00:06:32.308 ************************************ 00:06:32.308 START TEST env_memory 00:06:32.308 ************************************ 00:06:32.308 11:17:15 env.env_memory -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/env/memory/memory_ut 00:06:32.308 00:06:32.308 00:06:32.308 CUnit - A unit testing framework for C - Version 2.1-3 00:06:32.308 http://cunit.sourceforge.net/ 00:06:32.308 00:06:32.308 00:06:32.308 Suite: memory 00:06:32.309 Test: alloc and free memory map ...[2024-07-15 11:17:15.744000] /var/jenkins/workspace/crypto-phy-autotest/spdk/lib/env_dpdk/memory.c: 283:spdk_mem_map_alloc: *ERROR*: Initial mem_map notify failed 00:06:32.309 passed 00:06:32.309 Test: mem map translation ...[2024-07-15 11:17:15.773296] /var/jenkins/workspace/crypto-phy-autotest/spdk/lib/env_dpdk/memory.c: 590:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=2097152 len=1234 00:06:32.309 [2024-07-15 11:17:15.773319] /var/jenkins/workspace/crypto-phy-autotest/spdk/lib/env_dpdk/memory.c: 590:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=1234 len=2097152 00:06:32.309 [2024-07-15 11:17:15.773375] /var/jenkins/workspace/crypto-phy-autotest/spdk/lib/env_dpdk/memory.c: 584:spdk_mem_map_set_translation: *ERROR*: invalid usermode virtual address 281474976710656 00:06:32.309 [2024-07-15 11:17:15.773388] /var/jenkins/workspace/crypto-phy-autotest/spdk/lib/env_dpdk/memory.c: 600:spdk_mem_map_set_translation: *ERROR*: could not get 0xffffffe00000 map 00:06:32.309 passed 00:06:32.309 Test: mem map registration ...[2024-07-15 11:17:15.831192] /var/jenkins/workspace/crypto-phy-autotest/spdk/lib/env_dpdk/memory.c: 346:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=0x200000 len=1234 00:06:32.309 [2024-07-15 11:17:15.831216] /var/jenkins/workspace/crypto-phy-autotest/spdk/lib/env_dpdk/memory.c: 346:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=0x4d2 len=2097152 00:06:32.309 passed 00:06:32.567 Test: mem map adjacent registrations ...passed 00:06:32.567 00:06:32.567 Run Summary: Type Total Ran Passed Failed Inactive 00:06:32.567 suites 1 1 n/a 0 0 00:06:32.567 tests 4 4 4 0 0 00:06:32.567 asserts 152 152 152 0 n/a 00:06:32.567 00:06:32.567 Elapsed time = 0.200 seconds 00:06:32.567 00:06:32.567 real 0m0.216s 00:06:32.567 user 0m0.201s 00:06:32.567 sys 0m0.014s 00:06:32.567 11:17:15 env.env_memory -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:32.567 11:17:15 env.env_memory -- common/autotest_common.sh@10 -- # set +x 00:06:32.567 ************************************ 00:06:32.567 END TEST env_memory 00:06:32.567 ************************************ 00:06:32.567 11:17:15 env -- common/autotest_common.sh@1142 -- # return 0 00:06:32.567 11:17:15 env -- env/env.sh@11 -- # run_test env_vtophys /var/jenkins/workspace/crypto-phy-autotest/spdk/test/env/vtophys/vtophys 00:06:32.567 11:17:15 env -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:32.567 11:17:15 env -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:32.567 11:17:15 env -- common/autotest_common.sh@10 -- # set +x 00:06:32.567 ************************************ 00:06:32.567 START TEST env_vtophys 00:06:32.567 ************************************ 00:06:32.567 11:17:15 env.env_vtophys -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/env/vtophys/vtophys 00:06:32.567 EAL: lib.eal log level changed from notice to debug 00:06:32.567 EAL: Detected lcore 0 as core 0 on socket 0 00:06:32.567 EAL: Detected lcore 1 as core 1 on socket 0 00:06:32.567 EAL: Detected lcore 2 as core 2 on socket 0 00:06:32.567 EAL: Detected lcore 3 as core 3 on socket 0 00:06:32.567 EAL: Detected lcore 4 as core 4 on socket 0 00:06:32.567 EAL: Detected lcore 5 as core 8 on socket 0 00:06:32.567 EAL: Detected lcore 6 as core 9 on socket 0 00:06:32.567 EAL: Detected lcore 7 as core 10 on socket 0 00:06:32.567 EAL: Detected lcore 8 as core 11 on socket 0 00:06:32.567 EAL: Detected lcore 9 as core 16 on socket 0 00:06:32.567 EAL: Detected lcore 10 as core 17 on socket 0 00:06:32.567 EAL: Detected lcore 11 as core 18 on socket 0 00:06:32.567 EAL: Detected lcore 12 as core 19 on socket 0 00:06:32.567 EAL: Detected lcore 13 as core 20 on socket 0 00:06:32.567 EAL: Detected lcore 14 as core 24 on socket 0 00:06:32.567 EAL: Detected lcore 15 as core 25 on socket 0 00:06:32.567 EAL: Detected lcore 16 as core 26 on socket 0 00:06:32.567 EAL: Detected lcore 17 as core 27 on socket 0 00:06:32.567 EAL: Detected lcore 18 as core 0 on socket 1 00:06:32.567 EAL: Detected lcore 19 as core 1 on socket 1 00:06:32.567 EAL: Detected lcore 20 as core 2 on socket 1 00:06:32.567 EAL: Detected lcore 21 as core 3 on socket 1 00:06:32.567 EAL: Detected lcore 22 as core 4 on socket 1 00:06:32.567 EAL: Detected lcore 23 as core 8 on socket 1 00:06:32.567 EAL: Detected lcore 24 as core 9 on socket 1 00:06:32.567 EAL: Detected lcore 25 as core 10 on socket 1 00:06:32.567 EAL: Detected lcore 26 as core 11 on socket 1 00:06:32.567 EAL: Detected lcore 27 as core 16 on socket 1 00:06:32.567 EAL: Detected lcore 28 as core 17 on socket 1 00:06:32.567 EAL: Detected lcore 29 as core 18 on socket 1 00:06:32.567 EAL: Detected lcore 30 as core 19 on socket 1 00:06:32.567 EAL: Detected lcore 31 as core 20 on socket 1 00:06:32.567 EAL: Detected lcore 32 as core 24 on socket 1 00:06:32.567 EAL: Detected lcore 33 as core 25 on socket 1 00:06:32.567 EAL: Detected lcore 34 as core 26 on socket 1 00:06:32.567 EAL: Detected lcore 35 as core 27 on socket 1 00:06:32.567 EAL: Detected lcore 36 as core 0 on socket 0 00:06:32.567 EAL: Detected lcore 37 as core 1 on socket 0 00:06:32.567 EAL: Detected lcore 38 as core 2 on socket 0 00:06:32.567 EAL: Detected lcore 39 as core 3 on socket 0 00:06:32.567 EAL: Detected lcore 40 as core 4 on socket 0 00:06:32.567 EAL: Detected lcore 41 as core 8 on socket 0 00:06:32.567 EAL: Detected lcore 42 as core 9 on socket 0 00:06:32.567 EAL: Detected lcore 43 as core 10 on socket 0 00:06:32.567 EAL: Detected lcore 44 as core 11 on socket 0 00:06:32.567 EAL: Detected lcore 45 as core 16 on socket 0 00:06:32.567 EAL: Detected lcore 46 as core 17 on socket 0 00:06:32.567 EAL: Detected lcore 47 as core 18 on socket 0 00:06:32.567 EAL: Detected lcore 48 as core 19 on socket 0 00:06:32.567 EAL: Detected lcore 49 as core 20 on socket 0 00:06:32.567 EAL: Detected lcore 50 as core 24 on socket 0 00:06:32.567 EAL: Detected lcore 51 as core 25 on socket 0 00:06:32.567 EAL: Detected lcore 52 as core 26 on socket 0 00:06:32.567 EAL: Detected lcore 53 as core 27 on socket 0 00:06:32.567 EAL: Detected lcore 54 as core 0 on socket 1 00:06:32.567 EAL: Detected lcore 55 as core 1 on socket 1 00:06:32.567 EAL: Detected lcore 56 as core 2 on socket 1 00:06:32.567 EAL: Detected lcore 57 as core 3 on socket 1 00:06:32.567 EAL: Detected lcore 58 as core 4 on socket 1 00:06:32.567 EAL: Detected lcore 59 as core 8 on socket 1 00:06:32.568 EAL: Detected lcore 60 as core 9 on socket 1 00:06:32.568 EAL: Detected lcore 61 as core 10 on socket 1 00:06:32.568 EAL: Detected lcore 62 as core 11 on socket 1 00:06:32.568 EAL: Detected lcore 63 as core 16 on socket 1 00:06:32.568 EAL: Detected lcore 64 as core 17 on socket 1 00:06:32.568 EAL: Detected lcore 65 as core 18 on socket 1 00:06:32.568 EAL: Detected lcore 66 as core 19 on socket 1 00:06:32.568 EAL: Detected lcore 67 as core 20 on socket 1 00:06:32.568 EAL: Detected lcore 68 as core 24 on socket 1 00:06:32.568 EAL: Detected lcore 69 as core 25 on socket 1 00:06:32.568 EAL: Detected lcore 70 as core 26 on socket 1 00:06:32.568 EAL: Detected lcore 71 as core 27 on socket 1 00:06:32.568 EAL: Maximum logical cores by configuration: 128 00:06:32.568 EAL: Detected CPU lcores: 72 00:06:32.568 EAL: Detected NUMA nodes: 2 00:06:32.568 EAL: Checking presence of .so 'librte_eal.so.24.1' 00:06:32.568 EAL: Detected shared linkage of DPDK 00:06:32.568 EAL: No shared files mode enabled, IPC will be disabled 00:06:32.568 EAL: No shared files mode enabled, IPC is disabled 00:06:32.568 EAL: PCI driver qat for device 0000:3d:01.0 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3d:01.1 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3d:01.2 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3d:01.3 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3d:01.4 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3d:01.5 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3d:01.6 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3d:01.7 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3d:02.0 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3d:02.1 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3d:02.2 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3d:02.3 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3d:02.4 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3d:02.5 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3d:02.6 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3d:02.7 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3f:01.0 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3f:01.1 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3f:01.2 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3f:01.3 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3f:01.4 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3f:01.5 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3f:01.6 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3f:01.7 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3f:02.0 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3f:02.1 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3f:02.2 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3f:02.3 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3f:02.4 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3f:02.5 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3f:02.6 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:3f:02.7 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:da:01.0 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:da:01.1 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:da:01.2 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:da:01.3 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:da:01.4 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:da:01.5 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:da:01.6 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:da:01.7 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:da:02.0 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:da:02.1 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:da:02.2 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:da:02.3 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:da:02.4 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:da:02.5 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:da:02.6 wants IOVA as 'PA' 00:06:32.568 EAL: PCI driver qat for device 0000:da:02.7 wants IOVA as 'PA' 00:06:32.568 EAL: Bus pci wants IOVA as 'PA' 00:06:32.568 EAL: Bus auxiliary wants IOVA as 'DC' 00:06:32.568 EAL: Bus vdev wants IOVA as 'DC' 00:06:32.568 EAL: Selected IOVA mode 'PA' 00:06:32.568 EAL: Probing VFIO support... 00:06:32.568 EAL: IOMMU type 1 (Type 1) is supported 00:06:32.568 EAL: IOMMU type 7 (sPAPR) is not supported 00:06:32.568 EAL: IOMMU type 8 (No-IOMMU) is not supported 00:06:32.568 EAL: VFIO support initialized 00:06:32.568 EAL: Ask a virtual area of 0x2e000 bytes 00:06:32.568 EAL: Virtual area found at 0x200000000000 (size = 0x2e000) 00:06:32.568 EAL: Setting up physically contiguous memory... 00:06:32.568 EAL: Setting maximum number of open files to 524288 00:06:32.568 EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 00:06:32.568 EAL: Detected memory type: socket_id:1 hugepage_sz:2097152 00:06:32.568 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 00:06:32.568 EAL: Ask a virtual area of 0x61000 bytes 00:06:32.568 EAL: Virtual area found at 0x20000002e000 (size = 0x61000) 00:06:32.568 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:06:32.568 EAL: Ask a virtual area of 0x400000000 bytes 00:06:32.568 EAL: Virtual area found at 0x200000200000 (size = 0x400000000) 00:06:32.568 EAL: VA reserved for memseg list at 0x200000200000, size 400000000 00:06:32.568 EAL: Ask a virtual area of 0x61000 bytes 00:06:32.568 EAL: Virtual area found at 0x200400200000 (size = 0x61000) 00:06:32.568 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:06:32.568 EAL: Ask a virtual area of 0x400000000 bytes 00:06:32.568 EAL: Virtual area found at 0x200400400000 (size = 0x400000000) 00:06:32.568 EAL: VA reserved for memseg list at 0x200400400000, size 400000000 00:06:32.568 EAL: Ask a virtual area of 0x61000 bytes 00:06:32.568 EAL: Virtual area found at 0x200800400000 (size = 0x61000) 00:06:32.568 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:06:32.568 EAL: Ask a virtual area of 0x400000000 bytes 00:06:32.568 EAL: Virtual area found at 0x200800600000 (size = 0x400000000) 00:06:32.568 EAL: VA reserved for memseg list at 0x200800600000, size 400000000 00:06:32.568 EAL: Ask a virtual area of 0x61000 bytes 00:06:32.568 EAL: Virtual area found at 0x200c00600000 (size = 0x61000) 00:06:32.568 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:06:32.568 EAL: Ask a virtual area of 0x400000000 bytes 00:06:32.568 EAL: Virtual area found at 0x200c00800000 (size = 0x400000000) 00:06:32.568 EAL: VA reserved for memseg list at 0x200c00800000, size 400000000 00:06:32.568 EAL: Creating 4 segment lists: n_segs:8192 socket_id:1 hugepage_sz:2097152 00:06:32.568 EAL: Ask a virtual area of 0x61000 bytes 00:06:32.568 EAL: Virtual area found at 0x201000800000 (size = 0x61000) 00:06:32.568 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:06:32.568 EAL: Ask a virtual area of 0x400000000 bytes 00:06:32.568 EAL: Virtual area found at 0x201000a00000 (size = 0x400000000) 00:06:32.568 EAL: VA reserved for memseg list at 0x201000a00000, size 400000000 00:06:32.568 EAL: Ask a virtual area of 0x61000 bytes 00:06:32.568 EAL: Virtual area found at 0x201400a00000 (size = 0x61000) 00:06:32.568 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:06:32.568 EAL: Ask a virtual area of 0x400000000 bytes 00:06:32.568 EAL: Virtual area found at 0x201400c00000 (size = 0x400000000) 00:06:32.568 EAL: VA reserved for memseg list at 0x201400c00000, size 400000000 00:06:32.568 EAL: Ask a virtual area of 0x61000 bytes 00:06:32.568 EAL: Virtual area found at 0x201800c00000 (size = 0x61000) 00:06:32.568 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:06:32.568 EAL: Ask a virtual area of 0x400000000 bytes 00:06:32.568 EAL: Virtual area found at 0x201800e00000 (size = 0x400000000) 00:06:32.568 EAL: VA reserved for memseg list at 0x201800e00000, size 400000000 00:06:32.568 EAL: Ask a virtual area of 0x61000 bytes 00:06:32.568 EAL: Virtual area found at 0x201c00e00000 (size = 0x61000) 00:06:32.568 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:06:32.568 EAL: Ask a virtual area of 0x400000000 bytes 00:06:32.568 EAL: Virtual area found at 0x201c01000000 (size = 0x400000000) 00:06:32.568 EAL: VA reserved for memseg list at 0x201c01000000, size 400000000 00:06:32.568 EAL: Hugepages will be freed exactly as allocated. 00:06:32.568 EAL: No shared files mode enabled, IPC is disabled 00:06:32.568 EAL: No shared files mode enabled, IPC is disabled 00:06:32.568 EAL: TSC frequency is ~2300000 KHz 00:06:32.568 EAL: Main lcore 0 is ready (tid=7fb304c31b00;cpuset=[0]) 00:06:32.568 EAL: Trying to obtain current memory policy. 00:06:32.568 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:32.568 EAL: Restoring previous memory policy: 0 00:06:32.568 EAL: request: mp_malloc_sync 00:06:32.568 EAL: No shared files mode enabled, IPC is disabled 00:06:32.568 EAL: Heap on socket 0 was expanded by 2MB 00:06:32.568 EAL: PCI device 0000:3d:01.0 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001000000 00:06:32.568 EAL: PCI memory mapped at 0x202001001000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.0 (socket 0) 00:06:32.568 EAL: PCI device 0000:3d:01.1 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001002000 00:06:32.568 EAL: PCI memory mapped at 0x202001003000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.1 (socket 0) 00:06:32.568 EAL: PCI device 0000:3d:01.2 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001004000 00:06:32.568 EAL: PCI memory mapped at 0x202001005000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.2 (socket 0) 00:06:32.568 EAL: PCI device 0000:3d:01.3 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001006000 00:06:32.568 EAL: PCI memory mapped at 0x202001007000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.3 (socket 0) 00:06:32.568 EAL: PCI device 0000:3d:01.4 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001008000 00:06:32.568 EAL: PCI memory mapped at 0x202001009000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.4 (socket 0) 00:06:32.568 EAL: PCI device 0000:3d:01.5 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x20200100a000 00:06:32.568 EAL: PCI memory mapped at 0x20200100b000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.5 (socket 0) 00:06:32.568 EAL: PCI device 0000:3d:01.6 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x20200100c000 00:06:32.568 EAL: PCI memory mapped at 0x20200100d000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.6 (socket 0) 00:06:32.568 EAL: PCI device 0000:3d:01.7 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x20200100e000 00:06:32.568 EAL: PCI memory mapped at 0x20200100f000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.7 (socket 0) 00:06:32.568 EAL: PCI device 0000:3d:02.0 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001010000 00:06:32.568 EAL: PCI memory mapped at 0x202001011000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.0 (socket 0) 00:06:32.568 EAL: PCI device 0000:3d:02.1 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001012000 00:06:32.568 EAL: PCI memory mapped at 0x202001013000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.1 (socket 0) 00:06:32.568 EAL: PCI device 0000:3d:02.2 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001014000 00:06:32.568 EAL: PCI memory mapped at 0x202001015000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.2 (socket 0) 00:06:32.568 EAL: PCI device 0000:3d:02.3 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001016000 00:06:32.568 EAL: PCI memory mapped at 0x202001017000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.3 (socket 0) 00:06:32.568 EAL: PCI device 0000:3d:02.4 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001018000 00:06:32.568 EAL: PCI memory mapped at 0x202001019000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.4 (socket 0) 00:06:32.568 EAL: PCI device 0000:3d:02.5 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x20200101a000 00:06:32.568 EAL: PCI memory mapped at 0x20200101b000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.5 (socket 0) 00:06:32.568 EAL: PCI device 0000:3d:02.6 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x20200101c000 00:06:32.568 EAL: PCI memory mapped at 0x20200101d000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.6 (socket 0) 00:06:32.568 EAL: PCI device 0000:3d:02.7 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x20200101e000 00:06:32.568 EAL: PCI memory mapped at 0x20200101f000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.7 (socket 0) 00:06:32.568 EAL: PCI device 0000:3f:01.0 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001020000 00:06:32.568 EAL: PCI memory mapped at 0x202001021000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.0 (socket 0) 00:06:32.568 EAL: PCI device 0000:3f:01.1 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001022000 00:06:32.568 EAL: PCI memory mapped at 0x202001023000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.1 (socket 0) 00:06:32.568 EAL: PCI device 0000:3f:01.2 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001024000 00:06:32.568 EAL: PCI memory mapped at 0x202001025000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.2 (socket 0) 00:06:32.568 EAL: PCI device 0000:3f:01.3 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001026000 00:06:32.568 EAL: PCI memory mapped at 0x202001027000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.3 (socket 0) 00:06:32.568 EAL: PCI device 0000:3f:01.4 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001028000 00:06:32.568 EAL: PCI memory mapped at 0x202001029000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.4 (socket 0) 00:06:32.568 EAL: PCI device 0000:3f:01.5 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x20200102a000 00:06:32.568 EAL: PCI memory mapped at 0x20200102b000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.5 (socket 0) 00:06:32.568 EAL: PCI device 0000:3f:01.6 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x20200102c000 00:06:32.568 EAL: PCI memory mapped at 0x20200102d000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.6 (socket 0) 00:06:32.568 EAL: PCI device 0000:3f:01.7 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x20200102e000 00:06:32.568 EAL: PCI memory mapped at 0x20200102f000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.7 (socket 0) 00:06:32.568 EAL: PCI device 0000:3f:02.0 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001030000 00:06:32.568 EAL: PCI memory mapped at 0x202001031000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.0 (socket 0) 00:06:32.568 EAL: PCI device 0000:3f:02.1 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001032000 00:06:32.568 EAL: PCI memory mapped at 0x202001033000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.1 (socket 0) 00:06:32.568 EAL: PCI device 0000:3f:02.2 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001034000 00:06:32.568 EAL: PCI memory mapped at 0x202001035000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.2 (socket 0) 00:06:32.568 EAL: PCI device 0000:3f:02.3 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001036000 00:06:32.568 EAL: PCI memory mapped at 0x202001037000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.3 (socket 0) 00:06:32.568 EAL: PCI device 0000:3f:02.4 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001038000 00:06:32.568 EAL: PCI memory mapped at 0x202001039000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.4 (socket 0) 00:06:32.568 EAL: PCI device 0000:3f:02.5 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x20200103a000 00:06:32.568 EAL: PCI memory mapped at 0x20200103b000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.5 (socket 0) 00:06:32.568 EAL: PCI device 0000:3f:02.6 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x20200103c000 00:06:32.568 EAL: PCI memory mapped at 0x20200103d000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.6 (socket 0) 00:06:32.568 EAL: PCI device 0000:3f:02.7 on NUMA socket 0 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x20200103e000 00:06:32.568 EAL: PCI memory mapped at 0x20200103f000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.7 (socket 0) 00:06:32.568 EAL: PCI device 0000:da:01.0 on NUMA socket 1 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001040000 00:06:32.568 EAL: PCI memory mapped at 0x202001041000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.0 (socket 1) 00:06:32.568 EAL: Trying to obtain current memory policy. 00:06:32.568 EAL: Setting policy MPOL_PREFERRED for socket 1 00:06:32.568 EAL: Restoring previous memory policy: 4 00:06:32.568 EAL: request: mp_malloc_sync 00:06:32.568 EAL: No shared files mode enabled, IPC is disabled 00:06:32.568 EAL: Heap on socket 1 was expanded by 2MB 00:06:32.568 EAL: PCI device 0000:da:01.1 on NUMA socket 1 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001042000 00:06:32.568 EAL: PCI memory mapped at 0x202001043000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.1 (socket 1) 00:06:32.568 EAL: PCI device 0000:da:01.2 on NUMA socket 1 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001044000 00:06:32.568 EAL: PCI memory mapped at 0x202001045000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.2 (socket 1) 00:06:32.568 EAL: PCI device 0000:da:01.3 on NUMA socket 1 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001046000 00:06:32.568 EAL: PCI memory mapped at 0x202001047000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.3 (socket 1) 00:06:32.568 EAL: PCI device 0000:da:01.4 on NUMA socket 1 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x202001048000 00:06:32.568 EAL: PCI memory mapped at 0x202001049000 00:06:32.568 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.4 (socket 1) 00:06:32.568 EAL: PCI device 0000:da:01.5 on NUMA socket 1 00:06:32.568 EAL: probe driver: 8086:37c9 qat 00:06:32.568 EAL: PCI memory mapped at 0x20200104a000 00:06:32.569 EAL: PCI memory mapped at 0x20200104b000 00:06:32.569 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.5 (socket 1) 00:06:32.569 EAL: PCI device 0000:da:01.6 on NUMA socket 1 00:06:32.569 EAL: probe driver: 8086:37c9 qat 00:06:32.569 EAL: PCI memory mapped at 0x20200104c000 00:06:32.569 EAL: PCI memory mapped at 0x20200104d000 00:06:32.569 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.6 (socket 1) 00:06:32.569 EAL: PCI device 0000:da:01.7 on NUMA socket 1 00:06:32.569 EAL: probe driver: 8086:37c9 qat 00:06:32.569 EAL: PCI memory mapped at 0x20200104e000 00:06:32.569 EAL: PCI memory mapped at 0x20200104f000 00:06:32.569 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.7 (socket 1) 00:06:32.569 EAL: PCI device 0000:da:02.0 on NUMA socket 1 00:06:32.569 EAL: probe driver: 8086:37c9 qat 00:06:32.569 EAL: PCI memory mapped at 0x202001050000 00:06:32.569 EAL: PCI memory mapped at 0x202001051000 00:06:32.569 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.0 (socket 1) 00:06:32.569 EAL: PCI device 0000:da:02.1 on NUMA socket 1 00:06:32.569 EAL: probe driver: 8086:37c9 qat 00:06:32.569 EAL: PCI memory mapped at 0x202001052000 00:06:32.569 EAL: PCI memory mapped at 0x202001053000 00:06:32.569 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.1 (socket 1) 00:06:32.569 EAL: PCI device 0000:da:02.2 on NUMA socket 1 00:06:32.569 EAL: probe driver: 8086:37c9 qat 00:06:32.569 EAL: PCI memory mapped at 0x202001054000 00:06:32.569 EAL: PCI memory mapped at 0x202001055000 00:06:32.569 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.2 (socket 1) 00:06:32.569 EAL: PCI device 0000:da:02.3 on NUMA socket 1 00:06:32.569 EAL: probe driver: 8086:37c9 qat 00:06:32.569 EAL: PCI memory mapped at 0x202001056000 00:06:32.569 EAL: PCI memory mapped at 0x202001057000 00:06:32.569 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.3 (socket 1) 00:06:32.569 EAL: PCI device 0000:da:02.4 on NUMA socket 1 00:06:32.569 EAL: probe driver: 8086:37c9 qat 00:06:32.569 EAL: PCI memory mapped at 0x202001058000 00:06:32.569 EAL: PCI memory mapped at 0x202001059000 00:06:32.569 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.4 (socket 1) 00:06:32.569 EAL: PCI device 0000:da:02.5 on NUMA socket 1 00:06:32.569 EAL: probe driver: 8086:37c9 qat 00:06:32.569 EAL: PCI memory mapped at 0x20200105a000 00:06:32.569 EAL: PCI memory mapped at 0x20200105b000 00:06:32.569 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.5 (socket 1) 00:06:32.569 EAL: PCI device 0000:da:02.6 on NUMA socket 1 00:06:32.569 EAL: probe driver: 8086:37c9 qat 00:06:32.569 EAL: PCI memory mapped at 0x20200105c000 00:06:32.569 EAL: PCI memory mapped at 0x20200105d000 00:06:32.569 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.6 (socket 1) 00:06:32.569 EAL: PCI device 0000:da:02.7 on NUMA socket 1 00:06:32.569 EAL: probe driver: 8086:37c9 qat 00:06:32.569 EAL: PCI memory mapped at 0x20200105e000 00:06:32.569 EAL: PCI memory mapped at 0x20200105f000 00:06:32.569 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.7 (socket 1) 00:06:32.569 EAL: No shared files mode enabled, IPC is disabled 00:06:32.569 EAL: No shared files mode enabled, IPC is disabled 00:06:32.569 EAL: No PCI address specified using 'addr=' in: bus=pci 00:06:32.569 EAL: Mem event callback 'spdk:(nil)' registered 00:06:32.569 00:06:32.569 00:06:32.569 CUnit - A unit testing framework for C - Version 2.1-3 00:06:32.569 http://cunit.sourceforge.net/ 00:06:32.569 00:06:32.569 00:06:32.569 Suite: components_suite 00:06:32.569 Test: vtophys_malloc_test ...passed 00:06:32.569 Test: vtophys_spdk_malloc_test ...EAL: Trying to obtain current memory policy. 00:06:32.569 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:32.569 EAL: Restoring previous memory policy: 4 00:06:32.569 EAL: Calling mem event callback 'spdk:(nil)' 00:06:32.569 EAL: request: mp_malloc_sync 00:06:32.569 EAL: No shared files mode enabled, IPC is disabled 00:06:32.569 EAL: Heap on socket 0 was expanded by 4MB 00:06:32.569 EAL: Calling mem event callback 'spdk:(nil)' 00:06:32.569 EAL: request: mp_malloc_sync 00:06:32.569 EAL: No shared files mode enabled, IPC is disabled 00:06:32.569 EAL: Heap on socket 0 was shrunk by 4MB 00:06:32.569 EAL: Trying to obtain current memory policy. 00:06:32.569 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:32.569 EAL: Restoring previous memory policy: 4 00:06:32.569 EAL: Calling mem event callback 'spdk:(nil)' 00:06:32.569 EAL: request: mp_malloc_sync 00:06:32.569 EAL: No shared files mode enabled, IPC is disabled 00:06:32.569 EAL: Heap on socket 0 was expanded by 6MB 00:06:32.569 EAL: Calling mem event callback 'spdk:(nil)' 00:06:32.569 EAL: request: mp_malloc_sync 00:06:32.569 EAL: No shared files mode enabled, IPC is disabled 00:06:32.569 EAL: Heap on socket 0 was shrunk by 6MB 00:06:32.569 EAL: Trying to obtain current memory policy. 00:06:32.569 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:32.826 EAL: Restoring previous memory policy: 4 00:06:32.826 EAL: Calling mem event callback 'spdk:(nil)' 00:06:32.826 EAL: request: mp_malloc_sync 00:06:32.826 EAL: No shared files mode enabled, IPC is disabled 00:06:32.826 EAL: Heap on socket 0 was expanded by 10MB 00:06:32.826 EAL: Calling mem event callback 'spdk:(nil)' 00:06:32.826 EAL: request: mp_malloc_sync 00:06:32.826 EAL: No shared files mode enabled, IPC is disabled 00:06:32.826 EAL: Heap on socket 0 was shrunk by 10MB 00:06:32.826 EAL: Trying to obtain current memory policy. 00:06:32.826 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:32.826 EAL: Restoring previous memory policy: 4 00:06:32.826 EAL: Calling mem event callback 'spdk:(nil)' 00:06:32.826 EAL: request: mp_malloc_sync 00:06:32.826 EAL: No shared files mode enabled, IPC is disabled 00:06:32.826 EAL: Heap on socket 0 was expanded by 18MB 00:06:32.826 EAL: Calling mem event callback 'spdk:(nil)' 00:06:32.826 EAL: request: mp_malloc_sync 00:06:32.826 EAL: No shared files mode enabled, IPC is disabled 00:06:32.826 EAL: Heap on socket 0 was shrunk by 18MB 00:06:32.826 EAL: Trying to obtain current memory policy. 00:06:32.826 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:32.826 EAL: Restoring previous memory policy: 4 00:06:32.826 EAL: Calling mem event callback 'spdk:(nil)' 00:06:32.826 EAL: request: mp_malloc_sync 00:06:32.826 EAL: No shared files mode enabled, IPC is disabled 00:06:32.826 EAL: Heap on socket 0 was expanded by 34MB 00:06:32.826 EAL: Calling mem event callback 'spdk:(nil)' 00:06:32.826 EAL: request: mp_malloc_sync 00:06:32.826 EAL: No shared files mode enabled, IPC is disabled 00:06:32.826 EAL: Heap on socket 0 was shrunk by 34MB 00:06:32.826 EAL: Trying to obtain current memory policy. 00:06:32.826 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:32.826 EAL: Restoring previous memory policy: 4 00:06:32.826 EAL: Calling mem event callback 'spdk:(nil)' 00:06:32.826 EAL: request: mp_malloc_sync 00:06:32.826 EAL: No shared files mode enabled, IPC is disabled 00:06:32.826 EAL: Heap on socket 0 was expanded by 66MB 00:06:32.826 EAL: Calling mem event callback 'spdk:(nil)' 00:06:32.826 EAL: request: mp_malloc_sync 00:06:32.826 EAL: No shared files mode enabled, IPC is disabled 00:06:32.826 EAL: Heap on socket 0 was shrunk by 66MB 00:06:32.826 EAL: Trying to obtain current memory policy. 00:06:32.826 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:32.826 EAL: Restoring previous memory policy: 4 00:06:32.826 EAL: Calling mem event callback 'spdk:(nil)' 00:06:32.826 EAL: request: mp_malloc_sync 00:06:32.826 EAL: No shared files mode enabled, IPC is disabled 00:06:32.826 EAL: Heap on socket 0 was expanded by 130MB 00:06:32.826 EAL: Calling mem event callback 'spdk:(nil)' 00:06:32.826 EAL: request: mp_malloc_sync 00:06:32.826 EAL: No shared files mode enabled, IPC is disabled 00:06:32.826 EAL: Heap on socket 0 was shrunk by 130MB 00:06:32.826 EAL: Trying to obtain current memory policy. 00:06:32.826 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:32.826 EAL: Restoring previous memory policy: 4 00:06:32.826 EAL: Calling mem event callback 'spdk:(nil)' 00:06:32.826 EAL: request: mp_malloc_sync 00:06:32.826 EAL: No shared files mode enabled, IPC is disabled 00:06:32.826 EAL: Heap on socket 0 was expanded by 258MB 00:06:32.826 EAL: Calling mem event callback 'spdk:(nil)' 00:06:33.082 EAL: request: mp_malloc_sync 00:06:33.082 EAL: No shared files mode enabled, IPC is disabled 00:06:33.082 EAL: Heap on socket 0 was shrunk by 258MB 00:06:33.082 EAL: Trying to obtain current memory policy. 00:06:33.082 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:33.082 EAL: Restoring previous memory policy: 4 00:06:33.082 EAL: Calling mem event callback 'spdk:(nil)' 00:06:33.082 EAL: request: mp_malloc_sync 00:06:33.082 EAL: No shared files mode enabled, IPC is disabled 00:06:33.082 EAL: Heap on socket 0 was expanded by 514MB 00:06:33.082 EAL: Calling mem event callback 'spdk:(nil)' 00:06:33.339 EAL: request: mp_malloc_sync 00:06:33.339 EAL: No shared files mode enabled, IPC is disabled 00:06:33.339 EAL: Heap on socket 0 was shrunk by 514MB 00:06:33.339 EAL: Trying to obtain current memory policy. 00:06:33.339 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:33.597 EAL: Restoring previous memory policy: 4 00:06:33.598 EAL: Calling mem event callback 'spdk:(nil)' 00:06:33.598 EAL: request: mp_malloc_sync 00:06:33.598 EAL: No shared files mode enabled, IPC is disabled 00:06:33.598 EAL: Heap on socket 0 was expanded by 1026MB 00:06:33.598 EAL: Calling mem event callback 'spdk:(nil)' 00:06:33.855 EAL: request: mp_malloc_sync 00:06:33.855 EAL: No shared files mode enabled, IPC is disabled 00:06:33.855 EAL: Heap on socket 0 was shrunk by 1026MB 00:06:33.855 passed 00:06:33.855 00:06:33.855 Run Summary: Type Total Ran Passed Failed Inactive 00:06:33.855 suites 1 1 n/a 0 0 00:06:33.855 tests 2 2 2 0 0 00:06:33.855 asserts 5582 5582 5582 0 n/a 00:06:33.855 00:06:33.855 Elapsed time = 1.176 seconds 00:06:33.855 EAL: No shared files mode enabled, IPC is disabled 00:06:33.855 EAL: No shared files mode enabled, IPC is disabled 00:06:33.855 EAL: No shared files mode enabled, IPC is disabled 00:06:33.855 00:06:33.855 real 0m1.375s 00:06:33.855 user 0m0.769s 00:06:33.855 sys 0m0.574s 00:06:33.855 11:17:17 env.env_vtophys -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:33.855 11:17:17 env.env_vtophys -- common/autotest_common.sh@10 -- # set +x 00:06:33.855 ************************************ 00:06:33.855 END TEST env_vtophys 00:06:33.855 ************************************ 00:06:33.855 11:17:17 env -- common/autotest_common.sh@1142 -- # return 0 00:06:33.855 11:17:17 env -- env/env.sh@12 -- # run_test env_pci /var/jenkins/workspace/crypto-phy-autotest/spdk/test/env/pci/pci_ut 00:06:33.855 11:17:17 env -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:33.855 11:17:17 env -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:33.855 11:17:17 env -- common/autotest_common.sh@10 -- # set +x 00:06:33.855 ************************************ 00:06:33.855 START TEST env_pci 00:06:33.855 ************************************ 00:06:33.855 11:17:17 env.env_pci -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/env/pci/pci_ut 00:06:34.114 00:06:34.114 00:06:34.114 CUnit - A unit testing framework for C - Version 2.1-3 00:06:34.114 http://cunit.sourceforge.net/ 00:06:34.114 00:06:34.114 00:06:34.114 Suite: pci 00:06:34.114 Test: pci_hook ...[2024-07-15 11:17:17.456399] /var/jenkins/workspace/crypto-phy-autotest/spdk/lib/env_dpdk/pci.c:1040:spdk_pci_device_claim: *ERROR*: Cannot create lock on device /var/tmp/spdk_pci_lock_10000:00:01.0, probably process 821063 has claimed it 00:06:34.114 EAL: Cannot find device (10000:00:01.0) 00:06:34.114 EAL: Failed to attach device on primary process 00:06:34.114 passed 00:06:34.114 00:06:34.114 Run Summary: Type Total Ran Passed Failed Inactive 00:06:34.114 suites 1 1 n/a 0 0 00:06:34.114 tests 1 1 1 0 0 00:06:34.114 asserts 25 25 25 0 n/a 00:06:34.114 00:06:34.114 Elapsed time = 0.029 seconds 00:06:34.114 00:06:34.114 real 0m0.044s 00:06:34.114 user 0m0.007s 00:06:34.114 sys 0m0.037s 00:06:34.114 11:17:17 env.env_pci -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:34.114 11:17:17 env.env_pci -- common/autotest_common.sh@10 -- # set +x 00:06:34.114 ************************************ 00:06:34.114 END TEST env_pci 00:06:34.114 ************************************ 00:06:34.114 11:17:17 env -- common/autotest_common.sh@1142 -- # return 0 00:06:34.114 11:17:17 env -- env/env.sh@14 -- # argv='-c 0x1 ' 00:06:34.114 11:17:17 env -- env/env.sh@15 -- # uname 00:06:34.114 11:17:17 env -- env/env.sh@15 -- # '[' Linux = Linux ']' 00:06:34.114 11:17:17 env -- env/env.sh@22 -- # argv+=--base-virtaddr=0x200000000000 00:06:34.114 11:17:17 env -- env/env.sh@24 -- # run_test env_dpdk_post_init /var/jenkins/workspace/crypto-phy-autotest/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:06:34.114 11:17:17 env -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:06:34.114 11:17:17 env -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:34.114 11:17:17 env -- common/autotest_common.sh@10 -- # set +x 00:06:34.114 ************************************ 00:06:34.114 START TEST env_dpdk_post_init 00:06:34.114 ************************************ 00:06:34.114 11:17:17 env.env_dpdk_post_init -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:06:34.114 EAL: Detected CPU lcores: 72 00:06:34.114 EAL: Detected NUMA nodes: 2 00:06:34.114 EAL: Detected shared linkage of DPDK 00:06:34.114 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:06:34.114 EAL: Selected IOVA mode 'PA' 00:06:34.114 EAL: VFIO support initialized 00:06:34.114 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.0 (socket 0) 00:06:34.114 CRYPTODEV: Creating cryptodev 0000:3d:01.0_qat_asym 00:06:34.114 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.0_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.114 CRYPTODEV: Creating cryptodev 0000:3d:01.0_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.0_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.1 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:01.1_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.1_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:01.1_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.1_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.2 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:01.2_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.2_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:01.2_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.2_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.3 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:01.3_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.3_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:01.3_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.3_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.4 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:01.4_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.4_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:01.4_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.4_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.5 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:01.5_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.5_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:01.5_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.5_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.6 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:01.6_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.6_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:01.6_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.6_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.7 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:01.7_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.7_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:01.7_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.7_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.0 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:02.0_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.0_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:02.0_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.0_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.1 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:02.1_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.1_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:02.1_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.1_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.2 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:02.2_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.2_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:02.2_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.2_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.3 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:02.3_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.3_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:02.3_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.3_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.4 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:02.4_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.4_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:02.4_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.4_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.5 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:02.5_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.5_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:02.5_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.5_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.6 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:02.6_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.6_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:02.6_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.6_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.7 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:02.7_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.7_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3d:02.7_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.7_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.0 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:01.0_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.0_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:01.0_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.0_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.1 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:01.1_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.1_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:01.1_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.1_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.2 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:01.2_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.2_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:01.2_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.2_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.3 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:01.3_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.3_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:01.3_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.3_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.4 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:01.4_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.4_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:01.4_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.4_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.5 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:01.5_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.5_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:01.5_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.5_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.6 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:01.6_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.6_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:01.6_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.6_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.7 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:01.7_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.7_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:01.7_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.7_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.0 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:02.0_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.0_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:02.0_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.0_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.1 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:02.1_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.1_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:02.1_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.1_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.2 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:02.2_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.2_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:02.2_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.2_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.3 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:02.3_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.3_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:02.3_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.3_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.4 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:02.4_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.4_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:02.4_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.4_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.5 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:02.5_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.5_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:02.5_qat_sym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.5_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.115 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.6 (socket 0) 00:06:34.115 CRYPTODEV: Creating cryptodev 0000:3f:02.6_qat_asym 00:06:34.115 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.6_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:3f:02.6_qat_sym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.6_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.116 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.7 (socket 0) 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:3f:02.7_qat_asym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.7_qat_asym,socket id: 0, max queue pairs: 0 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:3f:02.7_qat_sym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.7_qat_sym,socket id: 0, max queue pairs: 0 00:06:34.116 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.0 (socket 1) 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:01.0_qat_asym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:01.0_qat_asym,socket id: 1, max queue pairs: 0 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:01.0_qat_sym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:01.0_qat_sym,socket id: 1, max queue pairs: 0 00:06:34.116 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.1 (socket 1) 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:01.1_qat_asym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:01.1_qat_asym,socket id: 1, max queue pairs: 0 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:01.1_qat_sym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:01.1_qat_sym,socket id: 1, max queue pairs: 0 00:06:34.116 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.2 (socket 1) 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:01.2_qat_asym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:01.2_qat_asym,socket id: 1, max queue pairs: 0 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:01.2_qat_sym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:01.2_qat_sym,socket id: 1, max queue pairs: 0 00:06:34.116 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.3 (socket 1) 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:01.3_qat_asym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:01.3_qat_asym,socket id: 1, max queue pairs: 0 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:01.3_qat_sym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:01.3_qat_sym,socket id: 1, max queue pairs: 0 00:06:34.116 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.4 (socket 1) 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:01.4_qat_asym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:01.4_qat_asym,socket id: 1, max queue pairs: 0 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:01.4_qat_sym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:01.4_qat_sym,socket id: 1, max queue pairs: 0 00:06:34.116 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.5 (socket 1) 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:01.5_qat_asym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:01.5_qat_asym,socket id: 1, max queue pairs: 0 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:01.5_qat_sym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:01.5_qat_sym,socket id: 1, max queue pairs: 0 00:06:34.116 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.6 (socket 1) 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:01.6_qat_asym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:01.6_qat_asym,socket id: 1, max queue pairs: 0 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:01.6_qat_sym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:01.6_qat_sym,socket id: 1, max queue pairs: 0 00:06:34.116 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.7 (socket 1) 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:01.7_qat_asym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:01.7_qat_asym,socket id: 1, max queue pairs: 0 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:01.7_qat_sym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:01.7_qat_sym,socket id: 1, max queue pairs: 0 00:06:34.116 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.0 (socket 1) 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:02.0_qat_asym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:02.0_qat_asym,socket id: 1, max queue pairs: 0 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:02.0_qat_sym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:02.0_qat_sym,socket id: 1, max queue pairs: 0 00:06:34.116 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.1 (socket 1) 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:02.1_qat_asym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:02.1_qat_asym,socket id: 1, max queue pairs: 0 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:02.1_qat_sym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:02.1_qat_sym,socket id: 1, max queue pairs: 0 00:06:34.116 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.2 (socket 1) 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:02.2_qat_asym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:02.2_qat_asym,socket id: 1, max queue pairs: 0 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:02.2_qat_sym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:02.2_qat_sym,socket id: 1, max queue pairs: 0 00:06:34.116 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.3 (socket 1) 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:02.3_qat_asym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:02.3_qat_asym,socket id: 1, max queue pairs: 0 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:02.3_qat_sym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:02.3_qat_sym,socket id: 1, max queue pairs: 0 00:06:34.116 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.4 (socket 1) 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:02.4_qat_asym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:02.4_qat_asym,socket id: 1, max queue pairs: 0 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:02.4_qat_sym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:02.4_qat_sym,socket id: 1, max queue pairs: 0 00:06:34.116 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.5 (socket 1) 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:02.5_qat_asym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:02.5_qat_asym,socket id: 1, max queue pairs: 0 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:02.5_qat_sym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:02.5_qat_sym,socket id: 1, max queue pairs: 0 00:06:34.116 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.6 (socket 1) 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:02.6_qat_asym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:02.6_qat_asym,socket id: 1, max queue pairs: 0 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:02.6_qat_sym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:02.6_qat_sym,socket id: 1, max queue pairs: 0 00:06:34.116 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.7 (socket 1) 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:02.7_qat_asym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:02.7_qat_asym,socket id: 1, max queue pairs: 0 00:06:34.116 CRYPTODEV: Creating cryptodev 0000:da:02.7_qat_sym 00:06:34.116 CRYPTODEV: Initialisation parameters - name: 0000:da:02.7_qat_sym,socket id: 1, max queue pairs: 0 00:06:34.116 TELEMETRY: No legacy callbacks, legacy socket not created 00:06:34.375 EAL: Using IOMMU type 1 (Type 1) 00:06:34.375 EAL: Ignore mapping IO port bar(1) 00:06:34.375 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:00:04.0 (socket 0) 00:06:34.375 EAL: Ignore mapping IO port bar(1) 00:06:34.375 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:00:04.1 (socket 0) 00:06:34.375 EAL: Ignore mapping IO port bar(1) 00:06:34.375 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:00:04.2 (socket 0) 00:06:34.375 EAL: Ignore mapping IO port bar(1) 00:06:34.375 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:00:04.3 (socket 0) 00:06:34.375 EAL: Ignore mapping IO port bar(1) 00:06:34.375 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:00:04.4 (socket 0) 00:06:34.375 EAL: Ignore mapping IO port bar(1) 00:06:34.375 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:00:04.5 (socket 0) 00:06:34.375 EAL: Ignore mapping IO port bar(1) 00:06:34.375 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:00:04.6 (socket 0) 00:06:34.375 EAL: Ignore mapping IO port bar(1) 00:06:34.375 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:00:04.7 (socket 0) 00:06:34.633 EAL: Probe PCI driver: spdk_nvme (8086:0b60) device: 0000:5e:00.0 (socket 0) 00:06:34.633 EAL: Ignore mapping IO port bar(1) 00:06:34.633 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:80:04.0 (socket 1) 00:06:34.633 EAL: Ignore mapping IO port bar(1) 00:06:34.633 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:80:04.1 (socket 1) 00:06:34.633 EAL: Ignore mapping IO port bar(1) 00:06:34.633 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:80:04.2 (socket 1) 00:06:34.633 EAL: Ignore mapping IO port bar(1) 00:06:34.633 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:80:04.3 (socket 1) 00:06:34.633 EAL: Ignore mapping IO port bar(1) 00:06:34.633 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:80:04.4 (socket 1) 00:06:34.633 EAL: Ignore mapping IO port bar(1) 00:06:34.633 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:80:04.5 (socket 1) 00:06:34.633 EAL: Ignore mapping IO port bar(1) 00:06:34.633 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:80:04.6 (socket 1) 00:06:34.633 EAL: Ignore mapping IO port bar(1) 00:06:34.633 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:80:04.7 (socket 1) 00:06:34.633 EAL: Ignore mapping IO port bar(1) 00:06:34.633 EAL: Ignore mapping IO port bar(5) 00:06:34.633 EAL: Probe PCI driver: spdk_vmd (8086:201d) device: 0000:85:05.5 (socket 1) 00:06:34.890 EAL: Ignore mapping IO port bar(1) 00:06:34.890 EAL: Ignore mapping IO port bar(5) 00:06:34.890 EAL: Probe PCI driver: spdk_vmd (8086:201d) device: 0000:d7:05.5 (socket 1) 00:06:37.417 EAL: Releasing PCI mapped resource for 0000:5e:00.0 00:06:37.417 EAL: Calling pci_unmap_resource for 0000:5e:00.0 at 0x202001080000 00:06:37.677 Starting DPDK initialization... 00:06:37.677 Starting SPDK post initialization... 00:06:37.677 SPDK NVMe probe 00:06:37.677 Attaching to 0000:5e:00.0 00:06:37.677 Attached to 0000:5e:00.0 00:06:37.677 Cleaning up... 00:06:37.677 00:06:37.677 real 0m3.501s 00:06:37.677 user 0m2.384s 00:06:37.677 sys 0m0.675s 00:06:37.677 11:17:21 env.env_dpdk_post_init -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:37.677 11:17:21 env.env_dpdk_post_init -- common/autotest_common.sh@10 -- # set +x 00:06:37.677 ************************************ 00:06:37.677 END TEST env_dpdk_post_init 00:06:37.677 ************************************ 00:06:37.677 11:17:21 env -- common/autotest_common.sh@1142 -- # return 0 00:06:37.677 11:17:21 env -- env/env.sh@26 -- # uname 00:06:37.677 11:17:21 env -- env/env.sh@26 -- # '[' Linux = Linux ']' 00:06:37.677 11:17:21 env -- env/env.sh@29 -- # run_test env_mem_callbacks /var/jenkins/workspace/crypto-phy-autotest/spdk/test/env/mem_callbacks/mem_callbacks 00:06:37.677 11:17:21 env -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:37.677 11:17:21 env -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:37.677 11:17:21 env -- common/autotest_common.sh@10 -- # set +x 00:06:37.677 ************************************ 00:06:37.677 START TEST env_mem_callbacks 00:06:37.677 ************************************ 00:06:37.677 11:17:21 env.env_mem_callbacks -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/env/mem_callbacks/mem_callbacks 00:06:37.677 EAL: Detected CPU lcores: 72 00:06:37.677 EAL: Detected NUMA nodes: 2 00:06:37.677 EAL: Detected shared linkage of DPDK 00:06:37.677 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:06:37.677 EAL: Selected IOVA mode 'PA' 00:06:37.677 EAL: VFIO support initialized 00:06:37.677 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.0 (socket 0) 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:01.0_qat_asym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.0_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:01.0_qat_sym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.0_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.677 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.1 (socket 0) 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:01.1_qat_asym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.1_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:01.1_qat_sym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.1_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.677 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.2 (socket 0) 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:01.2_qat_asym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.2_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:01.2_qat_sym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.2_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.677 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.3 (socket 0) 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:01.3_qat_asym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.3_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:01.3_qat_sym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.3_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.677 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.4 (socket 0) 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:01.4_qat_asym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.4_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:01.4_qat_sym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.4_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.677 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.5 (socket 0) 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:01.5_qat_asym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.5_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:01.5_qat_sym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.5_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.677 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.6 (socket 0) 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:01.6_qat_asym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.6_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:01.6_qat_sym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.6_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.677 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:01.7 (socket 0) 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:01.7_qat_asym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.7_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:01.7_qat_sym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:01.7_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.677 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.0 (socket 0) 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:02.0_qat_asym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.0_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:02.0_qat_sym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.0_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.677 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.1 (socket 0) 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:02.1_qat_asym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.1_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:02.1_qat_sym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.1_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.677 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.2 (socket 0) 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:02.2_qat_asym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.2_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:02.2_qat_sym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.2_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.677 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.3 (socket 0) 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:02.3_qat_asym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.3_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:02.3_qat_sym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.3_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.677 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.4 (socket 0) 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:02.4_qat_asym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.4_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:02.4_qat_sym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.4_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.677 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.5 (socket 0) 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:02.5_qat_asym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.5_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:02.5_qat_sym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.5_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.677 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.6 (socket 0) 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:02.6_qat_asym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.6_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:02.6_qat_sym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.6_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.677 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3d:02.7 (socket 0) 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:02.7_qat_asym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.7_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3d:02.7_qat_sym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3d:02.7_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.677 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.0 (socket 0) 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3f:01.0_qat_asym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.0_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3f:01.0_qat_sym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.0_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.677 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.1 (socket 0) 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3f:01.1_qat_asym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.1_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3f:01.1_qat_sym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.1_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.677 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.2 (socket 0) 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3f:01.2_qat_asym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.2_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3f:01.2_qat_sym 00:06:37.677 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.2_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.677 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.3 (socket 0) 00:06:37.677 CRYPTODEV: Creating cryptodev 0000:3f:01.3_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.3_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:01.3_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.3_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.4 (socket 0) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:01.4_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.4_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:01.4_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.4_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.5 (socket 0) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:01.5_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.5_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:01.5_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.5_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.6 (socket 0) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:01.6_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.6_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:01.6_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.6_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:01.7 (socket 0) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:01.7_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.7_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:01.7_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:01.7_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.0 (socket 0) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:02.0_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.0_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:02.0_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.0_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.1 (socket 0) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:02.1_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.1_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:02.1_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.1_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.2 (socket 0) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:02.2_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.2_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:02.2_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.2_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.3 (socket 0) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:02.3_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.3_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:02.3_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.3_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.4 (socket 0) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:02.4_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.4_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:02.4_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.4_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.5 (socket 0) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:02.5_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.5_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:02.5_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.5_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.6 (socket 0) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:02.6_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.6_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:02.6_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.6_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:3f:02.7 (socket 0) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:02.7_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.7_qat_asym,socket id: 0, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:3f:02.7_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:3f:02.7_qat_sym,socket id: 0, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.0 (socket 1) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:01.0_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:01.0_qat_asym,socket id: 1, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:01.0_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:01.0_qat_sym,socket id: 1, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.1 (socket 1) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:01.1_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:01.1_qat_asym,socket id: 1, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:01.1_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:01.1_qat_sym,socket id: 1, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.2 (socket 1) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:01.2_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:01.2_qat_asym,socket id: 1, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:01.2_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:01.2_qat_sym,socket id: 1, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.3 (socket 1) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:01.3_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:01.3_qat_asym,socket id: 1, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:01.3_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:01.3_qat_sym,socket id: 1, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.4 (socket 1) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:01.4_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:01.4_qat_asym,socket id: 1, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:01.4_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:01.4_qat_sym,socket id: 1, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.5 (socket 1) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:01.5_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:01.5_qat_asym,socket id: 1, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:01.5_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:01.5_qat_sym,socket id: 1, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.6 (socket 1) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:01.6_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:01.6_qat_asym,socket id: 1, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:01.6_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:01.6_qat_sym,socket id: 1, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:01.7 (socket 1) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:01.7_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:01.7_qat_asym,socket id: 1, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:01.7_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:01.7_qat_sym,socket id: 1, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.0 (socket 1) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:02.0_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:02.0_qat_asym,socket id: 1, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:02.0_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:02.0_qat_sym,socket id: 1, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.1 (socket 1) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:02.1_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:02.1_qat_asym,socket id: 1, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:02.1_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:02.1_qat_sym,socket id: 1, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.2 (socket 1) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:02.2_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:02.2_qat_asym,socket id: 1, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:02.2_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:02.2_qat_sym,socket id: 1, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.3 (socket 1) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:02.3_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:02.3_qat_asym,socket id: 1, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:02.3_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:02.3_qat_sym,socket id: 1, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.4 (socket 1) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:02.4_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:02.4_qat_asym,socket id: 1, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:02.4_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:02.4_qat_sym,socket id: 1, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.5 (socket 1) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:02.5_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:02.5_qat_asym,socket id: 1, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:02.5_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:02.5_qat_sym,socket id: 1, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.6 (socket 1) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:02.6_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:02.6_qat_asym,socket id: 1, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:02.6_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:02.6_qat_sym,socket id: 1, max queue pairs: 0 00:06:37.678 EAL: Probe PCI driver: qat (8086:37c9) device: 0000:da:02.7 (socket 1) 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:02.7_qat_asym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:02.7_qat_asym,socket id: 1, max queue pairs: 0 00:06:37.678 CRYPTODEV: Creating cryptodev 0000:da:02.7_qat_sym 00:06:37.678 CRYPTODEV: Initialisation parameters - name: 0000:da:02.7_qat_sym,socket id: 1, max queue pairs: 0 00:06:37.678 TELEMETRY: No legacy callbacks, legacy socket not created 00:06:37.678 00:06:37.678 00:06:37.678 CUnit - A unit testing framework for C - Version 2.1-3 00:06:37.678 http://cunit.sourceforge.net/ 00:06:37.678 00:06:37.678 00:06:37.678 Suite: memory 00:06:37.678 Test: test ... 00:06:37.678 register 0x200000200000 2097152 00:06:37.678 register 0x201000a00000 2097152 00:06:37.679 malloc 3145728 00:06:37.679 register 0x200000400000 4194304 00:06:37.679 buf 0x200000500000 len 3145728 PASSED 00:06:37.679 malloc 64 00:06:37.679 buf 0x2000004fff40 len 64 PASSED 00:06:37.679 malloc 4194304 00:06:37.679 register 0x200000800000 6291456 00:06:37.679 buf 0x200000a00000 len 4194304 PASSED 00:06:37.679 free 0x200000500000 3145728 00:06:37.679 free 0x2000004fff40 64 00:06:37.679 unregister 0x200000400000 4194304 PASSED 00:06:37.679 free 0x200000a00000 4194304 00:06:37.679 unregister 0x200000800000 6291456 PASSED 00:06:37.679 malloc 8388608 00:06:37.679 register 0x200000400000 10485760 00:06:37.679 buf 0x200000600000 len 8388608 PASSED 00:06:37.679 free 0x200000600000 8388608 00:06:37.679 unregister 0x200000400000 10485760 PASSED 00:06:37.679 passed 00:06:37.679 00:06:37.679 Run Summary: Type Total Ran Passed Failed Inactive 00:06:37.679 suites 1 1 n/a 0 0 00:06:37.679 tests 1 1 1 0 0 00:06:37.679 asserts 16 16 16 0 n/a 00:06:37.679 00:06:37.679 Elapsed time = 0.006 seconds 00:06:37.679 00:06:37.679 real 0m0.080s 00:06:37.679 user 0m0.022s 00:06:37.679 sys 0m0.057s 00:06:37.679 11:17:21 env.env_mem_callbacks -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:37.679 11:17:21 env.env_mem_callbacks -- common/autotest_common.sh@10 -- # set +x 00:06:37.679 ************************************ 00:06:37.679 END TEST env_mem_callbacks 00:06:37.679 ************************************ 00:06:37.679 11:17:21 env -- common/autotest_common.sh@1142 -- # return 0 00:06:37.679 00:06:37.679 real 0m5.704s 00:06:37.679 user 0m3.551s 00:06:37.679 sys 0m1.718s 00:06:37.679 11:17:21 env -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:37.679 11:17:21 env -- common/autotest_common.sh@10 -- # set +x 00:06:37.679 ************************************ 00:06:37.679 END TEST env 00:06:37.679 ************************************ 00:06:37.938 11:17:21 -- common/autotest_common.sh@1142 -- # return 0 00:06:37.938 11:17:21 -- spdk/autotest.sh@169 -- # run_test rpc /var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc/rpc.sh 00:06:37.938 11:17:21 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:37.938 11:17:21 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:37.938 11:17:21 -- common/autotest_common.sh@10 -- # set +x 00:06:37.938 ************************************ 00:06:37.938 START TEST rpc 00:06:37.938 ************************************ 00:06:37.938 11:17:21 rpc -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc/rpc.sh 00:06:37.938 * Looking for test storage... 00:06:37.938 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc 00:06:37.938 11:17:21 rpc -- rpc/rpc.sh@65 -- # spdk_pid=821693 00:06:37.938 11:17:21 rpc -- rpc/rpc.sh@66 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:06:37.938 11:17:21 rpc -- rpc/rpc.sh@67 -- # waitforlisten 821693 00:06:37.938 11:17:21 rpc -- common/autotest_common.sh@829 -- # '[' -z 821693 ']' 00:06:37.938 11:17:21 rpc -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:37.938 11:17:21 rpc -- common/autotest_common.sh@834 -- # local max_retries=100 00:06:37.938 11:17:21 rpc -- rpc/rpc.sh@64 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt -e bdev 00:06:37.938 11:17:21 rpc -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:37.938 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:37.938 11:17:21 rpc -- common/autotest_common.sh@838 -- # xtrace_disable 00:06:37.938 11:17:21 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:37.938 [2024-07-15 11:17:21.506674] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:06:37.938 [2024-07-15 11:17:21.506751] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid821693 ] 00:06:38.196 [2024-07-15 11:17:21.637086] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:38.196 [2024-07-15 11:17:21.743526] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask bdev specified. 00:06:38.196 [2024-07-15 11:17:21.743576] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s spdk_tgt -p 821693' to capture a snapshot of events at runtime. 00:06:38.196 [2024-07-15 11:17:21.743590] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:06:38.196 [2024-07-15 11:17:21.743603] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:06:38.196 [2024-07-15 11:17:21.743614] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/spdk_tgt_trace.pid821693 for offline analysis/debug. 00:06:38.196 [2024-07-15 11:17:21.743654] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:06:39.127 11:17:22 rpc -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:06:39.127 11:17:22 rpc -- common/autotest_common.sh@862 -- # return 0 00:06:39.127 11:17:22 rpc -- rpc/rpc.sh@69 -- # export PYTHONPATH=:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc 00:06:39.127 11:17:22 rpc -- rpc/rpc.sh@69 -- # PYTHONPATH=:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc 00:06:39.127 11:17:22 rpc -- rpc/rpc.sh@72 -- # rpc=rpc_cmd 00:06:39.127 11:17:22 rpc -- rpc/rpc.sh@73 -- # run_test rpc_integrity rpc_integrity 00:06:39.127 11:17:22 rpc -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:39.127 11:17:22 rpc -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:39.127 11:17:22 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:39.127 ************************************ 00:06:39.127 START TEST rpc_integrity 00:06:39.127 ************************************ 00:06:39.127 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@1123 -- # rpc_integrity 00:06:39.127 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:06:39.127 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:39.127 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.127 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:39.127 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:06:39.127 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # jq length 00:06:39.127 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:06:39.127 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:06:39.127 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:39.127 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.127 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:39.127 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc0 00:06:39.127 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:06:39.127 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:39.127 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.127 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:39.127 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:06:39.127 { 00:06:39.127 "name": "Malloc0", 00:06:39.127 "aliases": [ 00:06:39.127 "ba936787-cd94-445c-beb6-d576c1efc4ea" 00:06:39.127 ], 00:06:39.127 "product_name": "Malloc disk", 00:06:39.127 "block_size": 512, 00:06:39.127 "num_blocks": 16384, 00:06:39.127 "uuid": "ba936787-cd94-445c-beb6-d576c1efc4ea", 00:06:39.127 "assigned_rate_limits": { 00:06:39.127 "rw_ios_per_sec": 0, 00:06:39.127 "rw_mbytes_per_sec": 0, 00:06:39.127 "r_mbytes_per_sec": 0, 00:06:39.127 "w_mbytes_per_sec": 0 00:06:39.127 }, 00:06:39.127 "claimed": false, 00:06:39.127 "zoned": false, 00:06:39.127 "supported_io_types": { 00:06:39.127 "read": true, 00:06:39.127 "write": true, 00:06:39.127 "unmap": true, 00:06:39.127 "flush": true, 00:06:39.127 "reset": true, 00:06:39.127 "nvme_admin": false, 00:06:39.127 "nvme_io": false, 00:06:39.127 "nvme_io_md": false, 00:06:39.127 "write_zeroes": true, 00:06:39.127 "zcopy": true, 00:06:39.127 "get_zone_info": false, 00:06:39.127 "zone_management": false, 00:06:39.127 "zone_append": false, 00:06:39.127 "compare": false, 00:06:39.127 "compare_and_write": false, 00:06:39.127 "abort": true, 00:06:39.127 "seek_hole": false, 00:06:39.127 "seek_data": false, 00:06:39.127 "copy": true, 00:06:39.127 "nvme_iov_md": false 00:06:39.127 }, 00:06:39.127 "memory_domains": [ 00:06:39.127 { 00:06:39.127 "dma_device_id": "system", 00:06:39.127 "dma_device_type": 1 00:06:39.127 }, 00:06:39.127 { 00:06:39.127 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:39.127 "dma_device_type": 2 00:06:39.127 } 00:06:39.127 ], 00:06:39.127 "driver_specific": {} 00:06:39.127 } 00:06:39.127 ]' 00:06:39.127 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # jq length 00:06:39.127 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:06:39.127 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc0 -p Passthru0 00:06:39.127 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:39.127 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.127 [2024-07-15 11:17:22.617690] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc0 00:06:39.127 [2024-07-15 11:17:22.617729] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:39.127 [2024-07-15 11:17:22.617750] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x26cbeb0 00:06:39.127 [2024-07-15 11:17:22.617763] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:39.127 [2024-07-15 11:17:22.619258] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:39.127 [2024-07-15 11:17:22.619287] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:06:39.127 Passthru0 00:06:39.127 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:39.127 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:06:39.127 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:39.127 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.127 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:39.127 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:06:39.127 { 00:06:39.127 "name": "Malloc0", 00:06:39.127 "aliases": [ 00:06:39.127 "ba936787-cd94-445c-beb6-d576c1efc4ea" 00:06:39.127 ], 00:06:39.127 "product_name": "Malloc disk", 00:06:39.127 "block_size": 512, 00:06:39.127 "num_blocks": 16384, 00:06:39.127 "uuid": "ba936787-cd94-445c-beb6-d576c1efc4ea", 00:06:39.127 "assigned_rate_limits": { 00:06:39.127 "rw_ios_per_sec": 0, 00:06:39.127 "rw_mbytes_per_sec": 0, 00:06:39.127 "r_mbytes_per_sec": 0, 00:06:39.127 "w_mbytes_per_sec": 0 00:06:39.127 }, 00:06:39.127 "claimed": true, 00:06:39.127 "claim_type": "exclusive_write", 00:06:39.127 "zoned": false, 00:06:39.127 "supported_io_types": { 00:06:39.127 "read": true, 00:06:39.127 "write": true, 00:06:39.127 "unmap": true, 00:06:39.127 "flush": true, 00:06:39.127 "reset": true, 00:06:39.127 "nvme_admin": false, 00:06:39.127 "nvme_io": false, 00:06:39.127 "nvme_io_md": false, 00:06:39.127 "write_zeroes": true, 00:06:39.127 "zcopy": true, 00:06:39.127 "get_zone_info": false, 00:06:39.127 "zone_management": false, 00:06:39.127 "zone_append": false, 00:06:39.127 "compare": false, 00:06:39.127 "compare_and_write": false, 00:06:39.127 "abort": true, 00:06:39.127 "seek_hole": false, 00:06:39.127 "seek_data": false, 00:06:39.127 "copy": true, 00:06:39.127 "nvme_iov_md": false 00:06:39.127 }, 00:06:39.127 "memory_domains": [ 00:06:39.127 { 00:06:39.127 "dma_device_id": "system", 00:06:39.127 "dma_device_type": 1 00:06:39.127 }, 00:06:39.127 { 00:06:39.127 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:39.127 "dma_device_type": 2 00:06:39.127 } 00:06:39.127 ], 00:06:39.127 "driver_specific": {} 00:06:39.127 }, 00:06:39.127 { 00:06:39.127 "name": "Passthru0", 00:06:39.127 "aliases": [ 00:06:39.127 "95be4915-8b56-5773-9e2c-a36eda9efc52" 00:06:39.127 ], 00:06:39.127 "product_name": "passthru", 00:06:39.127 "block_size": 512, 00:06:39.127 "num_blocks": 16384, 00:06:39.127 "uuid": "95be4915-8b56-5773-9e2c-a36eda9efc52", 00:06:39.127 "assigned_rate_limits": { 00:06:39.127 "rw_ios_per_sec": 0, 00:06:39.127 "rw_mbytes_per_sec": 0, 00:06:39.127 "r_mbytes_per_sec": 0, 00:06:39.127 "w_mbytes_per_sec": 0 00:06:39.127 }, 00:06:39.127 "claimed": false, 00:06:39.127 "zoned": false, 00:06:39.127 "supported_io_types": { 00:06:39.127 "read": true, 00:06:39.127 "write": true, 00:06:39.127 "unmap": true, 00:06:39.127 "flush": true, 00:06:39.127 "reset": true, 00:06:39.127 "nvme_admin": false, 00:06:39.127 "nvme_io": false, 00:06:39.127 "nvme_io_md": false, 00:06:39.127 "write_zeroes": true, 00:06:39.127 "zcopy": true, 00:06:39.127 "get_zone_info": false, 00:06:39.127 "zone_management": false, 00:06:39.127 "zone_append": false, 00:06:39.127 "compare": false, 00:06:39.127 "compare_and_write": false, 00:06:39.127 "abort": true, 00:06:39.127 "seek_hole": false, 00:06:39.127 "seek_data": false, 00:06:39.127 "copy": true, 00:06:39.127 "nvme_iov_md": false 00:06:39.127 }, 00:06:39.127 "memory_domains": [ 00:06:39.127 { 00:06:39.127 "dma_device_id": "system", 00:06:39.127 "dma_device_type": 1 00:06:39.127 }, 00:06:39.127 { 00:06:39.127 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:39.127 "dma_device_type": 2 00:06:39.127 } 00:06:39.127 ], 00:06:39.127 "driver_specific": { 00:06:39.127 "passthru": { 00:06:39.127 "name": "Passthru0", 00:06:39.127 "base_bdev_name": "Malloc0" 00:06:39.127 } 00:06:39.127 } 00:06:39.127 } 00:06:39.127 ]' 00:06:39.127 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # jq length 00:06:39.127 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:06:39.127 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:06:39.127 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:39.127 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.128 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:39.128 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc0 00:06:39.128 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:39.128 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.128 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:39.128 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:06:39.128 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:39.128 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.128 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:39.128 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:06:39.128 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # jq length 00:06:39.384 11:17:22 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:06:39.384 00:06:39.384 real 0m0.290s 00:06:39.384 user 0m0.184s 00:06:39.384 sys 0m0.041s 00:06:39.384 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:39.384 11:17:22 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.384 ************************************ 00:06:39.384 END TEST rpc_integrity 00:06:39.384 ************************************ 00:06:39.384 11:17:22 rpc -- common/autotest_common.sh@1142 -- # return 0 00:06:39.384 11:17:22 rpc -- rpc/rpc.sh@74 -- # run_test rpc_plugins rpc_plugins 00:06:39.384 11:17:22 rpc -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:39.384 11:17:22 rpc -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:39.384 11:17:22 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:39.384 ************************************ 00:06:39.384 START TEST rpc_plugins 00:06:39.384 ************************************ 00:06:39.384 11:17:22 rpc.rpc_plugins -- common/autotest_common.sh@1123 -- # rpc_plugins 00:06:39.384 11:17:22 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # rpc_cmd --plugin rpc_plugin create_malloc 00:06:39.384 11:17:22 rpc.rpc_plugins -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:39.384 11:17:22 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:06:39.384 11:17:22 rpc.rpc_plugins -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:39.384 11:17:22 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # malloc=Malloc1 00:06:39.384 11:17:22 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # rpc_cmd bdev_get_bdevs 00:06:39.384 11:17:22 rpc.rpc_plugins -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:39.384 11:17:22 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:06:39.384 11:17:22 rpc.rpc_plugins -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:39.384 11:17:22 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # bdevs='[ 00:06:39.384 { 00:06:39.384 "name": "Malloc1", 00:06:39.384 "aliases": [ 00:06:39.384 "13628d34-05cc-4119-b9f3-5ebb9bf75dfc" 00:06:39.384 ], 00:06:39.384 "product_name": "Malloc disk", 00:06:39.384 "block_size": 4096, 00:06:39.384 "num_blocks": 256, 00:06:39.384 "uuid": "13628d34-05cc-4119-b9f3-5ebb9bf75dfc", 00:06:39.384 "assigned_rate_limits": { 00:06:39.384 "rw_ios_per_sec": 0, 00:06:39.384 "rw_mbytes_per_sec": 0, 00:06:39.384 "r_mbytes_per_sec": 0, 00:06:39.384 "w_mbytes_per_sec": 0 00:06:39.384 }, 00:06:39.384 "claimed": false, 00:06:39.384 "zoned": false, 00:06:39.384 "supported_io_types": { 00:06:39.384 "read": true, 00:06:39.384 "write": true, 00:06:39.384 "unmap": true, 00:06:39.384 "flush": true, 00:06:39.384 "reset": true, 00:06:39.384 "nvme_admin": false, 00:06:39.384 "nvme_io": false, 00:06:39.384 "nvme_io_md": false, 00:06:39.384 "write_zeroes": true, 00:06:39.384 "zcopy": true, 00:06:39.384 "get_zone_info": false, 00:06:39.384 "zone_management": false, 00:06:39.384 "zone_append": false, 00:06:39.384 "compare": false, 00:06:39.384 "compare_and_write": false, 00:06:39.384 "abort": true, 00:06:39.384 "seek_hole": false, 00:06:39.384 "seek_data": false, 00:06:39.384 "copy": true, 00:06:39.384 "nvme_iov_md": false 00:06:39.384 }, 00:06:39.384 "memory_domains": [ 00:06:39.384 { 00:06:39.384 "dma_device_id": "system", 00:06:39.384 "dma_device_type": 1 00:06:39.384 }, 00:06:39.384 { 00:06:39.384 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:39.384 "dma_device_type": 2 00:06:39.384 } 00:06:39.384 ], 00:06:39.384 "driver_specific": {} 00:06:39.384 } 00:06:39.384 ]' 00:06:39.384 11:17:22 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # jq length 00:06:39.384 11:17:22 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # '[' 1 == 1 ']' 00:06:39.385 11:17:22 rpc.rpc_plugins -- rpc/rpc.sh@34 -- # rpc_cmd --plugin rpc_plugin delete_malloc Malloc1 00:06:39.385 11:17:22 rpc.rpc_plugins -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:39.385 11:17:22 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:06:39.385 11:17:22 rpc.rpc_plugins -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:39.385 11:17:22 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # rpc_cmd bdev_get_bdevs 00:06:39.385 11:17:22 rpc.rpc_plugins -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:39.385 11:17:22 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:06:39.385 11:17:22 rpc.rpc_plugins -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:39.385 11:17:22 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # bdevs='[]' 00:06:39.385 11:17:22 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # jq length 00:06:39.385 11:17:22 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # '[' 0 == 0 ']' 00:06:39.385 00:06:39.385 real 0m0.143s 00:06:39.385 user 0m0.086s 00:06:39.385 sys 0m0.021s 00:06:39.385 11:17:22 rpc.rpc_plugins -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:39.385 11:17:22 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:06:39.385 ************************************ 00:06:39.385 END TEST rpc_plugins 00:06:39.385 ************************************ 00:06:39.642 11:17:23 rpc -- common/autotest_common.sh@1142 -- # return 0 00:06:39.642 11:17:23 rpc -- rpc/rpc.sh@75 -- # run_test rpc_trace_cmd_test rpc_trace_cmd_test 00:06:39.642 11:17:23 rpc -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:39.642 11:17:23 rpc -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:39.642 11:17:23 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:39.642 ************************************ 00:06:39.642 START TEST rpc_trace_cmd_test 00:06:39.642 ************************************ 00:06:39.642 11:17:23 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1123 -- # rpc_trace_cmd_test 00:06:39.642 11:17:23 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@40 -- # local info 00:06:39.642 11:17:23 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # rpc_cmd trace_get_info 00:06:39.642 11:17:23 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:39.642 11:17:23 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:06:39.642 11:17:23 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:39.642 11:17:23 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # info='{ 00:06:39.642 "tpoint_shm_path": "/dev/shm/spdk_tgt_trace.pid821693", 00:06:39.642 "tpoint_group_mask": "0x8", 00:06:39.642 "iscsi_conn": { 00:06:39.642 "mask": "0x2", 00:06:39.642 "tpoint_mask": "0x0" 00:06:39.642 }, 00:06:39.642 "scsi": { 00:06:39.642 "mask": "0x4", 00:06:39.642 "tpoint_mask": "0x0" 00:06:39.642 }, 00:06:39.642 "bdev": { 00:06:39.642 "mask": "0x8", 00:06:39.642 "tpoint_mask": "0xffffffffffffffff" 00:06:39.642 }, 00:06:39.642 "nvmf_rdma": { 00:06:39.642 "mask": "0x10", 00:06:39.642 "tpoint_mask": "0x0" 00:06:39.642 }, 00:06:39.642 "nvmf_tcp": { 00:06:39.642 "mask": "0x20", 00:06:39.642 "tpoint_mask": "0x0" 00:06:39.642 }, 00:06:39.642 "ftl": { 00:06:39.642 "mask": "0x40", 00:06:39.642 "tpoint_mask": "0x0" 00:06:39.642 }, 00:06:39.642 "blobfs": { 00:06:39.642 "mask": "0x80", 00:06:39.642 "tpoint_mask": "0x0" 00:06:39.642 }, 00:06:39.642 "dsa": { 00:06:39.642 "mask": "0x200", 00:06:39.642 "tpoint_mask": "0x0" 00:06:39.642 }, 00:06:39.642 "thread": { 00:06:39.642 "mask": "0x400", 00:06:39.642 "tpoint_mask": "0x0" 00:06:39.642 }, 00:06:39.642 "nvme_pcie": { 00:06:39.642 "mask": "0x800", 00:06:39.642 "tpoint_mask": "0x0" 00:06:39.642 }, 00:06:39.642 "iaa": { 00:06:39.642 "mask": "0x1000", 00:06:39.642 "tpoint_mask": "0x0" 00:06:39.642 }, 00:06:39.642 "nvme_tcp": { 00:06:39.642 "mask": "0x2000", 00:06:39.642 "tpoint_mask": "0x0" 00:06:39.642 }, 00:06:39.642 "bdev_nvme": { 00:06:39.642 "mask": "0x4000", 00:06:39.642 "tpoint_mask": "0x0" 00:06:39.642 }, 00:06:39.642 "sock": { 00:06:39.642 "mask": "0x8000", 00:06:39.642 "tpoint_mask": "0x0" 00:06:39.642 } 00:06:39.642 }' 00:06:39.642 11:17:23 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # jq length 00:06:39.642 11:17:23 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # '[' 16 -gt 2 ']' 00:06:39.642 11:17:23 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # jq 'has("tpoint_group_mask")' 00:06:39.642 11:17:23 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # '[' true = true ']' 00:06:39.642 11:17:23 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # jq 'has("tpoint_shm_path")' 00:06:39.642 11:17:23 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # '[' true = true ']' 00:06:39.642 11:17:23 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # jq 'has("bdev")' 00:06:39.899 11:17:23 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # '[' true = true ']' 00:06:39.899 11:17:23 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # jq -r .bdev.tpoint_mask 00:06:39.899 11:17:23 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # '[' 0xffffffffffffffff '!=' 0x0 ']' 00:06:39.899 00:06:39.899 real 0m0.235s 00:06:39.899 user 0m0.190s 00:06:39.899 sys 0m0.038s 00:06:39.899 11:17:23 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:39.899 11:17:23 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:06:39.899 ************************************ 00:06:39.899 END TEST rpc_trace_cmd_test 00:06:39.899 ************************************ 00:06:39.899 11:17:23 rpc -- common/autotest_common.sh@1142 -- # return 0 00:06:39.899 11:17:23 rpc -- rpc/rpc.sh@76 -- # [[ 0 -eq 1 ]] 00:06:39.899 11:17:23 rpc -- rpc/rpc.sh@80 -- # rpc=rpc_cmd 00:06:39.899 11:17:23 rpc -- rpc/rpc.sh@81 -- # run_test rpc_daemon_integrity rpc_integrity 00:06:39.899 11:17:23 rpc -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:39.899 11:17:23 rpc -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:39.899 11:17:23 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:39.899 ************************************ 00:06:39.899 START TEST rpc_daemon_integrity 00:06:39.899 ************************************ 00:06:39.899 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1123 -- # rpc_integrity 00:06:39.899 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:06:39.899 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:39.899 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.899 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:39.899 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:06:39.899 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # jq length 00:06:39.899 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:06:39.899 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:06:39.899 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:39.899 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.899 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:39.899 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc2 00:06:39.899 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:06:39.899 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:39.899 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.899 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:39.899 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:06:39.899 { 00:06:39.899 "name": "Malloc2", 00:06:39.899 "aliases": [ 00:06:39.899 "25756656-7022-458f-a6b6-baed0f7c748f" 00:06:39.899 ], 00:06:39.899 "product_name": "Malloc disk", 00:06:39.899 "block_size": 512, 00:06:39.899 "num_blocks": 16384, 00:06:39.899 "uuid": "25756656-7022-458f-a6b6-baed0f7c748f", 00:06:39.899 "assigned_rate_limits": { 00:06:39.899 "rw_ios_per_sec": 0, 00:06:39.899 "rw_mbytes_per_sec": 0, 00:06:39.899 "r_mbytes_per_sec": 0, 00:06:39.899 "w_mbytes_per_sec": 0 00:06:39.899 }, 00:06:39.899 "claimed": false, 00:06:39.899 "zoned": false, 00:06:39.899 "supported_io_types": { 00:06:39.899 "read": true, 00:06:39.899 "write": true, 00:06:39.899 "unmap": true, 00:06:39.899 "flush": true, 00:06:39.899 "reset": true, 00:06:39.899 "nvme_admin": false, 00:06:39.899 "nvme_io": false, 00:06:39.899 "nvme_io_md": false, 00:06:39.899 "write_zeroes": true, 00:06:39.899 "zcopy": true, 00:06:39.899 "get_zone_info": false, 00:06:39.899 "zone_management": false, 00:06:39.899 "zone_append": false, 00:06:39.899 "compare": false, 00:06:39.899 "compare_and_write": false, 00:06:39.899 "abort": true, 00:06:39.899 "seek_hole": false, 00:06:39.899 "seek_data": false, 00:06:39.899 "copy": true, 00:06:39.899 "nvme_iov_md": false 00:06:39.899 }, 00:06:39.899 "memory_domains": [ 00:06:39.899 { 00:06:39.899 "dma_device_id": "system", 00:06:39.899 "dma_device_type": 1 00:06:39.899 }, 00:06:39.899 { 00:06:39.899 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:39.899 "dma_device_type": 2 00:06:39.899 } 00:06:39.899 ], 00:06:39.899 "driver_specific": {} 00:06:39.899 } 00:06:39.899 ]' 00:06:39.899 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # jq length 00:06:40.155 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:06:40.155 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc2 -p Passthru0 00:06:40.155 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:40.155 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:40.155 [2024-07-15 11:17:23.520267] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc2 00:06:40.155 [2024-07-15 11:17:23.520304] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:40.155 [2024-07-15 11:17:23.520329] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x26ccb20 00:06:40.155 [2024-07-15 11:17:23.520342] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:40.155 [2024-07-15 11:17:23.521700] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:40.155 [2024-07-15 11:17:23.521727] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:06:40.155 Passthru0 00:06:40.155 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:40.155 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:06:40.155 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:40.155 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:40.155 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:40.155 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:06:40.155 { 00:06:40.155 "name": "Malloc2", 00:06:40.155 "aliases": [ 00:06:40.155 "25756656-7022-458f-a6b6-baed0f7c748f" 00:06:40.155 ], 00:06:40.155 "product_name": "Malloc disk", 00:06:40.155 "block_size": 512, 00:06:40.155 "num_blocks": 16384, 00:06:40.155 "uuid": "25756656-7022-458f-a6b6-baed0f7c748f", 00:06:40.155 "assigned_rate_limits": { 00:06:40.155 "rw_ios_per_sec": 0, 00:06:40.155 "rw_mbytes_per_sec": 0, 00:06:40.155 "r_mbytes_per_sec": 0, 00:06:40.155 "w_mbytes_per_sec": 0 00:06:40.155 }, 00:06:40.155 "claimed": true, 00:06:40.155 "claim_type": "exclusive_write", 00:06:40.155 "zoned": false, 00:06:40.155 "supported_io_types": { 00:06:40.155 "read": true, 00:06:40.155 "write": true, 00:06:40.155 "unmap": true, 00:06:40.155 "flush": true, 00:06:40.155 "reset": true, 00:06:40.155 "nvme_admin": false, 00:06:40.155 "nvme_io": false, 00:06:40.155 "nvme_io_md": false, 00:06:40.155 "write_zeroes": true, 00:06:40.155 "zcopy": true, 00:06:40.155 "get_zone_info": false, 00:06:40.155 "zone_management": false, 00:06:40.155 "zone_append": false, 00:06:40.155 "compare": false, 00:06:40.155 "compare_and_write": false, 00:06:40.155 "abort": true, 00:06:40.155 "seek_hole": false, 00:06:40.155 "seek_data": false, 00:06:40.155 "copy": true, 00:06:40.155 "nvme_iov_md": false 00:06:40.155 }, 00:06:40.155 "memory_domains": [ 00:06:40.155 { 00:06:40.155 "dma_device_id": "system", 00:06:40.155 "dma_device_type": 1 00:06:40.155 }, 00:06:40.155 { 00:06:40.155 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:40.155 "dma_device_type": 2 00:06:40.155 } 00:06:40.155 ], 00:06:40.155 "driver_specific": {} 00:06:40.155 }, 00:06:40.155 { 00:06:40.155 "name": "Passthru0", 00:06:40.155 "aliases": [ 00:06:40.155 "2809ff31-4d28-5294-b875-f3886818031c" 00:06:40.155 ], 00:06:40.155 "product_name": "passthru", 00:06:40.155 "block_size": 512, 00:06:40.155 "num_blocks": 16384, 00:06:40.156 "uuid": "2809ff31-4d28-5294-b875-f3886818031c", 00:06:40.156 "assigned_rate_limits": { 00:06:40.156 "rw_ios_per_sec": 0, 00:06:40.156 "rw_mbytes_per_sec": 0, 00:06:40.156 "r_mbytes_per_sec": 0, 00:06:40.156 "w_mbytes_per_sec": 0 00:06:40.156 }, 00:06:40.156 "claimed": false, 00:06:40.156 "zoned": false, 00:06:40.156 "supported_io_types": { 00:06:40.156 "read": true, 00:06:40.156 "write": true, 00:06:40.156 "unmap": true, 00:06:40.156 "flush": true, 00:06:40.156 "reset": true, 00:06:40.156 "nvme_admin": false, 00:06:40.156 "nvme_io": false, 00:06:40.156 "nvme_io_md": false, 00:06:40.156 "write_zeroes": true, 00:06:40.156 "zcopy": true, 00:06:40.156 "get_zone_info": false, 00:06:40.156 "zone_management": false, 00:06:40.156 "zone_append": false, 00:06:40.156 "compare": false, 00:06:40.156 "compare_and_write": false, 00:06:40.156 "abort": true, 00:06:40.156 "seek_hole": false, 00:06:40.156 "seek_data": false, 00:06:40.156 "copy": true, 00:06:40.156 "nvme_iov_md": false 00:06:40.156 }, 00:06:40.156 "memory_domains": [ 00:06:40.156 { 00:06:40.156 "dma_device_id": "system", 00:06:40.156 "dma_device_type": 1 00:06:40.156 }, 00:06:40.156 { 00:06:40.156 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:40.156 "dma_device_type": 2 00:06:40.156 } 00:06:40.156 ], 00:06:40.156 "driver_specific": { 00:06:40.156 "passthru": { 00:06:40.156 "name": "Passthru0", 00:06:40.156 "base_bdev_name": "Malloc2" 00:06:40.156 } 00:06:40.156 } 00:06:40.156 } 00:06:40.156 ]' 00:06:40.156 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # jq length 00:06:40.156 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:06:40.156 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:06:40.156 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:40.156 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:40.156 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:40.156 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc2 00:06:40.156 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:40.156 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:40.156 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:40.156 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:06:40.156 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:40.156 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:40.156 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:40.156 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:06:40.156 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # jq length 00:06:40.156 11:17:23 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:06:40.156 00:06:40.156 real 0m0.306s 00:06:40.156 user 0m0.203s 00:06:40.156 sys 0m0.039s 00:06:40.156 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:40.156 11:17:23 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:40.156 ************************************ 00:06:40.156 END TEST rpc_daemon_integrity 00:06:40.156 ************************************ 00:06:40.156 11:17:23 rpc -- common/autotest_common.sh@1142 -- # return 0 00:06:40.156 11:17:23 rpc -- rpc/rpc.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:06:40.156 11:17:23 rpc -- rpc/rpc.sh@84 -- # killprocess 821693 00:06:40.156 11:17:23 rpc -- common/autotest_common.sh@948 -- # '[' -z 821693 ']' 00:06:40.156 11:17:23 rpc -- common/autotest_common.sh@952 -- # kill -0 821693 00:06:40.156 11:17:23 rpc -- common/autotest_common.sh@953 -- # uname 00:06:40.156 11:17:23 rpc -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:06:40.156 11:17:23 rpc -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 821693 00:06:40.412 11:17:23 rpc -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:06:40.412 11:17:23 rpc -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:06:40.412 11:17:23 rpc -- common/autotest_common.sh@966 -- # echo 'killing process with pid 821693' 00:06:40.412 killing process with pid 821693 00:06:40.412 11:17:23 rpc -- common/autotest_common.sh@967 -- # kill 821693 00:06:40.412 11:17:23 rpc -- common/autotest_common.sh@972 -- # wait 821693 00:06:40.670 00:06:40.670 real 0m2.797s 00:06:40.670 user 0m3.544s 00:06:40.670 sys 0m0.907s 00:06:40.670 11:17:24 rpc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:40.670 11:17:24 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:40.670 ************************************ 00:06:40.670 END TEST rpc 00:06:40.670 ************************************ 00:06:40.670 11:17:24 -- common/autotest_common.sh@1142 -- # return 0 00:06:40.670 11:17:24 -- spdk/autotest.sh@170 -- # run_test skip_rpc /var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc/skip_rpc.sh 00:06:40.670 11:17:24 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:40.670 11:17:24 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:40.670 11:17:24 -- common/autotest_common.sh@10 -- # set +x 00:06:40.670 ************************************ 00:06:40.670 START TEST skip_rpc 00:06:40.670 ************************************ 00:06:40.670 11:17:24 skip_rpc -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc/skip_rpc.sh 00:06:40.927 * Looking for test storage... 00:06:40.927 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc 00:06:40.927 11:17:24 skip_rpc -- rpc/skip_rpc.sh@11 -- # CONFIG_PATH=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc/config.json 00:06:40.927 11:17:24 skip_rpc -- rpc/skip_rpc.sh@12 -- # LOG_PATH=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc/log.txt 00:06:40.927 11:17:24 skip_rpc -- rpc/skip_rpc.sh@73 -- # run_test skip_rpc test_skip_rpc 00:06:40.927 11:17:24 skip_rpc -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:40.927 11:17:24 skip_rpc -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:40.927 11:17:24 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:40.927 ************************************ 00:06:40.927 START TEST skip_rpc 00:06:40.927 ************************************ 00:06:40.927 11:17:24 skip_rpc.skip_rpc -- common/autotest_common.sh@1123 -- # test_skip_rpc 00:06:40.927 11:17:24 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@16 -- # local spdk_pid=822251 00:06:40.927 11:17:24 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@18 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:06:40.927 11:17:24 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@19 -- # sleep 5 00:06:40.928 11:17:24 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 00:06:40.928 [2024-07-15 11:17:24.441558] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:06:40.928 [2024-07-15 11:17:24.441623] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid822251 ] 00:06:41.184 [2024-07-15 11:17:24.569532] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:41.184 [2024-07-15 11:17:24.669762] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@21 -- # NOT rpc_cmd spdk_get_version 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@648 -- # local es=0 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@650 -- # valid_exec_arg rpc_cmd spdk_get_version 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@636 -- # local arg=rpc_cmd 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@640 -- # type -t rpc_cmd 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@651 -- # rpc_cmd spdk_get_version 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@587 -- # [[ 1 == 0 ]] 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@651 -- # es=1 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@22 -- # trap - SIGINT SIGTERM EXIT 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@23 -- # killprocess 822251 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@948 -- # '[' -z 822251 ']' 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@952 -- # kill -0 822251 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@953 -- # uname 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 822251 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@966 -- # echo 'killing process with pid 822251' 00:06:46.470 killing process with pid 822251 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@967 -- # kill 822251 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@972 -- # wait 822251 00:06:46.470 00:06:46.470 real 0m5.444s 00:06:46.470 user 0m5.056s 00:06:46.470 sys 0m0.393s 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:46.470 11:17:29 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:46.470 ************************************ 00:06:46.470 END TEST skip_rpc 00:06:46.470 ************************************ 00:06:46.470 11:17:29 skip_rpc -- common/autotest_common.sh@1142 -- # return 0 00:06:46.470 11:17:29 skip_rpc -- rpc/skip_rpc.sh@74 -- # run_test skip_rpc_with_json test_skip_rpc_with_json 00:06:46.470 11:17:29 skip_rpc -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:46.470 11:17:29 skip_rpc -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:46.470 11:17:29 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:46.470 ************************************ 00:06:46.470 START TEST skip_rpc_with_json 00:06:46.470 ************************************ 00:06:46.470 11:17:29 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1123 -- # test_skip_rpc_with_json 00:06:46.470 11:17:29 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@44 -- # gen_json_config 00:06:46.470 11:17:29 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@28 -- # local spdk_pid=822989 00:06:46.470 11:17:29 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@30 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:06:46.470 11:17:29 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@31 -- # waitforlisten 822989 00:06:46.470 11:17:29 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@829 -- # '[' -z 822989 ']' 00:06:46.470 11:17:29 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:46.470 11:17:29 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@834 -- # local max_retries=100 00:06:46.470 11:17:29 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:46.470 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:46.470 11:17:29 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@838 -- # xtrace_disable 00:06:46.470 11:17:29 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@27 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:06:46.470 11:17:29 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:46.470 [2024-07-15 11:17:30.014243] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:06:46.470 [2024-07-15 11:17:30.014379] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid822989 ] 00:06:46.727 [2024-07-15 11:17:30.211320] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:46.727 [2024-07-15 11:17:30.312228] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:06:47.291 11:17:30 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:06:47.291 11:17:30 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@862 -- # return 0 00:06:47.291 11:17:30 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_get_transports --trtype tcp 00:06:47.291 11:17:30 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:47.291 11:17:30 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:47.291 [2024-07-15 11:17:30.875031] nvmf_rpc.c:2562:rpc_nvmf_get_transports: *ERROR*: transport 'tcp' does not exist 00:06:47.292 request: 00:06:47.292 { 00:06:47.292 "trtype": "tcp", 00:06:47.292 "method": "nvmf_get_transports", 00:06:47.292 "req_id": 1 00:06:47.292 } 00:06:47.292 Got JSON-RPC error response 00:06:47.292 response: 00:06:47.292 { 00:06:47.292 "code": -19, 00:06:47.292 "message": "No such device" 00:06:47.292 } 00:06:47.292 11:17:30 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@587 -- # [[ 1 == 0 ]] 00:06:47.292 11:17:30 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_create_transport -t tcp 00:06:47.292 11:17:30 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:47.292 11:17:30 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:47.292 [2024-07-15 11:17:30.883165] tcp.c: 672:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:06:47.549 11:17:30 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:47.549 11:17:30 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@36 -- # rpc_cmd save_config 00:06:47.549 11:17:30 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:47.549 11:17:30 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:47.549 11:17:31 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:47.549 11:17:31 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@37 -- # cat /var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc/config.json 00:06:47.549 { 00:06:47.549 "subsystems": [ 00:06:47.549 { 00:06:47.549 "subsystem": "keyring", 00:06:47.549 "config": [] 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "subsystem": "iobuf", 00:06:47.549 "config": [ 00:06:47.549 { 00:06:47.549 "method": "iobuf_set_options", 00:06:47.549 "params": { 00:06:47.549 "small_pool_count": 8192, 00:06:47.549 "large_pool_count": 1024, 00:06:47.549 "small_bufsize": 8192, 00:06:47.549 "large_bufsize": 135168 00:06:47.549 } 00:06:47.549 } 00:06:47.549 ] 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "subsystem": "sock", 00:06:47.549 "config": [ 00:06:47.549 { 00:06:47.549 "method": "sock_set_default_impl", 00:06:47.549 "params": { 00:06:47.549 "impl_name": "posix" 00:06:47.549 } 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "method": "sock_impl_set_options", 00:06:47.549 "params": { 00:06:47.549 "impl_name": "ssl", 00:06:47.549 "recv_buf_size": 4096, 00:06:47.549 "send_buf_size": 4096, 00:06:47.549 "enable_recv_pipe": true, 00:06:47.549 "enable_quickack": false, 00:06:47.549 "enable_placement_id": 0, 00:06:47.549 "enable_zerocopy_send_server": true, 00:06:47.549 "enable_zerocopy_send_client": false, 00:06:47.549 "zerocopy_threshold": 0, 00:06:47.549 "tls_version": 0, 00:06:47.549 "enable_ktls": false 00:06:47.549 } 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "method": "sock_impl_set_options", 00:06:47.549 "params": { 00:06:47.549 "impl_name": "posix", 00:06:47.549 "recv_buf_size": 2097152, 00:06:47.549 "send_buf_size": 2097152, 00:06:47.549 "enable_recv_pipe": true, 00:06:47.549 "enable_quickack": false, 00:06:47.549 "enable_placement_id": 0, 00:06:47.549 "enable_zerocopy_send_server": true, 00:06:47.549 "enable_zerocopy_send_client": false, 00:06:47.549 "zerocopy_threshold": 0, 00:06:47.549 "tls_version": 0, 00:06:47.549 "enable_ktls": false 00:06:47.549 } 00:06:47.549 } 00:06:47.549 ] 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "subsystem": "vmd", 00:06:47.549 "config": [] 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "subsystem": "accel", 00:06:47.549 "config": [ 00:06:47.549 { 00:06:47.549 "method": "accel_set_options", 00:06:47.549 "params": { 00:06:47.549 "small_cache_size": 128, 00:06:47.549 "large_cache_size": 16, 00:06:47.549 "task_count": 2048, 00:06:47.549 "sequence_count": 2048, 00:06:47.549 "buf_count": 2048 00:06:47.549 } 00:06:47.549 } 00:06:47.549 ] 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "subsystem": "bdev", 00:06:47.549 "config": [ 00:06:47.549 { 00:06:47.549 "method": "bdev_set_options", 00:06:47.549 "params": { 00:06:47.549 "bdev_io_pool_size": 65535, 00:06:47.549 "bdev_io_cache_size": 256, 00:06:47.549 "bdev_auto_examine": true, 00:06:47.549 "iobuf_small_cache_size": 128, 00:06:47.549 "iobuf_large_cache_size": 16 00:06:47.549 } 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "method": "bdev_raid_set_options", 00:06:47.549 "params": { 00:06:47.549 "process_window_size_kb": 1024 00:06:47.549 } 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "method": "bdev_iscsi_set_options", 00:06:47.549 "params": { 00:06:47.549 "timeout_sec": 30 00:06:47.549 } 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "method": "bdev_nvme_set_options", 00:06:47.549 "params": { 00:06:47.549 "action_on_timeout": "none", 00:06:47.549 "timeout_us": 0, 00:06:47.549 "timeout_admin_us": 0, 00:06:47.549 "keep_alive_timeout_ms": 10000, 00:06:47.549 "arbitration_burst": 0, 00:06:47.549 "low_priority_weight": 0, 00:06:47.549 "medium_priority_weight": 0, 00:06:47.549 "high_priority_weight": 0, 00:06:47.549 "nvme_adminq_poll_period_us": 10000, 00:06:47.549 "nvme_ioq_poll_period_us": 0, 00:06:47.549 "io_queue_requests": 0, 00:06:47.549 "delay_cmd_submit": true, 00:06:47.549 "transport_retry_count": 4, 00:06:47.549 "bdev_retry_count": 3, 00:06:47.549 "transport_ack_timeout": 0, 00:06:47.549 "ctrlr_loss_timeout_sec": 0, 00:06:47.549 "reconnect_delay_sec": 0, 00:06:47.549 "fast_io_fail_timeout_sec": 0, 00:06:47.549 "disable_auto_failback": false, 00:06:47.549 "generate_uuids": false, 00:06:47.549 "transport_tos": 0, 00:06:47.549 "nvme_error_stat": false, 00:06:47.549 "rdma_srq_size": 0, 00:06:47.549 "io_path_stat": false, 00:06:47.549 "allow_accel_sequence": false, 00:06:47.549 "rdma_max_cq_size": 0, 00:06:47.549 "rdma_cm_event_timeout_ms": 0, 00:06:47.549 "dhchap_digests": [ 00:06:47.549 "sha256", 00:06:47.549 "sha384", 00:06:47.549 "sha512" 00:06:47.549 ], 00:06:47.549 "dhchap_dhgroups": [ 00:06:47.549 "null", 00:06:47.549 "ffdhe2048", 00:06:47.549 "ffdhe3072", 00:06:47.549 "ffdhe4096", 00:06:47.549 "ffdhe6144", 00:06:47.549 "ffdhe8192" 00:06:47.549 ] 00:06:47.549 } 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "method": "bdev_nvme_set_hotplug", 00:06:47.549 "params": { 00:06:47.549 "period_us": 100000, 00:06:47.549 "enable": false 00:06:47.549 } 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "method": "bdev_wait_for_examine" 00:06:47.549 } 00:06:47.549 ] 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "subsystem": "scsi", 00:06:47.549 "config": null 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "subsystem": "scheduler", 00:06:47.549 "config": [ 00:06:47.549 { 00:06:47.549 "method": "framework_set_scheduler", 00:06:47.549 "params": { 00:06:47.549 "name": "static" 00:06:47.549 } 00:06:47.549 } 00:06:47.549 ] 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "subsystem": "vhost_scsi", 00:06:47.549 "config": [] 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "subsystem": "vhost_blk", 00:06:47.549 "config": [] 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "subsystem": "ublk", 00:06:47.549 "config": [] 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "subsystem": "nbd", 00:06:47.549 "config": [] 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "subsystem": "nvmf", 00:06:47.549 "config": [ 00:06:47.549 { 00:06:47.549 "method": "nvmf_set_config", 00:06:47.549 "params": { 00:06:47.549 "discovery_filter": "match_any", 00:06:47.549 "admin_cmd_passthru": { 00:06:47.549 "identify_ctrlr": false 00:06:47.549 } 00:06:47.549 } 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "method": "nvmf_set_max_subsystems", 00:06:47.549 "params": { 00:06:47.549 "max_subsystems": 1024 00:06:47.549 } 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "method": "nvmf_set_crdt", 00:06:47.549 "params": { 00:06:47.549 "crdt1": 0, 00:06:47.549 "crdt2": 0, 00:06:47.549 "crdt3": 0 00:06:47.549 } 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "method": "nvmf_create_transport", 00:06:47.549 "params": { 00:06:47.549 "trtype": "TCP", 00:06:47.549 "max_queue_depth": 128, 00:06:47.549 "max_io_qpairs_per_ctrlr": 127, 00:06:47.549 "in_capsule_data_size": 4096, 00:06:47.549 "max_io_size": 131072, 00:06:47.549 "io_unit_size": 131072, 00:06:47.549 "max_aq_depth": 128, 00:06:47.549 "num_shared_buffers": 511, 00:06:47.549 "buf_cache_size": 4294967295, 00:06:47.549 "dif_insert_or_strip": false, 00:06:47.549 "zcopy": false, 00:06:47.549 "c2h_success": true, 00:06:47.549 "sock_priority": 0, 00:06:47.549 "abort_timeout_sec": 1, 00:06:47.549 "ack_timeout": 0, 00:06:47.549 "data_wr_pool_size": 0 00:06:47.549 } 00:06:47.549 } 00:06:47.549 ] 00:06:47.549 }, 00:06:47.549 { 00:06:47.549 "subsystem": "iscsi", 00:06:47.549 "config": [ 00:06:47.549 { 00:06:47.549 "method": "iscsi_set_options", 00:06:47.549 "params": { 00:06:47.549 "node_base": "iqn.2016-06.io.spdk", 00:06:47.549 "max_sessions": 128, 00:06:47.549 "max_connections_per_session": 2, 00:06:47.549 "max_queue_depth": 64, 00:06:47.549 "default_time2wait": 2, 00:06:47.549 "default_time2retain": 20, 00:06:47.549 "first_burst_length": 8192, 00:06:47.549 "immediate_data": true, 00:06:47.549 "allow_duplicated_isid": false, 00:06:47.549 "error_recovery_level": 0, 00:06:47.549 "nop_timeout": 60, 00:06:47.549 "nop_in_interval": 30, 00:06:47.549 "disable_chap": false, 00:06:47.549 "require_chap": false, 00:06:47.549 "mutual_chap": false, 00:06:47.549 "chap_group": 0, 00:06:47.549 "max_large_datain_per_connection": 64, 00:06:47.549 "max_r2t_per_connection": 4, 00:06:47.549 "pdu_pool_size": 36864, 00:06:47.549 "immediate_data_pool_size": 16384, 00:06:47.549 "data_out_pool_size": 2048 00:06:47.549 } 00:06:47.549 } 00:06:47.549 ] 00:06:47.549 } 00:06:47.549 ] 00:06:47.549 } 00:06:47.549 11:17:31 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:06:47.550 11:17:31 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@40 -- # killprocess 822989 00:06:47.550 11:17:31 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@948 -- # '[' -z 822989 ']' 00:06:47.550 11:17:31 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@952 -- # kill -0 822989 00:06:47.550 11:17:31 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@953 -- # uname 00:06:47.550 11:17:31 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:06:47.550 11:17:31 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 822989 00:06:47.550 11:17:31 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:06:47.550 11:17:31 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:06:47.550 11:17:31 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@966 -- # echo 'killing process with pid 822989' 00:06:47.550 killing process with pid 822989 00:06:47.550 11:17:31 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@967 -- # kill 822989 00:06:47.550 11:17:31 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # wait 822989 00:06:48.114 11:17:31 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@47 -- # local spdk_pid=823174 00:06:48.114 11:17:31 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@48 -- # sleep 5 00:06:48.114 11:17:31 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@46 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc/config.json 00:06:53.368 11:17:36 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@50 -- # killprocess 823174 00:06:53.368 11:17:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@948 -- # '[' -z 823174 ']' 00:06:53.368 11:17:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@952 -- # kill -0 823174 00:06:53.368 11:17:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@953 -- # uname 00:06:53.368 11:17:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:06:53.368 11:17:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 823174 00:06:53.368 11:17:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:06:53.368 11:17:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:06:53.368 11:17:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@966 -- # echo 'killing process with pid 823174' 00:06:53.368 killing process with pid 823174 00:06:53.368 11:17:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@967 -- # kill 823174 00:06:53.368 11:17:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # wait 823174 00:06:53.368 11:17:36 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@51 -- # grep -q 'TCP Transport Init' /var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc/log.txt 00:06:53.368 11:17:36 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@52 -- # rm /var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc/log.txt 00:06:53.368 00:06:53.368 real 0m7.002s 00:06:53.368 user 0m6.534s 00:06:53.368 sys 0m0.979s 00:06:53.368 11:17:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:53.368 11:17:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:53.368 ************************************ 00:06:53.368 END TEST skip_rpc_with_json 00:06:53.368 ************************************ 00:06:53.368 11:17:36 skip_rpc -- common/autotest_common.sh@1142 -- # return 0 00:06:53.368 11:17:36 skip_rpc -- rpc/skip_rpc.sh@75 -- # run_test skip_rpc_with_delay test_skip_rpc_with_delay 00:06:53.368 11:17:36 skip_rpc -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:53.368 11:17:36 skip_rpc -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:53.368 11:17:36 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:53.626 ************************************ 00:06:53.626 START TEST skip_rpc_with_delay 00:06:53.626 ************************************ 00:06:53.626 11:17:36 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1123 -- # test_skip_rpc_with_delay 00:06:53.626 11:17:36 skip_rpc.skip_rpc_with_delay -- rpc/skip_rpc.sh@57 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:53.626 11:17:36 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@648 -- # local es=0 00:06:53.626 11:17:36 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:53.626 11:17:36 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt 00:06:53.626 11:17:36 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:06:53.626 11:17:36 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt 00:06:53.626 11:17:36 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:06:53.626 11:17:36 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt 00:06:53.626 11:17:36 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:06:53.626 11:17:36 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt 00:06:53.626 11:17:36 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt ]] 00:06:53.626 11:17:36 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:53.626 [2024-07-15 11:17:37.050286] app.c: 832:spdk_app_start: *ERROR*: Cannot use '--wait-for-rpc' if no RPC server is going to be started. 00:06:53.626 [2024-07-15 11:17:37.050392] app.c: 711:unclaim_cpu_cores: *ERROR*: Failed to unlink lock fd for core 0, errno: 2 00:06:53.626 11:17:37 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@651 -- # es=1 00:06:53.626 11:17:37 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:06:53.626 11:17:37 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:06:53.626 11:17:37 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:06:53.626 00:06:53.626 real 0m0.095s 00:06:53.626 user 0m0.055s 00:06:53.626 sys 0m0.040s 00:06:53.626 11:17:37 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:53.626 11:17:37 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@10 -- # set +x 00:06:53.626 ************************************ 00:06:53.626 END TEST skip_rpc_with_delay 00:06:53.626 ************************************ 00:06:53.626 11:17:37 skip_rpc -- common/autotest_common.sh@1142 -- # return 0 00:06:53.626 11:17:37 skip_rpc -- rpc/skip_rpc.sh@77 -- # uname 00:06:53.626 11:17:37 skip_rpc -- rpc/skip_rpc.sh@77 -- # '[' Linux '!=' FreeBSD ']' 00:06:53.626 11:17:37 skip_rpc -- rpc/skip_rpc.sh@78 -- # run_test exit_on_failed_rpc_init test_exit_on_failed_rpc_init 00:06:53.626 11:17:37 skip_rpc -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:53.626 11:17:37 skip_rpc -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:53.626 11:17:37 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:53.626 ************************************ 00:06:53.626 START TEST exit_on_failed_rpc_init 00:06:53.626 ************************************ 00:06:53.626 11:17:37 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1123 -- # test_exit_on_failed_rpc_init 00:06:53.626 11:17:37 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@62 -- # local spdk_pid=823934 00:06:53.626 11:17:37 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@63 -- # waitforlisten 823934 00:06:53.626 11:17:37 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@61 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:06:53.626 11:17:37 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@829 -- # '[' -z 823934 ']' 00:06:53.626 11:17:37 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:53.626 11:17:37 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@834 -- # local max_retries=100 00:06:53.626 11:17:37 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:53.626 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:53.626 11:17:37 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@838 -- # xtrace_disable 00:06:53.626 11:17:37 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:06:53.884 [2024-07-15 11:17:37.231551] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:06:53.884 [2024-07-15 11:17:37.231621] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid823934 ] 00:06:53.884 [2024-07-15 11:17:37.361812] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:53.884 [2024-07-15 11:17:37.464209] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:06:54.818 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:06:54.818 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@862 -- # return 0 00:06:54.818 11:17:38 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@65 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:06:54.818 11:17:38 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@67 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:06:54.818 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@648 -- # local es=0 00:06:54.818 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:06:54.818 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt 00:06:54.818 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:06:54.818 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt 00:06:54.818 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:06:54.818 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt 00:06:54.818 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:06:54.818 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt 00:06:54.818 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt ]] 00:06:54.818 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:06:54.818 [2024-07-15 11:17:38.148304] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:06:54.818 [2024-07-15 11:17:38.148356] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid824109 ] 00:06:54.818 [2024-07-15 11:17:38.248953] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:54.818 [2024-07-15 11:17:38.347574] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:06:54.818 [2024-07-15 11:17:38.347668] rpc.c: 180:_spdk_rpc_listen: *ERROR*: RPC Unix domain socket path /var/tmp/spdk.sock in use. Specify another. 00:06:54.818 [2024-07-15 11:17:38.347685] rpc.c: 166:spdk_rpc_initialize: *ERROR*: Unable to start RPC service at /var/tmp/spdk.sock 00:06:54.818 [2024-07-15 11:17:38.347697] app.c:1053:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:06:55.076 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@651 -- # es=234 00:06:55.076 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:06:55.076 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@660 -- # es=106 00:06:55.076 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@661 -- # case "$es" in 00:06:55.076 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@668 -- # es=1 00:06:55.076 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:06:55.076 11:17:38 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:06:55.076 11:17:38 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@70 -- # killprocess 823934 00:06:55.076 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@948 -- # '[' -z 823934 ']' 00:06:55.076 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@952 -- # kill -0 823934 00:06:55.076 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@953 -- # uname 00:06:55.076 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:06:55.076 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 823934 00:06:55.076 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:06:55.076 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:06:55.076 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@966 -- # echo 'killing process with pid 823934' 00:06:55.076 killing process with pid 823934 00:06:55.076 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@967 -- # kill 823934 00:06:55.076 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@972 -- # wait 823934 00:06:55.334 00:06:55.334 real 0m1.704s 00:06:55.334 user 0m1.926s 00:06:55.334 sys 0m0.575s 00:06:55.334 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:55.334 11:17:38 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:06:55.334 ************************************ 00:06:55.334 END TEST exit_on_failed_rpc_init 00:06:55.334 ************************************ 00:06:55.334 11:17:38 skip_rpc -- common/autotest_common.sh@1142 -- # return 0 00:06:55.334 11:17:38 skip_rpc -- rpc/skip_rpc.sh@81 -- # rm /var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc/config.json 00:06:55.334 00:06:55.334 real 0m14.683s 00:06:55.334 user 0m13.722s 00:06:55.334 sys 0m2.311s 00:06:55.334 11:17:38 skip_rpc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:55.334 11:17:38 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:55.334 ************************************ 00:06:55.334 END TEST skip_rpc 00:06:55.334 ************************************ 00:06:55.591 11:17:38 -- common/autotest_common.sh@1142 -- # return 0 00:06:55.591 11:17:38 -- spdk/autotest.sh@171 -- # run_test rpc_client /var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_client/rpc_client.sh 00:06:55.591 11:17:38 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:55.591 11:17:38 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:55.591 11:17:38 -- common/autotest_common.sh@10 -- # set +x 00:06:55.591 ************************************ 00:06:55.591 START TEST rpc_client 00:06:55.591 ************************************ 00:06:55.591 11:17:39 rpc_client -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_client/rpc_client.sh 00:06:55.591 * Looking for test storage... 00:06:55.591 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_client 00:06:55.591 11:17:39 rpc_client -- rpc_client/rpc_client.sh@10 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_client/rpc_client_test 00:06:55.591 OK 00:06:55.591 11:17:39 rpc_client -- rpc_client/rpc_client.sh@12 -- # trap - SIGINT SIGTERM EXIT 00:06:55.591 00:06:55.591 real 0m0.139s 00:06:55.591 user 0m0.063s 00:06:55.591 sys 0m0.084s 00:06:55.591 11:17:39 rpc_client -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:55.591 11:17:39 rpc_client -- common/autotest_common.sh@10 -- # set +x 00:06:55.591 ************************************ 00:06:55.592 END TEST rpc_client 00:06:55.592 ************************************ 00:06:55.592 11:17:39 -- common/autotest_common.sh@1142 -- # return 0 00:06:55.592 11:17:39 -- spdk/autotest.sh@172 -- # run_test json_config /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/json_config.sh 00:06:55.592 11:17:39 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:55.592 11:17:39 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:55.592 11:17:39 -- common/autotest_common.sh@10 -- # set +x 00:06:55.849 ************************************ 00:06:55.849 START TEST json_config 00:06:55.849 ************************************ 00:06:55.849 11:17:39 json_config -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/json_config.sh 00:06:55.849 11:17:39 json_config -- json_config/json_config.sh@8 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/nvmf/common.sh 00:06:55.849 11:17:39 json_config -- nvmf/common.sh@7 -- # uname -s 00:06:55.849 11:17:39 json_config -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:55.849 11:17:39 json_config -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:55.849 11:17:39 json_config -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:55.849 11:17:39 json_config -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:55.849 11:17:39 json_config -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:55.849 11:17:39 json_config -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:55.849 11:17:39 json_config -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:55.849 11:17:39 json_config -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:55.849 11:17:39 json_config -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:55.849 11:17:39 json_config -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:55.849 11:17:39 json_config -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:809e2efd-7f71-e711-906e-0017a4403562 00:06:55.849 11:17:39 json_config -- nvmf/common.sh@18 -- # NVME_HOSTID=809e2efd-7f71-e711-906e-0017a4403562 00:06:55.849 11:17:39 json_config -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:55.849 11:17:39 json_config -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:55.849 11:17:39 json_config -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:06:55.849 11:17:39 json_config -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:55.849 11:17:39 json_config -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/common.sh 00:06:55.849 11:17:39 json_config -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:55.849 11:17:39 json_config -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:55.849 11:17:39 json_config -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:55.850 11:17:39 json_config -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:55.850 11:17:39 json_config -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:55.850 11:17:39 json_config -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:55.850 11:17:39 json_config -- paths/export.sh@5 -- # export PATH 00:06:55.850 11:17:39 json_config -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:55.850 11:17:39 json_config -- nvmf/common.sh@47 -- # : 0 00:06:55.850 11:17:39 json_config -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:06:55.850 11:17:39 json_config -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:06:55.850 11:17:39 json_config -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:55.850 11:17:39 json_config -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:55.850 11:17:39 json_config -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:55.850 11:17:39 json_config -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:06:55.850 11:17:39 json_config -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:06:55.850 11:17:39 json_config -- nvmf/common.sh@51 -- # have_pci_nics=0 00:06:55.850 11:17:39 json_config -- json_config/json_config.sh@9 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/common.sh 00:06:55.850 11:17:39 json_config -- json_config/json_config.sh@11 -- # [[ 0 -eq 1 ]] 00:06:55.850 11:17:39 json_config -- json_config/json_config.sh@15 -- # [[ 0 -ne 1 ]] 00:06:55.850 11:17:39 json_config -- json_config/json_config.sh@15 -- # [[ 0 -eq 1 ]] 00:06:55.850 11:17:39 json_config -- json_config/json_config.sh@26 -- # (( SPDK_TEST_BLOCKDEV + SPDK_TEST_ISCSI + SPDK_TEST_NVMF + SPDK_TEST_VHOST + SPDK_TEST_VHOST_INIT + SPDK_TEST_RBD == 0 )) 00:06:55.850 11:17:39 json_config -- json_config/json_config.sh@31 -- # app_pid=(['target']='' ['initiator']='') 00:06:55.850 11:17:39 json_config -- json_config/json_config.sh@31 -- # declare -A app_pid 00:06:55.850 11:17:39 json_config -- json_config/json_config.sh@32 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock' ['initiator']='/var/tmp/spdk_initiator.sock') 00:06:55.850 11:17:39 json_config -- json_config/json_config.sh@32 -- # declare -A app_socket 00:06:55.850 11:17:39 json_config -- json_config/json_config.sh@33 -- # app_params=(['target']='-m 0x1 -s 1024' ['initiator']='-m 0x2 -g -u -s 1024') 00:06:55.850 11:17:39 json_config -- json_config/json_config.sh@33 -- # declare -A app_params 00:06:55.850 11:17:39 json_config -- json_config/json_config.sh@34 -- # configs_path=(['target']='/var/jenkins/workspace/crypto-phy-autotest/spdk/spdk_tgt_config.json' ['initiator']='/var/jenkins/workspace/crypto-phy-autotest/spdk/spdk_initiator_config.json') 00:06:55.850 11:17:39 json_config -- json_config/json_config.sh@34 -- # declare -A configs_path 00:06:55.850 11:17:39 json_config -- json_config/json_config.sh@40 -- # last_event_id=0 00:06:55.850 11:17:39 json_config -- json_config/json_config.sh@355 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:06:55.850 11:17:39 json_config -- json_config/json_config.sh@356 -- # echo 'INFO: JSON configuration test init' 00:06:55.850 INFO: JSON configuration test init 00:06:55.850 11:17:39 json_config -- json_config/json_config.sh@357 -- # json_config_test_init 00:06:55.850 11:17:39 json_config -- json_config/json_config.sh@262 -- # timing_enter json_config_test_init 00:06:55.850 11:17:39 json_config -- common/autotest_common.sh@722 -- # xtrace_disable 00:06:55.850 11:17:39 json_config -- common/autotest_common.sh@10 -- # set +x 00:06:55.850 11:17:39 json_config -- json_config/json_config.sh@263 -- # timing_enter json_config_setup_target 00:06:55.850 11:17:39 json_config -- common/autotest_common.sh@722 -- # xtrace_disable 00:06:55.850 11:17:39 json_config -- common/autotest_common.sh@10 -- # set +x 00:06:55.850 11:17:39 json_config -- json_config/json_config.sh@265 -- # json_config_test_start_app target --wait-for-rpc 00:06:55.850 11:17:39 json_config -- json_config/common.sh@9 -- # local app=target 00:06:55.850 11:17:39 json_config -- json_config/common.sh@10 -- # shift 00:06:55.850 11:17:39 json_config -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:06:55.850 11:17:39 json_config -- json_config/common.sh@13 -- # [[ -z '' ]] 00:06:55.850 11:17:39 json_config -- json_config/common.sh@15 -- # local app_extra_params= 00:06:55.850 11:17:39 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:06:55.850 11:17:39 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:06:55.850 11:17:39 json_config -- json_config/common.sh@22 -- # app_pid["$app"]=824391 00:06:55.850 11:17:39 json_config -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:06:55.850 Waiting for target to run... 00:06:55.850 11:17:39 json_config -- json_config/common.sh@21 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --wait-for-rpc 00:06:55.850 11:17:39 json_config -- json_config/common.sh@25 -- # waitforlisten 824391 /var/tmp/spdk_tgt.sock 00:06:55.850 11:17:39 json_config -- common/autotest_common.sh@829 -- # '[' -z 824391 ']' 00:06:55.850 11:17:39 json_config -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:06:55.850 11:17:39 json_config -- common/autotest_common.sh@834 -- # local max_retries=100 00:06:55.850 11:17:39 json_config -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:06:55.850 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:06:55.850 11:17:39 json_config -- common/autotest_common.sh@838 -- # xtrace_disable 00:06:55.850 11:17:39 json_config -- common/autotest_common.sh@10 -- # set +x 00:06:55.850 [2024-07-15 11:17:39.405452] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:06:55.850 [2024-07-15 11:17:39.405525] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid824391 ] 00:06:56.415 [2024-07-15 11:17:39.962462] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:56.672 [2024-07-15 11:17:40.068259] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:06:56.930 11:17:40 json_config -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:06:56.930 11:17:40 json_config -- common/autotest_common.sh@862 -- # return 0 00:06:56.930 11:17:40 json_config -- json_config/common.sh@26 -- # echo '' 00:06:56.930 00:06:56.930 11:17:40 json_config -- json_config/json_config.sh@269 -- # create_accel_config 00:06:56.930 11:17:40 json_config -- json_config/json_config.sh@93 -- # timing_enter create_accel_config 00:06:56.930 11:17:40 json_config -- common/autotest_common.sh@722 -- # xtrace_disable 00:06:56.930 11:17:40 json_config -- common/autotest_common.sh@10 -- # set +x 00:06:56.930 11:17:40 json_config -- json_config/json_config.sh@95 -- # [[ 1 -eq 1 ]] 00:06:56.930 11:17:40 json_config -- json_config/json_config.sh@96 -- # tgt_rpc dpdk_cryptodev_scan_accel_module 00:06:56.930 11:17:40 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock dpdk_cryptodev_scan_accel_module 00:06:56.930 11:17:40 json_config -- json_config/json_config.sh@97 -- # tgt_rpc accel_assign_opc -o encrypt -m dpdk_cryptodev 00:06:56.930 11:17:40 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock accel_assign_opc -o encrypt -m dpdk_cryptodev 00:06:57.188 [2024-07-15 11:17:40.682299] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation encrypt will be assigned to module dpdk_cryptodev 00:06:57.188 11:17:40 json_config -- json_config/json_config.sh@98 -- # tgt_rpc accel_assign_opc -o decrypt -m dpdk_cryptodev 00:06:57.188 11:17:40 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock accel_assign_opc -o decrypt -m dpdk_cryptodev 00:06:57.445 [2024-07-15 11:17:40.926913] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation decrypt will be assigned to module dpdk_cryptodev 00:06:57.445 11:17:40 json_config -- json_config/json_config.sh@101 -- # timing_exit create_accel_config 00:06:57.445 11:17:40 json_config -- common/autotest_common.sh@728 -- # xtrace_disable 00:06:57.445 11:17:40 json_config -- common/autotest_common.sh@10 -- # set +x 00:06:57.445 11:17:40 json_config -- json_config/json_config.sh@273 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/gen_nvme.sh --json-with-subsystems 00:06:57.445 11:17:40 json_config -- json_config/json_config.sh@274 -- # tgt_rpc load_config 00:06:57.445 11:17:40 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock load_config 00:06:57.703 [2024-07-15 11:17:41.236403] accel_dpdk_cryptodev.c:1178:accel_dpdk_cryptodev_init: *NOTICE*: Found crypto devices: 97 00:07:00.227 11:17:43 json_config -- json_config/json_config.sh@276 -- # tgt_check_notification_types 00:07:00.227 11:17:43 json_config -- json_config/json_config.sh@43 -- # timing_enter tgt_check_notification_types 00:07:00.227 11:17:43 json_config -- common/autotest_common.sh@722 -- # xtrace_disable 00:07:00.227 11:17:43 json_config -- common/autotest_common.sh@10 -- # set +x 00:07:00.484 11:17:43 json_config -- json_config/json_config.sh@45 -- # local ret=0 00:07:00.484 11:17:43 json_config -- json_config/json_config.sh@46 -- # enabled_types=('bdev_register' 'bdev_unregister') 00:07:00.484 11:17:43 json_config -- json_config/json_config.sh@46 -- # local enabled_types 00:07:00.484 11:17:43 json_config -- json_config/json_config.sh@48 -- # tgt_rpc notify_get_types 00:07:00.484 11:17:43 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock notify_get_types 00:07:00.484 11:17:43 json_config -- json_config/json_config.sh@48 -- # jq -r '.[]' 00:07:00.741 11:17:44 json_config -- json_config/json_config.sh@48 -- # get_types=('bdev_register' 'bdev_unregister') 00:07:00.741 11:17:44 json_config -- json_config/json_config.sh@48 -- # local get_types 00:07:00.741 11:17:44 json_config -- json_config/json_config.sh@49 -- # [[ bdev_register bdev_unregister != \b\d\e\v\_\r\e\g\i\s\t\e\r\ \b\d\e\v\_\u\n\r\e\g\i\s\t\e\r ]] 00:07:00.741 11:17:44 json_config -- json_config/json_config.sh@54 -- # timing_exit tgt_check_notification_types 00:07:00.741 11:17:44 json_config -- common/autotest_common.sh@728 -- # xtrace_disable 00:07:00.741 11:17:44 json_config -- common/autotest_common.sh@10 -- # set +x 00:07:00.741 11:17:44 json_config -- json_config/json_config.sh@55 -- # return 0 00:07:00.741 11:17:44 json_config -- json_config/json_config.sh@278 -- # [[ 1 -eq 1 ]] 00:07:00.741 11:17:44 json_config -- json_config/json_config.sh@279 -- # create_bdev_subsystem_config 00:07:00.741 11:17:44 json_config -- json_config/json_config.sh@105 -- # timing_enter create_bdev_subsystem_config 00:07:00.741 11:17:44 json_config -- common/autotest_common.sh@722 -- # xtrace_disable 00:07:00.741 11:17:44 json_config -- common/autotest_common.sh@10 -- # set +x 00:07:00.742 11:17:44 json_config -- json_config/json_config.sh@107 -- # expected_notifications=() 00:07:00.742 11:17:44 json_config -- json_config/json_config.sh@107 -- # local expected_notifications 00:07:00.742 11:17:44 json_config -- json_config/json_config.sh@111 -- # expected_notifications+=($(get_notifications)) 00:07:00.742 11:17:44 json_config -- json_config/json_config.sh@111 -- # get_notifications 00:07:00.742 11:17:44 json_config -- json_config/json_config.sh@59 -- # local ev_type ev_ctx event_id 00:07:00.742 11:17:44 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:00.742 11:17:44 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:00.742 11:17:44 json_config -- json_config/json_config.sh@58 -- # tgt_rpc notify_get_notifications -i 0 00:07:00.742 11:17:44 json_config -- json_config/json_config.sh@58 -- # jq -r '.[] | "\(.type):\(.ctx):\(.id)"' 00:07:00.742 11:17:44 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock notify_get_notifications -i 0 00:07:00.999 11:17:44 json_config -- json_config/json_config.sh@62 -- # echo bdev_register:Nvme0n1 00:07:00.999 11:17:44 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:00.999 11:17:44 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:00.999 11:17:44 json_config -- json_config/json_config.sh@113 -- # [[ 1 -eq 1 ]] 00:07:00.999 11:17:44 json_config -- json_config/json_config.sh@114 -- # local lvol_store_base_bdev=Nvme0n1 00:07:00.999 11:17:44 json_config -- json_config/json_config.sh@116 -- # tgt_rpc bdev_split_create Nvme0n1 2 00:07:00.999 11:17:44 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_split_create Nvme0n1 2 00:07:01.257 Nvme0n1p0 Nvme0n1p1 00:07:01.257 11:17:44 json_config -- json_config/json_config.sh@117 -- # tgt_rpc bdev_split_create Malloc0 3 00:07:01.257 11:17:44 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_split_create Malloc0 3 00:07:01.514 [2024-07-15 11:17:44.858999] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:07:01.514 [2024-07-15 11:17:44.859057] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:07:01.514 00:07:01.514 11:17:44 json_config -- json_config/json_config.sh@118 -- # tgt_rpc bdev_malloc_create 8 4096 --name Malloc3 00:07:01.514 11:17:44 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 8 4096 --name Malloc3 00:07:01.514 Malloc3 00:07:01.772 11:17:45 json_config -- json_config/json_config.sh@119 -- # tgt_rpc bdev_passthru_create -b Malloc3 -p PTBdevFromMalloc3 00:07:01.772 11:17:45 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_passthru_create -b Malloc3 -p PTBdevFromMalloc3 00:07:01.772 [2024-07-15 11:17:45.332353] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc3 00:07:01.772 [2024-07-15 11:17:45.332405] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:01.772 [2024-07-15 11:17:45.332431] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x206da00 00:07:01.772 [2024-07-15 11:17:45.332444] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:01.772 [2024-07-15 11:17:45.334059] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:01.772 [2024-07-15 11:17:45.334090] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: PTBdevFromMalloc3 00:07:01.772 PTBdevFromMalloc3 00:07:01.772 11:17:45 json_config -- json_config/json_config.sh@121 -- # tgt_rpc bdev_null_create Null0 32 512 00:07:01.772 11:17:45 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_null_create Null0 32 512 00:07:02.030 Null0 00:07:02.030 11:17:45 json_config -- json_config/json_config.sh@123 -- # tgt_rpc bdev_malloc_create 32 512 --name Malloc0 00:07:02.030 11:17:45 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 32 512 --name Malloc0 00:07:02.287 Malloc0 00:07:02.287 11:17:45 json_config -- json_config/json_config.sh@124 -- # tgt_rpc bdev_malloc_create 16 4096 --name Malloc1 00:07:02.287 11:17:45 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 16 4096 --name Malloc1 00:07:02.544 Malloc1 00:07:02.544 11:17:46 json_config -- json_config/json_config.sh@137 -- # expected_notifications+=(bdev_register:${lvol_store_base_bdev}p1 bdev_register:${lvol_store_base_bdev}p0 bdev_register:Malloc3 bdev_register:PTBdevFromMalloc3 bdev_register:Null0 bdev_register:Malloc0 bdev_register:Malloc0p2 bdev_register:Malloc0p1 bdev_register:Malloc0p0 bdev_register:Malloc1) 00:07:02.544 11:17:46 json_config -- json_config/json_config.sh@140 -- # dd if=/dev/zero of=/sample_aio bs=1024 count=102400 00:07:03.170 102400+0 records in 00:07:03.170 102400+0 records out 00:07:03.170 104857600 bytes (105 MB, 100 MiB) copied, 0.308707 s, 340 MB/s 00:07:03.170 11:17:46 json_config -- json_config/json_config.sh@141 -- # tgt_rpc bdev_aio_create /sample_aio aio_disk 1024 00:07:03.170 11:17:46 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_aio_create /sample_aio aio_disk 1024 00:07:03.170 aio_disk 00:07:03.170 11:17:46 json_config -- json_config/json_config.sh@142 -- # expected_notifications+=(bdev_register:aio_disk) 00:07:03.170 11:17:46 json_config -- json_config/json_config.sh@147 -- # tgt_rpc bdev_lvol_create_lvstore -c 1048576 Nvme0n1p0 lvs_test 00:07:03.170 11:17:46 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_lvol_create_lvstore -c 1048576 Nvme0n1p0 lvs_test 00:07:08.427 1681c52d-f9db-4afb-a6d7-14a141766525 00:07:08.427 11:17:51 json_config -- json_config/json_config.sh@154 -- # expected_notifications+=("bdev_register:$(tgt_rpc bdev_lvol_create -l lvs_test lvol0 32)" "bdev_register:$(tgt_rpc bdev_lvol_create -l lvs_test -t lvol1 32)" "bdev_register:$(tgt_rpc bdev_lvol_snapshot lvs_test/lvol0 snapshot0)" "bdev_register:$(tgt_rpc bdev_lvol_clone lvs_test/snapshot0 clone0)") 00:07:08.427 11:17:51 json_config -- json_config/json_config.sh@154 -- # tgt_rpc bdev_lvol_create -l lvs_test lvol0 32 00:07:08.427 11:17:51 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_lvol_create -l lvs_test lvol0 32 00:07:08.427 11:17:51 json_config -- json_config/json_config.sh@154 -- # tgt_rpc bdev_lvol_create -l lvs_test -t lvol1 32 00:07:08.427 11:17:51 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_lvol_create -l lvs_test -t lvol1 32 00:07:08.427 11:17:51 json_config -- json_config/json_config.sh@154 -- # tgt_rpc bdev_lvol_snapshot lvs_test/lvol0 snapshot0 00:07:08.427 11:17:51 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_lvol_snapshot lvs_test/lvol0 snapshot0 00:07:08.685 11:17:52 json_config -- json_config/json_config.sh@154 -- # tgt_rpc bdev_lvol_clone lvs_test/snapshot0 clone0 00:07:08.685 11:17:52 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_lvol_clone lvs_test/snapshot0 clone0 00:07:08.943 11:17:52 json_config -- json_config/json_config.sh@157 -- # [[ 1 -eq 1 ]] 00:07:08.943 11:17:52 json_config -- json_config/json_config.sh@158 -- # tgt_rpc bdev_malloc_create 8 1024 --name MallocForCryptoBdev 00:07:08.943 11:17:52 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 8 1024 --name MallocForCryptoBdev 00:07:09.200 MallocForCryptoBdev 00:07:09.200 11:17:52 json_config -- json_config/json_config.sh@159 -- # lspci -d:37c8 00:07:09.200 11:17:52 json_config -- json_config/json_config.sh@159 -- # wc -l 00:07:09.200 11:17:52 json_config -- json_config/json_config.sh@159 -- # [[ 3 -eq 0 ]] 00:07:09.200 11:17:52 json_config -- json_config/json_config.sh@162 -- # local crypto_driver=crypto_qat 00:07:09.200 11:17:52 json_config -- json_config/json_config.sh@165 -- # tgt_rpc bdev_crypto_create MallocForCryptoBdev CryptoMallocBdev -p crypto_qat -k 01234567891234560123456789123456 00:07:09.200 11:17:52 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_crypto_create MallocForCryptoBdev CryptoMallocBdev -p crypto_qat -k 01234567891234560123456789123456 00:07:09.457 [2024-07-15 11:17:52.825266] vbdev_crypto_rpc.c: 136:rpc_bdev_crypto_create: *WARNING*: "crypto_pmd" parameters is obsolete and ignored 00:07:09.457 CryptoMallocBdev 00:07:09.457 11:17:52 json_config -- json_config/json_config.sh@169 -- # expected_notifications+=(bdev_register:MallocForCryptoBdev bdev_register:CryptoMallocBdev) 00:07:09.457 11:17:52 json_config -- json_config/json_config.sh@172 -- # [[ 0 -eq 1 ]] 00:07:09.457 11:17:52 json_config -- json_config/json_config.sh@178 -- # tgt_check_notifications bdev_register:Nvme0n1 bdev_register:Nvme0n1p1 bdev_register:Nvme0n1p0 bdev_register:Malloc3 bdev_register:PTBdevFromMalloc3 bdev_register:Null0 bdev_register:Malloc0 bdev_register:Malloc0p2 bdev_register:Malloc0p1 bdev_register:Malloc0p0 bdev_register:Malloc1 bdev_register:aio_disk bdev_register:635e020f-2df2-442c-a686-9c4a3693428b bdev_register:9f761a71-5301-47a6-b672-523287690cfb bdev_register:aa616c70-3dcd-442b-a2cd-386b7636fd14 bdev_register:5edb3c1d-d24a-48cd-a475-2eb8fbe903e5 bdev_register:MallocForCryptoBdev bdev_register:CryptoMallocBdev 00:07:09.457 11:17:52 json_config -- json_config/json_config.sh@67 -- # local events_to_check 00:07:09.457 11:17:52 json_config -- json_config/json_config.sh@68 -- # local recorded_events 00:07:09.457 11:17:52 json_config -- json_config/json_config.sh@71 -- # events_to_check=($(printf '%s\n' "$@" | sort)) 00:07:09.458 11:17:52 json_config -- json_config/json_config.sh@71 -- # printf '%s\n' bdev_register:Nvme0n1 bdev_register:Nvme0n1p1 bdev_register:Nvme0n1p0 bdev_register:Malloc3 bdev_register:PTBdevFromMalloc3 bdev_register:Null0 bdev_register:Malloc0 bdev_register:Malloc0p2 bdev_register:Malloc0p1 bdev_register:Malloc0p0 bdev_register:Malloc1 bdev_register:aio_disk bdev_register:635e020f-2df2-442c-a686-9c4a3693428b bdev_register:9f761a71-5301-47a6-b672-523287690cfb bdev_register:aa616c70-3dcd-442b-a2cd-386b7636fd14 bdev_register:5edb3c1d-d24a-48cd-a475-2eb8fbe903e5 bdev_register:MallocForCryptoBdev bdev_register:CryptoMallocBdev 00:07:09.458 11:17:52 json_config -- json_config/json_config.sh@71 -- # sort 00:07:09.458 11:17:52 json_config -- json_config/json_config.sh@72 -- # recorded_events=($(get_notifications | sort)) 00:07:09.458 11:17:52 json_config -- json_config/json_config.sh@72 -- # get_notifications 00:07:09.458 11:17:52 json_config -- json_config/json_config.sh@59 -- # local ev_type ev_ctx event_id 00:07:09.458 11:17:52 json_config -- json_config/json_config.sh@72 -- # sort 00:07:09.458 11:17:52 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:09.458 11:17:52 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:09.458 11:17:52 json_config -- json_config/json_config.sh@58 -- # jq -r '.[] | "\(.type):\(.ctx):\(.id)"' 00:07:09.458 11:17:52 json_config -- json_config/json_config.sh@58 -- # tgt_rpc notify_get_notifications -i 0 00:07:09.458 11:17:52 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock notify_get_notifications -i 0 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@62 -- # echo bdev_register:Nvme0n1 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@62 -- # echo bdev_register:Nvme0n1p1 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@62 -- # echo bdev_register:Nvme0n1p0 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@62 -- # echo bdev_register:Malloc3 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@62 -- # echo bdev_register:PTBdevFromMalloc3 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@62 -- # echo bdev_register:Null0 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@62 -- # echo bdev_register:Malloc0 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@62 -- # echo bdev_register:Malloc0p2 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@62 -- # echo bdev_register:Malloc0p1 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@62 -- # echo bdev_register:Malloc0p0 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@62 -- # echo bdev_register:Malloc1 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@62 -- # echo bdev_register:aio_disk 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@62 -- # echo bdev_register:635e020f-2df2-442c-a686-9c4a3693428b 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@62 -- # echo bdev_register:9f761a71-5301-47a6-b672-523287690cfb 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@62 -- # echo bdev_register:aa616c70-3dcd-442b-a2cd-386b7636fd14 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@62 -- # echo bdev_register:5edb3c1d-d24a-48cd-a475-2eb8fbe903e5 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@62 -- # echo bdev_register:MallocForCryptoBdev 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@62 -- # echo bdev_register:CryptoMallocBdev 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # IFS=: 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@61 -- # read -r ev_type ev_ctx event_id 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@74 -- # [[ bdev_register:5edb3c1d-d24a-48cd-a475-2eb8fbe903e5 bdev_register:635e020f-2df2-442c-a686-9c4a3693428b bdev_register:9f761a71-5301-47a6-b672-523287690cfb bdev_register:aa616c70-3dcd-442b-a2cd-386b7636fd14 bdev_register:aio_disk bdev_register:CryptoMallocBdev bdev_register:Malloc0 bdev_register:Malloc0p0 bdev_register:Malloc0p1 bdev_register:Malloc0p2 bdev_register:Malloc1 bdev_register:Malloc3 bdev_register:MallocForCryptoBdev bdev_register:Null0 bdev_register:Nvme0n1 bdev_register:Nvme0n1p0 bdev_register:Nvme0n1p1 bdev_register:PTBdevFromMalloc3 != \b\d\e\v\_\r\e\g\i\s\t\e\r\:\5\e\d\b\3\c\1\d\-\d\2\4\a\-\4\8\c\d\-\a\4\7\5\-\2\e\b\8\f\b\e\9\0\3\e\5\ \b\d\e\v\_\r\e\g\i\s\t\e\r\:\6\3\5\e\0\2\0\f\-\2\d\f\2\-\4\4\2\c\-\a\6\8\6\-\9\c\4\a\3\6\9\3\4\2\8\b\ \b\d\e\v\_\r\e\g\i\s\t\e\r\:\9\f\7\6\1\a\7\1\-\5\3\0\1\-\4\7\a\6\-\b\6\7\2\-\5\2\3\2\8\7\6\9\0\c\f\b\ \b\d\e\v\_\r\e\g\i\s\t\e\r\:\a\a\6\1\6\c\7\0\-\3\d\c\d\-\4\4\2\b\-\a\2\c\d\-\3\8\6\b\7\6\3\6\f\d\1\4\ \b\d\e\v\_\r\e\g\i\s\t\e\r\:\a\i\o\_\d\i\s\k\ \b\d\e\v\_\r\e\g\i\s\t\e\r\:\C\r\y\p\t\o\M\a\l\l\o\c\B\d\e\v\ \b\d\e\v\_\r\e\g\i\s\t\e\r\:\M\a\l\l\o\c\0\ \b\d\e\v\_\r\e\g\i\s\t\e\r\:\M\a\l\l\o\c\0\p\0\ \b\d\e\v\_\r\e\g\i\s\t\e\r\:\M\a\l\l\o\c\0\p\1\ \b\d\e\v\_\r\e\g\i\s\t\e\r\:\M\a\l\l\o\c\0\p\2\ \b\d\e\v\_\r\e\g\i\s\t\e\r\:\M\a\l\l\o\c\1\ \b\d\e\v\_\r\e\g\i\s\t\e\r\:\M\a\l\l\o\c\3\ \b\d\e\v\_\r\e\g\i\s\t\e\r\:\M\a\l\l\o\c\F\o\r\C\r\y\p\t\o\B\d\e\v\ \b\d\e\v\_\r\e\g\i\s\t\e\r\:\N\u\l\l\0\ \b\d\e\v\_\r\e\g\i\s\t\e\r\:\N\v\m\e\0\n\1\ \b\d\e\v\_\r\e\g\i\s\t\e\r\:\N\v\m\e\0\n\1\p\0\ \b\d\e\v\_\r\e\g\i\s\t\e\r\:\N\v\m\e\0\n\1\p\1\ \b\d\e\v\_\r\e\g\i\s\t\e\r\:\P\T\B\d\e\v\F\r\o\m\M\a\l\l\o\c\3 ]] 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@86 -- # cat 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@86 -- # printf ' %s\n' bdev_register:5edb3c1d-d24a-48cd-a475-2eb8fbe903e5 bdev_register:635e020f-2df2-442c-a686-9c4a3693428b bdev_register:9f761a71-5301-47a6-b672-523287690cfb bdev_register:aa616c70-3dcd-442b-a2cd-386b7636fd14 bdev_register:aio_disk bdev_register:CryptoMallocBdev bdev_register:Malloc0 bdev_register:Malloc0p0 bdev_register:Malloc0p1 bdev_register:Malloc0p2 bdev_register:Malloc1 bdev_register:Malloc3 bdev_register:MallocForCryptoBdev bdev_register:Null0 bdev_register:Nvme0n1 bdev_register:Nvme0n1p0 bdev_register:Nvme0n1p1 bdev_register:PTBdevFromMalloc3 00:07:09.717 Expected events matched: 00:07:09.717 bdev_register:5edb3c1d-d24a-48cd-a475-2eb8fbe903e5 00:07:09.717 bdev_register:635e020f-2df2-442c-a686-9c4a3693428b 00:07:09.717 bdev_register:9f761a71-5301-47a6-b672-523287690cfb 00:07:09.717 bdev_register:aa616c70-3dcd-442b-a2cd-386b7636fd14 00:07:09.717 bdev_register:aio_disk 00:07:09.717 bdev_register:CryptoMallocBdev 00:07:09.717 bdev_register:Malloc0 00:07:09.717 bdev_register:Malloc0p0 00:07:09.717 bdev_register:Malloc0p1 00:07:09.717 bdev_register:Malloc0p2 00:07:09.717 bdev_register:Malloc1 00:07:09.717 bdev_register:Malloc3 00:07:09.717 bdev_register:MallocForCryptoBdev 00:07:09.717 bdev_register:Null0 00:07:09.717 bdev_register:Nvme0n1 00:07:09.717 bdev_register:Nvme0n1p0 00:07:09.717 bdev_register:Nvme0n1p1 00:07:09.717 bdev_register:PTBdevFromMalloc3 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@180 -- # timing_exit create_bdev_subsystem_config 00:07:09.717 11:17:53 json_config -- common/autotest_common.sh@728 -- # xtrace_disable 00:07:09.717 11:17:53 json_config -- common/autotest_common.sh@10 -- # set +x 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@282 -- # [[ 0 -eq 1 ]] 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@286 -- # [[ 0 -eq 1 ]] 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@290 -- # [[ 0 -eq 1 ]] 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@293 -- # timing_exit json_config_setup_target 00:07:09.717 11:17:53 json_config -- common/autotest_common.sh@728 -- # xtrace_disable 00:07:09.717 11:17:53 json_config -- common/autotest_common.sh@10 -- # set +x 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@295 -- # [[ 0 -eq 1 ]] 00:07:09.717 11:17:53 json_config -- json_config/json_config.sh@300 -- # tgt_rpc bdev_malloc_create 8 512 --name MallocBdevForConfigChangeCheck 00:07:09.717 11:17:53 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 8 512 --name MallocBdevForConfigChangeCheck 00:07:09.975 MallocBdevForConfigChangeCheck 00:07:09.975 11:17:53 json_config -- json_config/json_config.sh@302 -- # timing_exit json_config_test_init 00:07:09.975 11:17:53 json_config -- common/autotest_common.sh@728 -- # xtrace_disable 00:07:09.975 11:17:53 json_config -- common/autotest_common.sh@10 -- # set +x 00:07:09.975 11:17:53 json_config -- json_config/json_config.sh@359 -- # tgt_rpc save_config 00:07:09.975 11:17:53 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:07:10.232 11:17:53 json_config -- json_config/json_config.sh@361 -- # echo 'INFO: shutting down applications...' 00:07:10.232 INFO: shutting down applications... 00:07:10.232 11:17:53 json_config -- json_config/json_config.sh@362 -- # [[ 0 -eq 1 ]] 00:07:10.232 11:17:53 json_config -- json_config/json_config.sh@368 -- # json_config_clear target 00:07:10.232 11:17:53 json_config -- json_config/json_config.sh@332 -- # [[ -n 22 ]] 00:07:10.232 11:17:53 json_config -- json_config/json_config.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/clear_config.py -s /var/tmp/spdk_tgt.sock clear_config 00:07:10.489 [2024-07-15 11:17:53.980839] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev Nvme0n1p0 being removed: closing lvstore lvs_test 00:07:13.768 Calling clear_iscsi_subsystem 00:07:13.768 Calling clear_nvmf_subsystem 00:07:13.768 Calling clear_nbd_subsystem 00:07:13.768 Calling clear_ublk_subsystem 00:07:13.768 Calling clear_vhost_blk_subsystem 00:07:13.768 Calling clear_vhost_scsi_subsystem 00:07:13.768 Calling clear_bdev_subsystem 00:07:13.768 11:17:56 json_config -- json_config/json_config.sh@337 -- # local config_filter=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/config_filter.py 00:07:13.768 11:17:56 json_config -- json_config/json_config.sh@343 -- # count=100 00:07:13.768 11:17:56 json_config -- json_config/json_config.sh@344 -- # '[' 100 -gt 0 ']' 00:07:13.768 11:17:56 json_config -- json_config/json_config.sh@345 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:07:13.768 11:17:56 json_config -- json_config/json_config.sh@345 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/config_filter.py -method delete_global_parameters 00:07:13.768 11:17:56 json_config -- json_config/json_config.sh@345 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/config_filter.py -method check_empty 00:07:13.768 11:17:57 json_config -- json_config/json_config.sh@345 -- # break 00:07:13.768 11:17:57 json_config -- json_config/json_config.sh@350 -- # '[' 100 -eq 0 ']' 00:07:13.768 11:17:57 json_config -- json_config/json_config.sh@369 -- # json_config_test_shutdown_app target 00:07:13.768 11:17:57 json_config -- json_config/common.sh@31 -- # local app=target 00:07:13.768 11:17:57 json_config -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:07:13.768 11:17:57 json_config -- json_config/common.sh@35 -- # [[ -n 824391 ]] 00:07:13.768 11:17:57 json_config -- json_config/common.sh@38 -- # kill -SIGINT 824391 00:07:13.768 11:17:57 json_config -- json_config/common.sh@40 -- # (( i = 0 )) 00:07:13.768 11:17:57 json_config -- json_config/common.sh@40 -- # (( i < 30 )) 00:07:13.768 11:17:57 json_config -- json_config/common.sh@41 -- # kill -0 824391 00:07:13.768 11:17:57 json_config -- json_config/common.sh@45 -- # sleep 0.5 00:07:14.336 11:17:57 json_config -- json_config/common.sh@40 -- # (( i++ )) 00:07:14.336 11:17:57 json_config -- json_config/common.sh@40 -- # (( i < 30 )) 00:07:14.336 11:17:57 json_config -- json_config/common.sh@41 -- # kill -0 824391 00:07:14.336 11:17:57 json_config -- json_config/common.sh@42 -- # app_pid["$app"]= 00:07:14.336 11:17:57 json_config -- json_config/common.sh@43 -- # break 00:07:14.336 11:17:57 json_config -- json_config/common.sh@48 -- # [[ -n '' ]] 00:07:14.336 11:17:57 json_config -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:07:14.336 SPDK target shutdown done 00:07:14.336 11:17:57 json_config -- json_config/json_config.sh@371 -- # echo 'INFO: relaunching applications...' 00:07:14.336 INFO: relaunching applications... 00:07:14.336 11:17:57 json_config -- json_config/json_config.sh@372 -- # json_config_test_start_app target --json /var/jenkins/workspace/crypto-phy-autotest/spdk/spdk_tgt_config.json 00:07:14.336 11:17:57 json_config -- json_config/common.sh@9 -- # local app=target 00:07:14.336 11:17:57 json_config -- json_config/common.sh@10 -- # shift 00:07:14.336 11:17:57 json_config -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:07:14.336 11:17:57 json_config -- json_config/common.sh@13 -- # [[ -z '' ]] 00:07:14.336 11:17:57 json_config -- json_config/common.sh@15 -- # local app_extra_params= 00:07:14.336 11:17:57 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:07:14.336 11:17:57 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:07:14.336 11:17:57 json_config -- json_config/common.sh@22 -- # app_pid["$app"]=827002 00:07:14.336 11:17:57 json_config -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:07:14.336 Waiting for target to run... 00:07:14.336 11:17:57 json_config -- json_config/common.sh@21 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /var/jenkins/workspace/crypto-phy-autotest/spdk/spdk_tgt_config.json 00:07:14.336 11:17:57 json_config -- json_config/common.sh@25 -- # waitforlisten 827002 /var/tmp/spdk_tgt.sock 00:07:14.336 11:17:57 json_config -- common/autotest_common.sh@829 -- # '[' -z 827002 ']' 00:07:14.336 11:17:57 json_config -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:07:14.336 11:17:57 json_config -- common/autotest_common.sh@834 -- # local max_retries=100 00:07:14.336 11:17:57 json_config -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:07:14.336 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:07:14.336 11:17:57 json_config -- common/autotest_common.sh@838 -- # xtrace_disable 00:07:14.336 11:17:57 json_config -- common/autotest_common.sh@10 -- # set +x 00:07:14.336 [2024-07-15 11:17:57.808363] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:07:14.336 [2024-07-15 11:17:57.808430] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid827002 ] 00:07:14.904 [2024-07-15 11:17:58.426889] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:15.161 [2024-07-15 11:17:58.525165] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:07:15.161 [2024-07-15 11:17:58.579325] accel_dpdk_cryptodev.c: 223:accel_dpdk_cryptodev_set_driver: *NOTICE*: Using driver crypto_aesni_mb 00:07:15.161 [2024-07-15 11:17:58.587363] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation encrypt will be assigned to module dpdk_cryptodev 00:07:15.161 [2024-07-15 11:17:58.595378] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation decrypt will be assigned to module dpdk_cryptodev 00:07:15.161 [2024-07-15 11:17:58.676614] accel_dpdk_cryptodev.c:1178:accel_dpdk_cryptodev_init: *NOTICE*: Found crypto devices: 97 00:07:17.689 [2024-07-15 11:18:00.887507] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc3 00:07:17.689 [2024-07-15 11:18:00.887564] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc3 00:07:17.689 [2024-07-15 11:18:00.887580] vbdev_passthru.c: 735:bdev_passthru_create_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:07:17.689 [2024-07-15 11:18:00.895523] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Nvme0n1 00:07:17.689 [2024-07-15 11:18:00.895550] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Nvme0n1 00:07:17.689 [2024-07-15 11:18:00.903539] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:07:17.689 [2024-07-15 11:18:00.903564] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:07:17.689 [2024-07-15 11:18:00.911574] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "CryptoMallocBdev_AES_CBC" 00:07:17.689 [2024-07-15 11:18:00.911601] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: MallocForCryptoBdev 00:07:17.689 [2024-07-15 11:18:00.911614] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:07:17.947 [2024-07-15 11:18:01.289769] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc3 00:07:17.947 [2024-07-15 11:18:01.289819] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:17.947 [2024-07-15 11:18:01.289837] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xdecb90 00:07:17.947 [2024-07-15 11:18:01.289849] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:17.947 [2024-07-15 11:18:01.290150] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:17.947 [2024-07-15 11:18:01.290168] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: PTBdevFromMalloc3 00:07:17.947 11:18:01 json_config -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:07:17.947 11:18:01 json_config -- common/autotest_common.sh@862 -- # return 0 00:07:17.947 11:18:01 json_config -- json_config/common.sh@26 -- # echo '' 00:07:17.947 00:07:17.947 11:18:01 json_config -- json_config/json_config.sh@373 -- # [[ 0 -eq 1 ]] 00:07:17.947 11:18:01 json_config -- json_config/json_config.sh@377 -- # echo 'INFO: Checking if target configuration is the same...' 00:07:17.947 INFO: Checking if target configuration is the same... 00:07:17.947 11:18:01 json_config -- json_config/json_config.sh@378 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/json_diff.sh /dev/fd/62 /var/jenkins/workspace/crypto-phy-autotest/spdk/spdk_tgt_config.json 00:07:17.947 11:18:01 json_config -- json_config/json_config.sh@378 -- # tgt_rpc save_config 00:07:17.947 11:18:01 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:07:17.947 + '[' 2 -ne 2 ']' 00:07:17.947 +++ dirname /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/json_diff.sh 00:07:17.947 ++ readlink -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/../.. 00:07:17.947 + rootdir=/var/jenkins/workspace/crypto-phy-autotest/spdk 00:07:17.947 +++ basename /dev/fd/62 00:07:17.947 ++ mktemp /tmp/62.XXX 00:07:17.947 + tmp_file_1=/tmp/62.Bqk 00:07:17.947 +++ basename /var/jenkins/workspace/crypto-phy-autotest/spdk/spdk_tgt_config.json 00:07:17.947 ++ mktemp /tmp/spdk_tgt_config.json.XXX 00:07:17.947 + tmp_file_2=/tmp/spdk_tgt_config.json.Ce6 00:07:17.947 + ret=0 00:07:17.947 + /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:07:18.205 + /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:07:18.462 + diff -u /tmp/62.Bqk /tmp/spdk_tgt_config.json.Ce6 00:07:18.462 + echo 'INFO: JSON config files are the same' 00:07:18.462 INFO: JSON config files are the same 00:07:18.462 + rm /tmp/62.Bqk /tmp/spdk_tgt_config.json.Ce6 00:07:18.462 + exit 0 00:07:18.462 11:18:01 json_config -- json_config/json_config.sh@379 -- # [[ 0 -eq 1 ]] 00:07:18.462 11:18:01 json_config -- json_config/json_config.sh@384 -- # echo 'INFO: changing configuration and checking if this can be detected...' 00:07:18.462 INFO: changing configuration and checking if this can be detected... 00:07:18.462 11:18:01 json_config -- json_config/json_config.sh@386 -- # tgt_rpc bdev_malloc_delete MallocBdevForConfigChangeCheck 00:07:18.462 11:18:01 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_delete MallocBdevForConfigChangeCheck 00:07:18.720 11:18:02 json_config -- json_config/json_config.sh@387 -- # tgt_rpc save_config 00:07:18.720 11:18:02 json_config -- json_config/json_config.sh@387 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/json_diff.sh /dev/fd/62 /var/jenkins/workspace/crypto-phy-autotest/spdk/spdk_tgt_config.json 00:07:18.720 11:18:02 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:07:18.720 + '[' 2 -ne 2 ']' 00:07:18.720 +++ dirname /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/json_diff.sh 00:07:18.720 ++ readlink -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/../.. 00:07:18.720 + rootdir=/var/jenkins/workspace/crypto-phy-autotest/spdk 00:07:18.720 +++ basename /dev/fd/62 00:07:18.720 ++ mktemp /tmp/62.XXX 00:07:18.720 + tmp_file_1=/tmp/62.A7u 00:07:18.720 +++ basename /var/jenkins/workspace/crypto-phy-autotest/spdk/spdk_tgt_config.json 00:07:18.720 ++ mktemp /tmp/spdk_tgt_config.json.XXX 00:07:18.720 + tmp_file_2=/tmp/spdk_tgt_config.json.sgr 00:07:18.720 + ret=0 00:07:18.720 + /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:07:18.978 + /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:07:18.978 + diff -u /tmp/62.A7u /tmp/spdk_tgt_config.json.sgr 00:07:18.978 + ret=1 00:07:18.978 + echo '=== Start of file: /tmp/62.A7u ===' 00:07:18.978 + cat /tmp/62.A7u 00:07:18.978 + echo '=== End of file: /tmp/62.A7u ===' 00:07:18.978 + echo '' 00:07:18.978 + echo '=== Start of file: /tmp/spdk_tgt_config.json.sgr ===' 00:07:18.978 + cat /tmp/spdk_tgt_config.json.sgr 00:07:18.978 + echo '=== End of file: /tmp/spdk_tgt_config.json.sgr ===' 00:07:18.978 + echo '' 00:07:18.978 + rm /tmp/62.A7u /tmp/spdk_tgt_config.json.sgr 00:07:18.978 + exit 1 00:07:18.978 11:18:02 json_config -- json_config/json_config.sh@391 -- # echo 'INFO: configuration change detected.' 00:07:18.978 INFO: configuration change detected. 00:07:18.978 11:18:02 json_config -- json_config/json_config.sh@394 -- # json_config_test_fini 00:07:18.978 11:18:02 json_config -- json_config/json_config.sh@306 -- # timing_enter json_config_test_fini 00:07:18.978 11:18:02 json_config -- common/autotest_common.sh@722 -- # xtrace_disable 00:07:18.978 11:18:02 json_config -- common/autotest_common.sh@10 -- # set +x 00:07:19.236 11:18:02 json_config -- json_config/json_config.sh@307 -- # local ret=0 00:07:19.236 11:18:02 json_config -- json_config/json_config.sh@309 -- # [[ -n '' ]] 00:07:19.236 11:18:02 json_config -- json_config/json_config.sh@317 -- # [[ -n 827002 ]] 00:07:19.236 11:18:02 json_config -- json_config/json_config.sh@320 -- # cleanup_bdev_subsystem_config 00:07:19.236 11:18:02 json_config -- json_config/json_config.sh@184 -- # timing_enter cleanup_bdev_subsystem_config 00:07:19.236 11:18:02 json_config -- common/autotest_common.sh@722 -- # xtrace_disable 00:07:19.236 11:18:02 json_config -- common/autotest_common.sh@10 -- # set +x 00:07:19.236 11:18:02 json_config -- json_config/json_config.sh@186 -- # [[ 1 -eq 1 ]] 00:07:19.236 11:18:02 json_config -- json_config/json_config.sh@187 -- # tgt_rpc bdev_lvol_delete lvs_test/clone0 00:07:19.236 11:18:02 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_lvol_delete lvs_test/clone0 00:07:19.236 11:18:02 json_config -- json_config/json_config.sh@188 -- # tgt_rpc bdev_lvol_delete lvs_test/lvol0 00:07:19.236 11:18:02 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_lvol_delete lvs_test/lvol0 00:07:19.494 11:18:03 json_config -- json_config/json_config.sh@189 -- # tgt_rpc bdev_lvol_delete lvs_test/snapshot0 00:07:19.494 11:18:03 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_lvol_delete lvs_test/snapshot0 00:07:19.752 11:18:03 json_config -- json_config/json_config.sh@190 -- # tgt_rpc bdev_lvol_delete_lvstore -l lvs_test 00:07:19.752 11:18:03 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_lvol_delete_lvstore -l lvs_test 00:07:20.009 11:18:03 json_config -- json_config/json_config.sh@193 -- # uname -s 00:07:20.009 11:18:03 json_config -- json_config/json_config.sh@193 -- # [[ Linux = Linux ]] 00:07:20.009 11:18:03 json_config -- json_config/json_config.sh@194 -- # rm -f /sample_aio 00:07:20.009 11:18:03 json_config -- json_config/json_config.sh@197 -- # [[ 0 -eq 1 ]] 00:07:20.009 11:18:03 json_config -- json_config/json_config.sh@201 -- # timing_exit cleanup_bdev_subsystem_config 00:07:20.009 11:18:03 json_config -- common/autotest_common.sh@728 -- # xtrace_disable 00:07:20.009 11:18:03 json_config -- common/autotest_common.sh@10 -- # set +x 00:07:20.009 11:18:03 json_config -- json_config/json_config.sh@323 -- # killprocess 827002 00:07:20.009 11:18:03 json_config -- common/autotest_common.sh@948 -- # '[' -z 827002 ']' 00:07:20.009 11:18:03 json_config -- common/autotest_common.sh@952 -- # kill -0 827002 00:07:20.009 11:18:03 json_config -- common/autotest_common.sh@953 -- # uname 00:07:20.009 11:18:03 json_config -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:07:20.009 11:18:03 json_config -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 827002 00:07:20.009 11:18:03 json_config -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:07:20.009 11:18:03 json_config -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:07:20.009 11:18:03 json_config -- common/autotest_common.sh@966 -- # echo 'killing process with pid 827002' 00:07:20.009 killing process with pid 827002 00:07:20.009 11:18:03 json_config -- common/autotest_common.sh@967 -- # kill 827002 00:07:20.009 11:18:03 json_config -- common/autotest_common.sh@972 -- # wait 827002 00:07:23.285 11:18:06 json_config -- json_config/json_config.sh@326 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/spdk_initiator_config.json /var/jenkins/workspace/crypto-phy-autotest/spdk/spdk_tgt_config.json 00:07:23.285 11:18:06 json_config -- json_config/json_config.sh@327 -- # timing_exit json_config_test_fini 00:07:23.285 11:18:06 json_config -- common/autotest_common.sh@728 -- # xtrace_disable 00:07:23.285 11:18:06 json_config -- common/autotest_common.sh@10 -- # set +x 00:07:23.285 11:18:06 json_config -- json_config/json_config.sh@328 -- # return 0 00:07:23.285 11:18:06 json_config -- json_config/json_config.sh@396 -- # echo 'INFO: Success' 00:07:23.285 INFO: Success 00:07:23.285 00:07:23.285 real 0m27.580s 00:07:23.285 user 0m32.805s 00:07:23.285 sys 0m4.066s 00:07:23.285 11:18:06 json_config -- common/autotest_common.sh@1124 -- # xtrace_disable 00:07:23.285 11:18:06 json_config -- common/autotest_common.sh@10 -- # set +x 00:07:23.285 ************************************ 00:07:23.285 END TEST json_config 00:07:23.285 ************************************ 00:07:23.285 11:18:06 -- common/autotest_common.sh@1142 -- # return 0 00:07:23.285 11:18:06 -- spdk/autotest.sh@173 -- # run_test json_config_extra_key /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/json_config_extra_key.sh 00:07:23.285 11:18:06 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:07:23.285 11:18:06 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:07:23.285 11:18:06 -- common/autotest_common.sh@10 -- # set +x 00:07:23.285 ************************************ 00:07:23.285 START TEST json_config_extra_key 00:07:23.285 ************************************ 00:07:23.285 11:18:06 json_config_extra_key -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/json_config_extra_key.sh 00:07:23.545 11:18:06 json_config_extra_key -- json_config/json_config_extra_key.sh@9 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/nvmf/common.sh 00:07:23.545 11:18:06 json_config_extra_key -- nvmf/common.sh@7 -- # uname -s 00:07:23.545 11:18:06 json_config_extra_key -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:07:23.545 11:18:06 json_config_extra_key -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:07:23.545 11:18:06 json_config_extra_key -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:07:23.545 11:18:06 json_config_extra_key -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:07:23.545 11:18:06 json_config_extra_key -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:07:23.545 11:18:06 json_config_extra_key -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:07:23.545 11:18:06 json_config_extra_key -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:07:23.545 11:18:06 json_config_extra_key -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:07:23.545 11:18:06 json_config_extra_key -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:07:23.545 11:18:06 json_config_extra_key -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:07:23.545 11:18:06 json_config_extra_key -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:809e2efd-7f71-e711-906e-0017a4403562 00:07:23.545 11:18:06 json_config_extra_key -- nvmf/common.sh@18 -- # NVME_HOSTID=809e2efd-7f71-e711-906e-0017a4403562 00:07:23.545 11:18:06 json_config_extra_key -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:07:23.545 11:18:06 json_config_extra_key -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:07:23.545 11:18:06 json_config_extra_key -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:07:23.545 11:18:06 json_config_extra_key -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:07:23.545 11:18:06 json_config_extra_key -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/common.sh 00:07:23.545 11:18:06 json_config_extra_key -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:07:23.545 11:18:06 json_config_extra_key -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:07:23.545 11:18:06 json_config_extra_key -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:07:23.546 11:18:06 json_config_extra_key -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:23.546 11:18:06 json_config_extra_key -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:23.546 11:18:06 json_config_extra_key -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:23.546 11:18:06 json_config_extra_key -- paths/export.sh@5 -- # export PATH 00:07:23.546 11:18:06 json_config_extra_key -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:23.546 11:18:06 json_config_extra_key -- nvmf/common.sh@47 -- # : 0 00:07:23.546 11:18:06 json_config_extra_key -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:07:23.546 11:18:06 json_config_extra_key -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:07:23.546 11:18:06 json_config_extra_key -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:07:23.546 11:18:06 json_config_extra_key -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:07:23.546 11:18:06 json_config_extra_key -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:07:23.546 11:18:06 json_config_extra_key -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:07:23.546 11:18:06 json_config_extra_key -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:07:23.546 11:18:06 json_config_extra_key -- nvmf/common.sh@51 -- # have_pci_nics=0 00:07:23.546 11:18:06 json_config_extra_key -- json_config/json_config_extra_key.sh@10 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/common.sh 00:07:23.546 11:18:06 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # app_pid=(['target']='') 00:07:23.546 11:18:06 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # declare -A app_pid 00:07:23.546 11:18:06 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock') 00:07:23.546 11:18:06 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # declare -A app_socket 00:07:23.546 11:18:06 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # app_params=(['target']='-m 0x1 -s 1024') 00:07:23.546 11:18:06 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # declare -A app_params 00:07:23.546 11:18:06 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # configs_path=(['target']='/var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/extra_key.json') 00:07:23.546 11:18:06 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # declare -A configs_path 00:07:23.546 11:18:06 json_config_extra_key -- json_config/json_config_extra_key.sh@22 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:07:23.546 11:18:06 json_config_extra_key -- json_config/json_config_extra_key.sh@24 -- # echo 'INFO: launching applications...' 00:07:23.546 INFO: launching applications... 00:07:23.546 11:18:06 json_config_extra_key -- json_config/json_config_extra_key.sh@25 -- # json_config_test_start_app target --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/extra_key.json 00:07:23.546 11:18:06 json_config_extra_key -- json_config/common.sh@9 -- # local app=target 00:07:23.546 11:18:06 json_config_extra_key -- json_config/common.sh@10 -- # shift 00:07:23.546 11:18:06 json_config_extra_key -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:07:23.546 11:18:06 json_config_extra_key -- json_config/common.sh@13 -- # [[ -z '' ]] 00:07:23.546 11:18:06 json_config_extra_key -- json_config/common.sh@15 -- # local app_extra_params= 00:07:23.546 11:18:06 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:07:23.546 11:18:06 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:07:23.546 11:18:06 json_config_extra_key -- json_config/common.sh@22 -- # app_pid["$app"]=828852 00:07:23.546 11:18:06 json_config_extra_key -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:07:23.546 Waiting for target to run... 00:07:23.546 11:18:06 json_config_extra_key -- json_config/common.sh@25 -- # waitforlisten 828852 /var/tmp/spdk_tgt.sock 00:07:23.546 11:18:06 json_config_extra_key -- common/autotest_common.sh@829 -- # '[' -z 828852 ']' 00:07:23.546 11:18:06 json_config_extra_key -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:07:23.546 11:18:06 json_config_extra_key -- json_config/common.sh@21 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/extra_key.json 00:07:23.546 11:18:06 json_config_extra_key -- common/autotest_common.sh@834 -- # local max_retries=100 00:07:23.546 11:18:06 json_config_extra_key -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:07:23.546 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:07:23.546 11:18:06 json_config_extra_key -- common/autotest_common.sh@838 -- # xtrace_disable 00:07:23.546 11:18:06 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:07:23.546 [2024-07-15 11:18:07.045154] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:07:23.546 [2024-07-15 11:18:07.045232] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid828852 ] 00:07:23.842 [2024-07-15 11:18:07.398779] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:24.099 [2024-07-15 11:18:07.491292] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:07:24.664 11:18:07 json_config_extra_key -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:07:24.664 11:18:07 json_config_extra_key -- common/autotest_common.sh@862 -- # return 0 00:07:24.664 11:18:07 json_config_extra_key -- json_config/common.sh@26 -- # echo '' 00:07:24.664 00:07:24.664 11:18:07 json_config_extra_key -- json_config/json_config_extra_key.sh@27 -- # echo 'INFO: shutting down applications...' 00:07:24.664 INFO: shutting down applications... 00:07:24.664 11:18:07 json_config_extra_key -- json_config/json_config_extra_key.sh@28 -- # json_config_test_shutdown_app target 00:07:24.664 11:18:07 json_config_extra_key -- json_config/common.sh@31 -- # local app=target 00:07:24.664 11:18:07 json_config_extra_key -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:07:24.664 11:18:07 json_config_extra_key -- json_config/common.sh@35 -- # [[ -n 828852 ]] 00:07:24.664 11:18:07 json_config_extra_key -- json_config/common.sh@38 -- # kill -SIGINT 828852 00:07:24.664 11:18:07 json_config_extra_key -- json_config/common.sh@40 -- # (( i = 0 )) 00:07:24.664 11:18:07 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:07:24.664 11:18:07 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 828852 00:07:24.664 11:18:07 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:07:24.920 11:18:08 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:07:24.920 11:18:08 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:07:24.920 11:18:08 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 828852 00:07:24.920 11:18:08 json_config_extra_key -- json_config/common.sh@42 -- # app_pid["$app"]= 00:07:24.920 11:18:08 json_config_extra_key -- json_config/common.sh@43 -- # break 00:07:24.920 11:18:08 json_config_extra_key -- json_config/common.sh@48 -- # [[ -n '' ]] 00:07:24.920 11:18:08 json_config_extra_key -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:07:24.920 SPDK target shutdown done 00:07:24.920 11:18:08 json_config_extra_key -- json_config/json_config_extra_key.sh@30 -- # echo Success 00:07:24.920 Success 00:07:24.920 00:07:24.920 real 0m1.606s 00:07:24.920 user 0m1.307s 00:07:24.920 sys 0m0.494s 00:07:24.920 11:18:08 json_config_extra_key -- common/autotest_common.sh@1124 -- # xtrace_disable 00:07:24.920 11:18:08 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:07:24.920 ************************************ 00:07:24.920 END TEST json_config_extra_key 00:07:24.920 ************************************ 00:07:25.176 11:18:08 -- common/autotest_common.sh@1142 -- # return 0 00:07:25.176 11:18:08 -- spdk/autotest.sh@174 -- # run_test alias_rpc /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:07:25.176 11:18:08 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:07:25.176 11:18:08 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:07:25.176 11:18:08 -- common/autotest_common.sh@10 -- # set +x 00:07:25.176 ************************************ 00:07:25.176 START TEST alias_rpc 00:07:25.176 ************************************ 00:07:25.176 11:18:08 alias_rpc -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:07:25.176 * Looking for test storage... 00:07:25.176 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/alias_rpc 00:07:25.176 11:18:08 alias_rpc -- alias_rpc/alias_rpc.sh@10 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:07:25.176 11:18:08 alias_rpc -- alias_rpc/alias_rpc.sh@13 -- # spdk_tgt_pid=829087 00:07:25.176 11:18:08 alias_rpc -- alias_rpc/alias_rpc.sh@14 -- # waitforlisten 829087 00:07:25.176 11:18:08 alias_rpc -- common/autotest_common.sh@829 -- # '[' -z 829087 ']' 00:07:25.176 11:18:08 alias_rpc -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:25.176 11:18:08 alias_rpc -- common/autotest_common.sh@834 -- # local max_retries=100 00:07:25.176 11:18:08 alias_rpc -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:25.176 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:25.176 11:18:08 alias_rpc -- common/autotest_common.sh@838 -- # xtrace_disable 00:07:25.176 11:18:08 alias_rpc -- alias_rpc/alias_rpc.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt 00:07:25.176 11:18:08 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:25.176 [2024-07-15 11:18:08.735064] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:07:25.176 [2024-07-15 11:18:08.735134] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid829087 ] 00:07:25.433 [2024-07-15 11:18:08.862785] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:25.433 [2024-07-15 11:18:08.967009] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:07:26.365 11:18:09 alias_rpc -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:07:26.365 11:18:09 alias_rpc -- common/autotest_common.sh@862 -- # return 0 00:07:26.365 11:18:09 alias_rpc -- alias_rpc/alias_rpc.sh@17 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py load_config -i 00:07:26.929 11:18:10 alias_rpc -- alias_rpc/alias_rpc.sh@19 -- # killprocess 829087 00:07:26.929 11:18:10 alias_rpc -- common/autotest_common.sh@948 -- # '[' -z 829087 ']' 00:07:26.929 11:18:10 alias_rpc -- common/autotest_common.sh@952 -- # kill -0 829087 00:07:26.929 11:18:10 alias_rpc -- common/autotest_common.sh@953 -- # uname 00:07:26.929 11:18:10 alias_rpc -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:07:26.929 11:18:10 alias_rpc -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 829087 00:07:26.929 11:18:10 alias_rpc -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:07:26.929 11:18:10 alias_rpc -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:07:26.929 11:18:10 alias_rpc -- common/autotest_common.sh@966 -- # echo 'killing process with pid 829087' 00:07:26.929 killing process with pid 829087 00:07:26.929 11:18:10 alias_rpc -- common/autotest_common.sh@967 -- # kill 829087 00:07:26.929 11:18:10 alias_rpc -- common/autotest_common.sh@972 -- # wait 829087 00:07:27.494 00:07:27.494 real 0m2.337s 00:07:27.494 user 0m2.970s 00:07:27.494 sys 0m0.639s 00:07:27.494 11:18:10 alias_rpc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:07:27.494 11:18:10 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:27.494 ************************************ 00:07:27.494 END TEST alias_rpc 00:07:27.494 ************************************ 00:07:27.494 11:18:10 -- common/autotest_common.sh@1142 -- # return 0 00:07:27.494 11:18:10 -- spdk/autotest.sh@176 -- # [[ 0 -eq 0 ]] 00:07:27.494 11:18:10 -- spdk/autotest.sh@177 -- # run_test spdkcli_tcp /var/jenkins/workspace/crypto-phy-autotest/spdk/test/spdkcli/tcp.sh 00:07:27.494 11:18:10 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:07:27.494 11:18:10 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:07:27.494 11:18:10 -- common/autotest_common.sh@10 -- # set +x 00:07:27.494 ************************************ 00:07:27.494 START TEST spdkcli_tcp 00:07:27.494 ************************************ 00:07:27.494 11:18:10 spdkcli_tcp -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/spdkcli/tcp.sh 00:07:27.494 * Looking for test storage... 00:07:27.494 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/spdkcli 00:07:27.494 11:18:11 spdkcli_tcp -- spdkcli/tcp.sh@9 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/spdkcli/common.sh 00:07:27.494 11:18:11 spdkcli_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/spdkcli/spdkcli_job.py 00:07:27.494 11:18:11 spdkcli_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/json_config/clear_config.py 00:07:27.494 11:18:11 spdkcli_tcp -- spdkcli/tcp.sh@18 -- # IP_ADDRESS=127.0.0.1 00:07:27.494 11:18:11 spdkcli_tcp -- spdkcli/tcp.sh@19 -- # PORT=9998 00:07:27.494 11:18:11 spdkcli_tcp -- spdkcli/tcp.sh@21 -- # trap 'err_cleanup; exit 1' SIGINT SIGTERM EXIT 00:07:27.494 11:18:11 spdkcli_tcp -- spdkcli/tcp.sh@23 -- # timing_enter run_spdk_tgt_tcp 00:07:27.494 11:18:11 spdkcli_tcp -- common/autotest_common.sh@722 -- # xtrace_disable 00:07:27.494 11:18:11 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:07:27.494 11:18:11 spdkcli_tcp -- spdkcli/tcp.sh@25 -- # spdk_tgt_pid=829488 00:07:27.494 11:18:11 spdkcli_tcp -- spdkcli/tcp.sh@24 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:07:27.494 11:18:11 spdkcli_tcp -- spdkcli/tcp.sh@27 -- # waitforlisten 829488 00:07:27.494 11:18:11 spdkcli_tcp -- common/autotest_common.sh@829 -- # '[' -z 829488 ']' 00:07:27.494 11:18:11 spdkcli_tcp -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:27.494 11:18:11 spdkcli_tcp -- common/autotest_common.sh@834 -- # local max_retries=100 00:07:27.494 11:18:11 spdkcli_tcp -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:27.494 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:27.494 11:18:11 spdkcli_tcp -- common/autotest_common.sh@838 -- # xtrace_disable 00:07:27.495 11:18:11 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:07:27.752 [2024-07-15 11:18:11.207421] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:07:27.752 [2024-07-15 11:18:11.207561] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid829488 ] 00:07:28.010 [2024-07-15 11:18:11.404035] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:07:28.010 [2024-07-15 11:18:11.505751] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:07:28.010 [2024-07-15 11:18:11.505757] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:07:28.574 11:18:12 spdkcli_tcp -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:07:28.574 11:18:12 spdkcli_tcp -- common/autotest_common.sh@862 -- # return 0 00:07:28.574 11:18:12 spdkcli_tcp -- spdkcli/tcp.sh@31 -- # socat_pid=829585 00:07:28.574 11:18:12 spdkcli_tcp -- spdkcli/tcp.sh@33 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -r 100 -t 2 -s 127.0.0.1 -p 9998 rpc_get_methods 00:07:28.574 11:18:12 spdkcli_tcp -- spdkcli/tcp.sh@30 -- # socat TCP-LISTEN:9998 UNIX-CONNECT:/var/tmp/spdk.sock 00:07:28.831 [ 00:07:28.831 "bdev_malloc_delete", 00:07:28.831 "bdev_malloc_create", 00:07:28.831 "bdev_null_resize", 00:07:28.831 "bdev_null_delete", 00:07:28.831 "bdev_null_create", 00:07:28.831 "bdev_nvme_cuse_unregister", 00:07:28.831 "bdev_nvme_cuse_register", 00:07:28.831 "bdev_opal_new_user", 00:07:28.831 "bdev_opal_set_lock_state", 00:07:28.831 "bdev_opal_delete", 00:07:28.831 "bdev_opal_get_info", 00:07:28.831 "bdev_opal_create", 00:07:28.831 "bdev_nvme_opal_revert", 00:07:28.831 "bdev_nvme_opal_init", 00:07:28.831 "bdev_nvme_send_cmd", 00:07:28.831 "bdev_nvme_get_path_iostat", 00:07:28.831 "bdev_nvme_get_mdns_discovery_info", 00:07:28.831 "bdev_nvme_stop_mdns_discovery", 00:07:28.831 "bdev_nvme_start_mdns_discovery", 00:07:28.831 "bdev_nvme_set_multipath_policy", 00:07:28.831 "bdev_nvme_set_preferred_path", 00:07:28.831 "bdev_nvme_get_io_paths", 00:07:28.831 "bdev_nvme_remove_error_injection", 00:07:28.831 "bdev_nvme_add_error_injection", 00:07:28.831 "bdev_nvme_get_discovery_info", 00:07:28.831 "bdev_nvme_stop_discovery", 00:07:28.831 "bdev_nvme_start_discovery", 00:07:28.831 "bdev_nvme_get_controller_health_info", 00:07:28.831 "bdev_nvme_disable_controller", 00:07:28.831 "bdev_nvme_enable_controller", 00:07:28.831 "bdev_nvme_reset_controller", 00:07:28.831 "bdev_nvme_get_transport_statistics", 00:07:28.831 "bdev_nvme_apply_firmware", 00:07:28.831 "bdev_nvme_detach_controller", 00:07:28.831 "bdev_nvme_get_controllers", 00:07:28.831 "bdev_nvme_attach_controller", 00:07:28.831 "bdev_nvme_set_hotplug", 00:07:28.831 "bdev_nvme_set_options", 00:07:28.831 "bdev_passthru_delete", 00:07:28.831 "bdev_passthru_create", 00:07:28.831 "bdev_lvol_set_parent_bdev", 00:07:28.831 "bdev_lvol_set_parent", 00:07:28.831 "bdev_lvol_check_shallow_copy", 00:07:28.831 "bdev_lvol_start_shallow_copy", 00:07:28.831 "bdev_lvol_grow_lvstore", 00:07:28.831 "bdev_lvol_get_lvols", 00:07:28.831 "bdev_lvol_get_lvstores", 00:07:28.831 "bdev_lvol_delete", 00:07:28.831 "bdev_lvol_set_read_only", 00:07:28.831 "bdev_lvol_resize", 00:07:28.831 "bdev_lvol_decouple_parent", 00:07:28.831 "bdev_lvol_inflate", 00:07:28.831 "bdev_lvol_rename", 00:07:28.831 "bdev_lvol_clone_bdev", 00:07:28.831 "bdev_lvol_clone", 00:07:28.831 "bdev_lvol_snapshot", 00:07:28.831 "bdev_lvol_create", 00:07:28.831 "bdev_lvol_delete_lvstore", 00:07:28.831 "bdev_lvol_rename_lvstore", 00:07:28.831 "bdev_lvol_create_lvstore", 00:07:28.831 "bdev_raid_set_options", 00:07:28.831 "bdev_raid_remove_base_bdev", 00:07:28.831 "bdev_raid_add_base_bdev", 00:07:28.831 "bdev_raid_delete", 00:07:28.831 "bdev_raid_create", 00:07:28.831 "bdev_raid_get_bdevs", 00:07:28.831 "bdev_error_inject_error", 00:07:28.831 "bdev_error_delete", 00:07:28.831 "bdev_error_create", 00:07:28.831 "bdev_split_delete", 00:07:28.831 "bdev_split_create", 00:07:28.831 "bdev_delay_delete", 00:07:28.831 "bdev_delay_create", 00:07:28.831 "bdev_delay_update_latency", 00:07:28.831 "bdev_zone_block_delete", 00:07:28.831 "bdev_zone_block_create", 00:07:28.831 "blobfs_create", 00:07:28.831 "blobfs_detect", 00:07:28.831 "blobfs_set_cache_size", 00:07:28.831 "bdev_crypto_delete", 00:07:28.831 "bdev_crypto_create", 00:07:28.831 "bdev_compress_delete", 00:07:28.831 "bdev_compress_create", 00:07:28.831 "bdev_compress_get_orphans", 00:07:28.831 "bdev_aio_delete", 00:07:28.831 "bdev_aio_rescan", 00:07:28.831 "bdev_aio_create", 00:07:28.831 "bdev_ftl_set_property", 00:07:28.831 "bdev_ftl_get_properties", 00:07:28.831 "bdev_ftl_get_stats", 00:07:28.831 "bdev_ftl_unmap", 00:07:28.831 "bdev_ftl_unload", 00:07:28.831 "bdev_ftl_delete", 00:07:28.831 "bdev_ftl_load", 00:07:28.831 "bdev_ftl_create", 00:07:28.831 "bdev_virtio_attach_controller", 00:07:28.831 "bdev_virtio_scsi_get_devices", 00:07:28.831 "bdev_virtio_detach_controller", 00:07:28.831 "bdev_virtio_blk_set_hotplug", 00:07:28.831 "bdev_iscsi_delete", 00:07:28.831 "bdev_iscsi_create", 00:07:28.832 "bdev_iscsi_set_options", 00:07:28.832 "accel_error_inject_error", 00:07:28.832 "ioat_scan_accel_module", 00:07:28.832 "dsa_scan_accel_module", 00:07:28.832 "iaa_scan_accel_module", 00:07:28.832 "dpdk_cryptodev_get_driver", 00:07:28.832 "dpdk_cryptodev_set_driver", 00:07:28.832 "dpdk_cryptodev_scan_accel_module", 00:07:28.832 "compressdev_scan_accel_module", 00:07:28.832 "keyring_file_remove_key", 00:07:28.832 "keyring_file_add_key", 00:07:28.832 "keyring_linux_set_options", 00:07:28.832 "iscsi_get_histogram", 00:07:28.832 "iscsi_enable_histogram", 00:07:28.832 "iscsi_set_options", 00:07:28.832 "iscsi_get_auth_groups", 00:07:28.832 "iscsi_auth_group_remove_secret", 00:07:28.832 "iscsi_auth_group_add_secret", 00:07:28.832 "iscsi_delete_auth_group", 00:07:28.832 "iscsi_create_auth_group", 00:07:28.832 "iscsi_set_discovery_auth", 00:07:28.832 "iscsi_get_options", 00:07:28.832 "iscsi_target_node_request_logout", 00:07:28.832 "iscsi_target_node_set_redirect", 00:07:28.832 "iscsi_target_node_set_auth", 00:07:28.832 "iscsi_target_node_add_lun", 00:07:28.832 "iscsi_get_stats", 00:07:28.832 "iscsi_get_connections", 00:07:28.832 "iscsi_portal_group_set_auth", 00:07:28.832 "iscsi_start_portal_group", 00:07:28.832 "iscsi_delete_portal_group", 00:07:28.832 "iscsi_create_portal_group", 00:07:28.832 "iscsi_get_portal_groups", 00:07:28.832 "iscsi_delete_target_node", 00:07:28.832 "iscsi_target_node_remove_pg_ig_maps", 00:07:28.832 "iscsi_target_node_add_pg_ig_maps", 00:07:28.832 "iscsi_create_target_node", 00:07:28.832 "iscsi_get_target_nodes", 00:07:28.832 "iscsi_delete_initiator_group", 00:07:28.832 "iscsi_initiator_group_remove_initiators", 00:07:28.832 "iscsi_initiator_group_add_initiators", 00:07:28.832 "iscsi_create_initiator_group", 00:07:28.832 "iscsi_get_initiator_groups", 00:07:28.832 "nvmf_set_crdt", 00:07:28.832 "nvmf_set_config", 00:07:28.832 "nvmf_set_max_subsystems", 00:07:28.832 "nvmf_stop_mdns_prr", 00:07:28.832 "nvmf_publish_mdns_prr", 00:07:28.832 "nvmf_subsystem_get_listeners", 00:07:28.832 "nvmf_subsystem_get_qpairs", 00:07:28.832 "nvmf_subsystem_get_controllers", 00:07:28.832 "nvmf_get_stats", 00:07:28.832 "nvmf_get_transports", 00:07:28.832 "nvmf_create_transport", 00:07:28.832 "nvmf_get_targets", 00:07:28.832 "nvmf_delete_target", 00:07:28.832 "nvmf_create_target", 00:07:28.832 "nvmf_subsystem_allow_any_host", 00:07:28.832 "nvmf_subsystem_remove_host", 00:07:28.832 "nvmf_subsystem_add_host", 00:07:28.832 "nvmf_ns_remove_host", 00:07:28.832 "nvmf_ns_add_host", 00:07:28.832 "nvmf_subsystem_remove_ns", 00:07:28.832 "nvmf_subsystem_add_ns", 00:07:28.832 "nvmf_subsystem_listener_set_ana_state", 00:07:28.832 "nvmf_discovery_get_referrals", 00:07:28.832 "nvmf_discovery_remove_referral", 00:07:28.832 "nvmf_discovery_add_referral", 00:07:28.832 "nvmf_subsystem_remove_listener", 00:07:28.832 "nvmf_subsystem_add_listener", 00:07:28.832 "nvmf_delete_subsystem", 00:07:28.832 "nvmf_create_subsystem", 00:07:28.832 "nvmf_get_subsystems", 00:07:28.832 "env_dpdk_get_mem_stats", 00:07:28.832 "nbd_get_disks", 00:07:28.832 "nbd_stop_disk", 00:07:28.832 "nbd_start_disk", 00:07:28.832 "ublk_recover_disk", 00:07:28.832 "ublk_get_disks", 00:07:28.832 "ublk_stop_disk", 00:07:28.832 "ublk_start_disk", 00:07:28.832 "ublk_destroy_target", 00:07:28.832 "ublk_create_target", 00:07:28.832 "virtio_blk_create_transport", 00:07:28.832 "virtio_blk_get_transports", 00:07:28.832 "vhost_controller_set_coalescing", 00:07:28.832 "vhost_get_controllers", 00:07:28.832 "vhost_delete_controller", 00:07:28.832 "vhost_create_blk_controller", 00:07:28.832 "vhost_scsi_controller_remove_target", 00:07:28.832 "vhost_scsi_controller_add_target", 00:07:28.832 "vhost_start_scsi_controller", 00:07:28.832 "vhost_create_scsi_controller", 00:07:28.832 "thread_set_cpumask", 00:07:28.832 "framework_get_governor", 00:07:28.832 "framework_get_scheduler", 00:07:28.832 "framework_set_scheduler", 00:07:28.832 "framework_get_reactors", 00:07:28.832 "thread_get_io_channels", 00:07:28.832 "thread_get_pollers", 00:07:28.832 "thread_get_stats", 00:07:28.832 "framework_monitor_context_switch", 00:07:28.832 "spdk_kill_instance", 00:07:28.832 "log_enable_timestamps", 00:07:28.832 "log_get_flags", 00:07:28.832 "log_clear_flag", 00:07:28.832 "log_set_flag", 00:07:28.832 "log_get_level", 00:07:28.832 "log_set_level", 00:07:28.832 "log_get_print_level", 00:07:28.832 "log_set_print_level", 00:07:28.832 "framework_enable_cpumask_locks", 00:07:28.832 "framework_disable_cpumask_locks", 00:07:28.832 "framework_wait_init", 00:07:28.832 "framework_start_init", 00:07:28.832 "scsi_get_devices", 00:07:28.832 "bdev_get_histogram", 00:07:28.832 "bdev_enable_histogram", 00:07:28.832 "bdev_set_qos_limit", 00:07:28.832 "bdev_set_qd_sampling_period", 00:07:28.832 "bdev_get_bdevs", 00:07:28.832 "bdev_reset_iostat", 00:07:28.832 "bdev_get_iostat", 00:07:28.832 "bdev_examine", 00:07:28.832 "bdev_wait_for_examine", 00:07:28.832 "bdev_set_options", 00:07:28.832 "notify_get_notifications", 00:07:28.832 "notify_get_types", 00:07:28.832 "accel_get_stats", 00:07:28.832 "accel_set_options", 00:07:28.832 "accel_set_driver", 00:07:28.832 "accel_crypto_key_destroy", 00:07:28.832 "accel_crypto_keys_get", 00:07:28.832 "accel_crypto_key_create", 00:07:28.832 "accel_assign_opc", 00:07:28.832 "accel_get_module_info", 00:07:28.832 "accel_get_opc_assignments", 00:07:28.832 "vmd_rescan", 00:07:28.832 "vmd_remove_device", 00:07:28.832 "vmd_enable", 00:07:28.832 "sock_get_default_impl", 00:07:28.832 "sock_set_default_impl", 00:07:28.832 "sock_impl_set_options", 00:07:28.832 "sock_impl_get_options", 00:07:28.832 "iobuf_get_stats", 00:07:28.832 "iobuf_set_options", 00:07:28.832 "framework_get_pci_devices", 00:07:28.832 "framework_get_config", 00:07:28.832 "framework_get_subsystems", 00:07:28.832 "trace_get_info", 00:07:28.832 "trace_get_tpoint_group_mask", 00:07:28.832 "trace_disable_tpoint_group", 00:07:28.832 "trace_enable_tpoint_group", 00:07:28.832 "trace_clear_tpoint_mask", 00:07:28.832 "trace_set_tpoint_mask", 00:07:28.832 "keyring_get_keys", 00:07:28.832 "spdk_get_version", 00:07:28.832 "rpc_get_methods" 00:07:28.832 ] 00:07:28.832 11:18:12 spdkcli_tcp -- spdkcli/tcp.sh@35 -- # timing_exit run_spdk_tgt_tcp 00:07:28.832 11:18:12 spdkcli_tcp -- common/autotest_common.sh@728 -- # xtrace_disable 00:07:28.832 11:18:12 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:07:28.832 11:18:12 spdkcli_tcp -- spdkcli/tcp.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:07:28.832 11:18:12 spdkcli_tcp -- spdkcli/tcp.sh@38 -- # killprocess 829488 00:07:28.832 11:18:12 spdkcli_tcp -- common/autotest_common.sh@948 -- # '[' -z 829488 ']' 00:07:28.832 11:18:12 spdkcli_tcp -- common/autotest_common.sh@952 -- # kill -0 829488 00:07:28.832 11:18:12 spdkcli_tcp -- common/autotest_common.sh@953 -- # uname 00:07:28.832 11:18:12 spdkcli_tcp -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:07:28.832 11:18:12 spdkcli_tcp -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 829488 00:07:28.832 11:18:12 spdkcli_tcp -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:07:28.832 11:18:12 spdkcli_tcp -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:07:28.832 11:18:12 spdkcli_tcp -- common/autotest_common.sh@966 -- # echo 'killing process with pid 829488' 00:07:28.832 killing process with pid 829488 00:07:28.832 11:18:12 spdkcli_tcp -- common/autotest_common.sh@967 -- # kill 829488 00:07:29.088 11:18:12 spdkcli_tcp -- common/autotest_common.sh@972 -- # wait 829488 00:07:29.347 00:07:29.347 real 0m1.855s 00:07:29.347 user 0m3.197s 00:07:29.347 sys 0m0.674s 00:07:29.347 11:18:12 spdkcli_tcp -- common/autotest_common.sh@1124 -- # xtrace_disable 00:07:29.347 11:18:12 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:07:29.347 ************************************ 00:07:29.347 END TEST spdkcli_tcp 00:07:29.347 ************************************ 00:07:29.347 11:18:12 -- common/autotest_common.sh@1142 -- # return 0 00:07:29.347 11:18:12 -- spdk/autotest.sh@180 -- # run_test dpdk_mem_utility /var/jenkins/workspace/crypto-phy-autotest/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:07:29.347 11:18:12 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:07:29.347 11:18:12 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:07:29.347 11:18:12 -- common/autotest_common.sh@10 -- # set +x 00:07:29.347 ************************************ 00:07:29.347 START TEST dpdk_mem_utility 00:07:29.347 ************************************ 00:07:29.347 11:18:12 dpdk_mem_utility -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:07:29.604 * Looking for test storage... 00:07:29.604 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/dpdk_memory_utility 00:07:29.604 11:18:13 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@10 -- # MEM_SCRIPT=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/dpdk_mem_info.py 00:07:29.604 11:18:13 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@13 -- # spdkpid=829740 00:07:29.604 11:18:13 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@15 -- # waitforlisten 829740 00:07:29.604 11:18:13 dpdk_mem_utility -- common/autotest_common.sh@829 -- # '[' -z 829740 ']' 00:07:29.604 11:18:13 dpdk_mem_utility -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:29.604 11:18:13 dpdk_mem_utility -- common/autotest_common.sh@834 -- # local max_retries=100 00:07:29.604 11:18:13 dpdk_mem_utility -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:29.604 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:29.604 11:18:13 dpdk_mem_utility -- common/autotest_common.sh@838 -- # xtrace_disable 00:07:29.604 11:18:13 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:07:29.604 11:18:13 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt 00:07:29.604 [2024-07-15 11:18:13.145486] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:07:29.604 [2024-07-15 11:18:13.145623] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid829740 ] 00:07:29.860 [2024-07-15 11:18:13.339770] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:29.860 [2024-07-15 11:18:13.440606] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:07:30.791 11:18:14 dpdk_mem_utility -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:07:30.791 11:18:14 dpdk_mem_utility -- common/autotest_common.sh@862 -- # return 0 00:07:30.791 11:18:14 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@17 -- # trap 'killprocess $spdkpid' SIGINT SIGTERM EXIT 00:07:30.791 11:18:14 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@19 -- # rpc_cmd env_dpdk_get_mem_stats 00:07:30.791 11:18:14 dpdk_mem_utility -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:30.791 11:18:14 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:07:30.791 { 00:07:30.791 "filename": "/tmp/spdk_mem_dump.txt" 00:07:30.791 } 00:07:30.791 11:18:14 dpdk_mem_utility -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:30.791 11:18:14 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@21 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/dpdk_mem_info.py 00:07:30.791 DPDK memory size 816.000000 MiB in 2 heap(s) 00:07:30.791 2 heaps totaling size 816.000000 MiB 00:07:30.791 size: 814.000000 MiB heap id: 0 00:07:30.791 size: 2.000000 MiB heap id: 1 00:07:30.791 end heaps---------- 00:07:30.791 8 mempools totaling size 598.116089 MiB 00:07:30.791 size: 212.674988 MiB name: PDU_immediate_data_Pool 00:07:30.791 size: 158.602051 MiB name: PDU_data_out_Pool 00:07:30.791 size: 84.521057 MiB name: bdev_io_829740 00:07:30.791 size: 51.011292 MiB name: evtpool_829740 00:07:30.791 size: 50.003479 MiB name: msgpool_829740 00:07:30.791 size: 21.763794 MiB name: PDU_Pool 00:07:30.791 size: 19.513306 MiB name: SCSI_TASK_Pool 00:07:30.791 size: 0.026123 MiB name: Session_Pool 00:07:30.791 end mempools------- 00:07:30.791 201 memzones totaling size 4.176453 MiB 00:07:30.791 size: 1.000366 MiB name: RG_ring_0_829740 00:07:30.791 size: 1.000366 MiB name: RG_ring_1_829740 00:07:30.791 size: 1.000366 MiB name: RG_ring_4_829740 00:07:30.791 size: 1.000366 MiB name: RG_ring_5_829740 00:07:30.791 size: 0.125366 MiB name: RG_ring_2_829740 00:07:30.791 size: 0.015991 MiB name: RG_ring_3_829740 00:07:30.791 size: 0.001160 MiB name: QAT_SYM_CAPA_GEN_1 00:07:30.791 size: 0.000305 MiB name: 0000:3d:01.0_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3d:01.1_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3d:01.2_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3d:01.3_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3d:01.4_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3d:01.5_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3d:01.6_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3d:01.7_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3d:02.0_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3d:02.1_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3d:02.2_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3d:02.3_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3d:02.4_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3d:02.5_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3d:02.6_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3d:02.7_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3f:01.0_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3f:01.1_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3f:01.2_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3f:01.3_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3f:01.4_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3f:01.5_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3f:01.6_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3f:01.7_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3f:02.0_qat 00:07:30.791 size: 0.000305 MiB name: 0000:3f:02.1_qat 00:07:30.792 size: 0.000305 MiB name: 0000:3f:02.2_qat 00:07:30.792 size: 0.000305 MiB name: 0000:3f:02.3_qat 00:07:30.792 size: 0.000305 MiB name: 0000:3f:02.4_qat 00:07:30.792 size: 0.000305 MiB name: 0000:3f:02.5_qat 00:07:30.792 size: 0.000305 MiB name: 0000:3f:02.6_qat 00:07:30.792 size: 0.000305 MiB name: 0000:3f:02.7_qat 00:07:30.792 size: 0.000305 MiB name: 0000:da:01.0_qat 00:07:30.792 size: 0.000305 MiB name: 0000:da:01.1_qat 00:07:30.792 size: 0.000305 MiB name: 0000:da:01.2_qat 00:07:30.792 size: 0.000305 MiB name: 0000:da:01.3_qat 00:07:30.792 size: 0.000305 MiB name: 0000:da:01.4_qat 00:07:30.792 size: 0.000305 MiB name: 0000:da:01.5_qat 00:07:30.792 size: 0.000305 MiB name: 0000:da:01.6_qat 00:07:30.792 size: 0.000305 MiB name: 0000:da:01.7_qat 00:07:30.792 size: 0.000305 MiB name: 0000:da:02.0_qat 00:07:30.792 size: 0.000305 MiB name: 0000:da:02.1_qat 00:07:30.792 size: 0.000305 MiB name: 0000:da:02.2_qat 00:07:30.792 size: 0.000305 MiB name: 0000:da:02.3_qat 00:07:30.792 size: 0.000305 MiB name: 0000:da:02.4_qat 00:07:30.792 size: 0.000305 MiB name: 0000:da:02.5_qat 00:07:30.792 size: 0.000305 MiB name: 0000:da:02.6_qat 00:07:30.792 size: 0.000305 MiB name: 0000:da:02.7_qat 00:07:30.792 size: 0.000183 MiB name: QAT_ASYM_CAPA_GEN_1 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_0 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_1 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_0 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_2 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_3 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_1 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_4 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_5 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_2 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_6 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_7 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_3 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_8 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_9 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_4 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_10 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_11 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_5 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_12 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_13 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_6 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_14 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_15 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_7 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_16 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_17 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_8 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_18 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_19 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_9 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_20 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_21 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_10 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_22 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_23 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_11 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_24 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_25 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_12 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_26 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_27 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_13 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_28 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_29 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_14 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_30 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_31 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_15 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_32 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_33 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_16 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_34 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_35 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_17 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_36 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_37 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_18 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_38 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_39 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_19 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_40 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_41 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_20 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_42 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_43 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_21 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_44 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_45 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_22 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_46 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_47 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_23 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_48 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_49 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_24 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_50 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_51 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_25 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_52 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_53 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_26 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_54 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_55 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_27 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_56 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_57 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_28 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_58 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_59 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_29 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_60 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_61 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_30 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_62 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_63 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_31 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_64 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_65 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_32 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_66 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_67 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_33 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_68 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_69 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_34 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_70 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_71 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_35 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_72 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_73 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_36 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_74 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_75 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_37 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_76 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_77 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_38 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_78 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_79 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_39 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_80 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_81 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_40 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_82 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_83 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_41 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_84 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_85 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_42 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_86 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_87 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_43 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_88 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_89 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_44 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_90 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_91 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_45 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_92 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_93 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_46 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_94 00:07:30.792 size: 0.000122 MiB name: rte_cryptodev_data_95 00:07:30.792 size: 0.000122 MiB name: rte_compressdev_data_47 00:07:30.792 size: 0.000061 MiB name: QAT_COMP_CAPA_GEN_1 00:07:30.792 end memzones------- 00:07:30.792 11:18:14 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@23 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/dpdk_mem_info.py -m 0 00:07:31.052 heap id: 0 total size: 814.000000 MiB number of busy elements: 518 number of free elements: 14 00:07:31.052 list of free elements. size: 11.815002 MiB 00:07:31.052 element at address: 0x200000400000 with size: 1.999512 MiB 00:07:31.052 element at address: 0x200018e00000 with size: 0.999878 MiB 00:07:31.052 element at address: 0x200019000000 with size: 0.999878 MiB 00:07:31.052 element at address: 0x200003e00000 with size: 0.996460 MiB 00:07:31.052 element at address: 0x200031c00000 with size: 0.994446 MiB 00:07:31.052 element at address: 0x200013800000 with size: 0.978882 MiB 00:07:31.052 element at address: 0x200007000000 with size: 0.960022 MiB 00:07:31.052 element at address: 0x200019200000 with size: 0.937256 MiB 00:07:31.052 element at address: 0x20001aa00000 with size: 0.583252 MiB 00:07:31.052 element at address: 0x200003a00000 with size: 0.498535 MiB 00:07:31.052 element at address: 0x20000b200000 with size: 0.491272 MiB 00:07:31.052 element at address: 0x200000800000 with size: 0.486694 MiB 00:07:31.052 element at address: 0x200019400000 with size: 0.485840 MiB 00:07:31.052 element at address: 0x200027e00000 with size: 0.403076 MiB 00:07:31.052 list of standard malloc elements. size: 199.876709 MiB 00:07:31.052 element at address: 0x20000b3fff80 with size: 132.000122 MiB 00:07:31.052 element at address: 0x2000071fff80 with size: 64.000122 MiB 00:07:31.052 element at address: 0x200018efff80 with size: 1.000122 MiB 00:07:31.052 element at address: 0x2000190fff80 with size: 1.000122 MiB 00:07:31.052 element at address: 0x2000192fff80 with size: 1.000122 MiB 00:07:31.052 element at address: 0x2000003d9f00 with size: 0.140747 MiB 00:07:31.052 element at address: 0x2000192eff00 with size: 0.062622 MiB 00:07:31.052 element at address: 0x2000003fdf80 with size: 0.007935 MiB 00:07:31.052 element at address: 0x200000330b40 with size: 0.004395 MiB 00:07:31.052 element at address: 0x2000003340c0 with size: 0.004395 MiB 00:07:31.052 element at address: 0x200000337640 with size: 0.004395 MiB 00:07:31.053 element at address: 0x20000033abc0 with size: 0.004395 MiB 00:07:31.053 element at address: 0x20000033e140 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003416c0 with size: 0.004395 MiB 00:07:31.053 element at address: 0x200000344c40 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003481c0 with size: 0.004395 MiB 00:07:31.053 element at address: 0x20000034b740 with size: 0.004395 MiB 00:07:31.053 element at address: 0x20000034ecc0 with size: 0.004395 MiB 00:07:31.053 element at address: 0x200000352240 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003557c0 with size: 0.004395 MiB 00:07:31.053 element at address: 0x200000358d40 with size: 0.004395 MiB 00:07:31.053 element at address: 0x20000035c2c0 with size: 0.004395 MiB 00:07:31.053 element at address: 0x20000035f840 with size: 0.004395 MiB 00:07:31.053 element at address: 0x200000362dc0 with size: 0.004395 MiB 00:07:31.053 element at address: 0x200000366880 with size: 0.004395 MiB 00:07:31.053 element at address: 0x20000036a340 with size: 0.004395 MiB 00:07:31.053 element at address: 0x20000036de00 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003718c0 with size: 0.004395 MiB 00:07:31.053 element at address: 0x200000375380 with size: 0.004395 MiB 00:07:31.053 element at address: 0x200000378e40 with size: 0.004395 MiB 00:07:31.053 element at address: 0x20000037c900 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003803c0 with size: 0.004395 MiB 00:07:31.053 element at address: 0x200000383e80 with size: 0.004395 MiB 00:07:31.053 element at address: 0x200000387940 with size: 0.004395 MiB 00:07:31.053 element at address: 0x20000038b400 with size: 0.004395 MiB 00:07:31.053 element at address: 0x20000038eec0 with size: 0.004395 MiB 00:07:31.053 element at address: 0x200000392980 with size: 0.004395 MiB 00:07:31.053 element at address: 0x200000396440 with size: 0.004395 MiB 00:07:31.053 element at address: 0x200000399f00 with size: 0.004395 MiB 00:07:31.053 element at address: 0x20000039d9c0 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003a1480 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003a4f40 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003a8a00 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003ac4c0 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003aff80 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003b3a40 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003b7500 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003bafc0 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003bea80 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003c2540 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003c6000 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003c9ac0 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003cd580 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003d1040 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003d4b00 with size: 0.004395 MiB 00:07:31.053 element at address: 0x2000003d8d00 with size: 0.004395 MiB 00:07:31.053 element at address: 0x20000032ea40 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000032fac0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000331fc0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000333040 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000335540 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003365c0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000338ac0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000339b40 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000033c040 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000033d0c0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000033f5c0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000340640 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000342b40 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000343bc0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003460c0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000347140 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000349640 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000034a6c0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000034cbc0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000034dc40 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000350140 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003511c0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003536c0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000354740 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000356c40 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000357cc0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000035a1c0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000035b240 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000035d740 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000035e7c0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000360cc0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000361d40 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000364780 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000365800 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000368240 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003692c0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000036bd00 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000036cd80 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000036f7c0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000370840 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000373280 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000374300 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000376d40 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000377dc0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000037a800 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000037b880 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000037e2c0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000037f340 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000381d80 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000382e00 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000385840 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003868c0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000389300 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000038a380 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000038cdc0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000038de40 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000390880 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000391900 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000394340 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003953c0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000397e00 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000398e80 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000039b8c0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000039c940 with size: 0.004028 MiB 00:07:31.053 element at address: 0x20000039f380 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003a0400 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003a2e40 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003a3ec0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003a6900 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003a7980 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003aa3c0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003ab440 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003ade80 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003aef00 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003b1940 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003b29c0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003b5400 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003b6480 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003b8ec0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003b9f40 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003bc980 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003bda00 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003c0440 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003c14c0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003c3f00 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003c4f80 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003c79c0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003c8a40 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003cb480 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003cc500 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003cef40 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003cffc0 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003d2a00 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003d3a80 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003d6c00 with size: 0.004028 MiB 00:07:31.053 element at address: 0x2000003d7c80 with size: 0.004028 MiB 00:07:31.053 element at address: 0x200000204f80 with size: 0.000305 MiB 00:07:31.053 element at address: 0x200000200000 with size: 0.000183 MiB 00:07:31.053 element at address: 0x2000002000c0 with size: 0.000183 MiB 00:07:31.053 element at address: 0x200000200180 with size: 0.000183 MiB 00:07:31.053 element at address: 0x200000200240 with size: 0.000183 MiB 00:07:31.053 element at address: 0x200000200300 with size: 0.000183 MiB 00:07:31.053 element at address: 0x2000002003c0 with size: 0.000183 MiB 00:07:31.053 element at address: 0x200000200480 with size: 0.000183 MiB 00:07:31.053 element at address: 0x200000200540 with size: 0.000183 MiB 00:07:31.053 element at address: 0x200000200600 with size: 0.000183 MiB 00:07:31.053 element at address: 0x2000002006c0 with size: 0.000183 MiB 00:07:31.053 element at address: 0x200000200780 with size: 0.000183 MiB 00:07:31.053 element at address: 0x200000200840 with size: 0.000183 MiB 00:07:31.053 element at address: 0x200000200900 with size: 0.000183 MiB 00:07:31.053 element at address: 0x2000002009c0 with size: 0.000183 MiB 00:07:31.053 element at address: 0x200000200a80 with size: 0.000183 MiB 00:07:31.053 element at address: 0x200000200b40 with size: 0.000183 MiB 00:07:31.053 element at address: 0x200000200c00 with size: 0.000183 MiB 00:07:31.053 element at address: 0x200000200cc0 with size: 0.000183 MiB 00:07:31.053 element at address: 0x200000200d80 with size: 0.000183 MiB 00:07:31.053 element at address: 0x200000200e40 with size: 0.000183 MiB 00:07:31.053 element at address: 0x200000200f00 with size: 0.000183 MiB 00:07:31.053 element at address: 0x200000200fc0 with size: 0.000183 MiB 00:07:31.053 element at address: 0x200000201080 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000201140 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000201200 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000002012c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000201380 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000201440 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000201500 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000002015c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000201680 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000201740 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000201800 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000002018c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000201980 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000201a40 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000201b00 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000201bc0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000201c80 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000201d40 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000201e00 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000201ec0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000201f80 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000202040 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000202100 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000002021c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000202280 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000202340 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000202400 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000002024c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000202580 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000202640 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000202700 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000002027c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000202880 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000202940 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000202a00 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000202ac0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000202b80 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000202c40 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000202d00 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000202dc0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000202e80 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000202f40 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000203000 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000002030c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000203180 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000203240 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000203300 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000002033c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000203480 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000203540 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000203600 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000002036c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000203780 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000203840 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000203900 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000002039c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000203a80 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000203b40 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000203c00 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000203cc0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000203d80 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000203e40 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000203f00 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000203fc0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000204080 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000204140 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000204200 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000002042c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000204380 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000204440 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000204500 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000002045c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000204680 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000204740 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000204800 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000002048c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000204980 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000204a40 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000204b00 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000204bc0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000204c80 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000204d40 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000204e00 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000204ec0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000002050c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000205180 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000205240 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000205300 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000002053c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000205480 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000205540 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000205600 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000002056c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000205780 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000205840 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000205900 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000002059c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000205a80 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000205b40 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000205c00 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000205cc0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000205d80 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000205e40 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000205f00 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000205fc0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000206080 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000206140 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000206200 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000002062c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000002064c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000020a780 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022aa40 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022ab00 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022abc0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022ac80 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022ad40 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022ae00 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022aec0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022af80 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022b040 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022b100 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022b1c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022b280 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022b340 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022b400 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022b4c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022b580 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022b640 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022b700 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022b900 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022b9c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022ba80 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022bb40 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022bc00 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022bcc0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022bd80 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022be40 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022bf00 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022bfc0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022c080 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022c140 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022c200 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022c2c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022c380 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022c440 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000022c500 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000032e700 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000032e7c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000331d40 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000003352c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000338840 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000033bdc0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000033f340 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000003428c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000345e40 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000003493c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000034c940 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000034fec0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000353440 with size: 0.000183 MiB 00:07:31.054 element at address: 0x2000003569c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000359f40 with size: 0.000183 MiB 00:07:31.054 element at address: 0x20000035d4c0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000360a40 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000363fc0 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000364180 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000364240 with size: 0.000183 MiB 00:07:31.054 element at address: 0x200000364400 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000367a80 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000367c40 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000367d00 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000367ec0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000036b540 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000036b700 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000036b7c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000036b980 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000036f000 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000036f1c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000036f280 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000036f440 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000372ac0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000372c80 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000372d40 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000372f00 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000376580 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000376740 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000376800 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003769c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000037a040 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000037a200 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000037a2c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000037a480 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000037db00 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000037dcc0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000037dd80 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000037df40 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003815c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000381780 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000381840 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000381a00 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000385080 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000385240 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000385300 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003854c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000388b40 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000388d00 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000388dc0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000388f80 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000038c600 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000038c7c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000038c880 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000038ca40 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003900c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000390280 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000390340 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000390500 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000393b80 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000393d40 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000393e00 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000393fc0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000397640 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000397800 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003978c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200000397a80 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000039b100 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000039b2c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000039b380 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000039b540 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000039ebc0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000039ed80 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000039ee40 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000039f000 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003a2680 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003a2840 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003a2900 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003a2ac0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003a6140 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003a6300 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003a63c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003a6580 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003a9c00 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003a9dc0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003a9e80 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003aa040 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003ad6c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003ad880 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003ad940 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003adb00 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003b1180 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003b1340 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003b1400 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003b15c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003b4c40 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003b4e00 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003b4ec0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003b5080 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003b8700 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003b88c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003b8980 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003b8b40 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003bc1c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003bc380 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003bc440 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003bc600 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003bfc80 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003bfe40 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003bff00 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003c00c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003c3740 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003c3900 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003c39c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003c3b80 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003c7200 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003c73c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003c7480 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003c7640 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003cacc0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003cae80 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003caf40 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003cb100 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003ce780 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003ce940 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003cea00 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003cebc0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003d2240 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003d2400 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003d24c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003d2680 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003d5dc0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003d64c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003d6580 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000003d6880 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000087c980 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000087ca40 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000087cb00 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000087cbc0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000087cc80 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000087cd40 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000087ce00 with size: 0.000183 MiB 00:07:31.055 element at address: 0x20000087cec0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x2000008fd180 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e67300 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e673c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6dfc0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6e1c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6e280 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6e340 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6e400 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6e4c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6e580 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6e640 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6e700 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6e7c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6e880 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6e940 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6ea00 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6eac0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6eb80 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6ec40 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6ed00 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6edc0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6ee80 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6ef40 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6f000 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6f0c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6f180 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6f240 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6f300 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6f3c0 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6f480 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6f540 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6f600 with size: 0.000183 MiB 00:07:31.055 element at address: 0x200027e6f6c0 with size: 0.000183 MiB 00:07:31.056 element at address: 0x200027e6f780 with size: 0.000183 MiB 00:07:31.056 element at address: 0x200027e6f840 with size: 0.000183 MiB 00:07:31.056 element at address: 0x200027e6f900 with size: 0.000183 MiB 00:07:31.056 element at address: 0x200027e6f9c0 with size: 0.000183 MiB 00:07:31.056 element at address: 0x200027e6fa80 with size: 0.000183 MiB 00:07:31.056 element at address: 0x200027e6fb40 with size: 0.000183 MiB 00:07:31.056 element at address: 0x200027e6fc00 with size: 0.000183 MiB 00:07:31.056 element at address: 0x200027e6fcc0 with size: 0.000183 MiB 00:07:31.056 element at address: 0x200027e6fd80 with size: 0.000183 MiB 00:07:31.056 element at address: 0x200027e6fe40 with size: 0.000183 MiB 00:07:31.056 element at address: 0x200027e6ff00 with size: 0.000183 MiB 00:07:31.056 list of memzone associated elements. size: 602.308289 MiB 00:07:31.056 element at address: 0x20001aa95500 with size: 211.416748 MiB 00:07:31.056 associated memzone info: size: 211.416626 MiB name: MP_PDU_immediate_data_Pool_0 00:07:31.056 element at address: 0x200027e6ffc0 with size: 157.562561 MiB 00:07:31.056 associated memzone info: size: 157.562439 MiB name: MP_PDU_data_out_Pool_0 00:07:31.056 element at address: 0x2000139fab80 with size: 84.020630 MiB 00:07:31.056 associated memzone info: size: 84.020508 MiB name: MP_bdev_io_829740_0 00:07:31.056 element at address: 0x2000009ff380 with size: 48.003052 MiB 00:07:31.056 associated memzone info: size: 48.002930 MiB name: MP_evtpool_829740_0 00:07:31.056 element at address: 0x200003fff380 with size: 48.003052 MiB 00:07:31.056 associated memzone info: size: 48.002930 MiB name: MP_msgpool_829740_0 00:07:31.056 element at address: 0x2000195be940 with size: 20.255554 MiB 00:07:31.056 associated memzone info: size: 20.255432 MiB name: MP_PDU_Pool_0 00:07:31.056 element at address: 0x200031dfeb40 with size: 18.005066 MiB 00:07:31.056 associated memzone info: size: 18.004944 MiB name: MP_SCSI_TASK_Pool_0 00:07:31.056 element at address: 0x2000005ffe00 with size: 2.000488 MiB 00:07:31.056 associated memzone info: size: 2.000366 MiB name: RG_MP_evtpool_829740 00:07:31.056 element at address: 0x200003bffe00 with size: 2.000488 MiB 00:07:31.056 associated memzone info: size: 2.000366 MiB name: RG_MP_msgpool_829740 00:07:31.056 element at address: 0x20000022c5c0 with size: 1.008118 MiB 00:07:31.056 associated memzone info: size: 1.007996 MiB name: MP_evtpool_829740 00:07:31.056 element at address: 0x20000b2fde40 with size: 1.008118 MiB 00:07:31.056 associated memzone info: size: 1.007996 MiB name: MP_PDU_Pool 00:07:31.056 element at address: 0x2000194bc800 with size: 1.008118 MiB 00:07:31.056 associated memzone info: size: 1.007996 MiB name: MP_PDU_immediate_data_Pool 00:07:31.056 element at address: 0x2000070fde40 with size: 1.008118 MiB 00:07:31.056 associated memzone info: size: 1.007996 MiB name: MP_PDU_data_out_Pool 00:07:31.056 element at address: 0x2000008fd240 with size: 1.008118 MiB 00:07:31.056 associated memzone info: size: 1.007996 MiB name: MP_SCSI_TASK_Pool 00:07:31.056 element at address: 0x200003eff180 with size: 1.000488 MiB 00:07:31.056 associated memzone info: size: 1.000366 MiB name: RG_ring_0_829740 00:07:31.056 element at address: 0x200003affc00 with size: 1.000488 MiB 00:07:31.056 associated memzone info: size: 1.000366 MiB name: RG_ring_1_829740 00:07:31.056 element at address: 0x2000138fa980 with size: 1.000488 MiB 00:07:31.056 associated memzone info: size: 1.000366 MiB name: RG_ring_4_829740 00:07:31.056 element at address: 0x200031cfe940 with size: 1.000488 MiB 00:07:31.056 associated memzone info: size: 1.000366 MiB name: RG_ring_5_829740 00:07:31.056 element at address: 0x200003a7fa00 with size: 0.500488 MiB 00:07:31.056 associated memzone info: size: 0.500366 MiB name: RG_MP_bdev_io_829740 00:07:31.056 element at address: 0x20000b27dc40 with size: 0.500488 MiB 00:07:31.056 associated memzone info: size: 0.500366 MiB name: RG_MP_PDU_Pool 00:07:31.056 element at address: 0x20000087cf80 with size: 0.500488 MiB 00:07:31.056 associated memzone info: size: 0.500366 MiB name: RG_MP_SCSI_TASK_Pool 00:07:31.056 element at address: 0x20001947c600 with size: 0.250488 MiB 00:07:31.056 associated memzone info: size: 0.250366 MiB name: RG_MP_PDU_immediate_data_Pool 00:07:31.056 element at address: 0x20000020a840 with size: 0.125488 MiB 00:07:31.056 associated memzone info: size: 0.125366 MiB name: RG_ring_2_829740 00:07:31.056 element at address: 0x2000070f5c40 with size: 0.031738 MiB 00:07:31.056 associated memzone info: size: 0.031616 MiB name: RG_MP_PDU_data_out_Pool 00:07:31.056 element at address: 0x200027e67480 with size: 0.023743 MiB 00:07:31.056 associated memzone info: size: 0.023621 MiB name: MP_Session_Pool_0 00:07:31.056 element at address: 0x200000206580 with size: 0.016113 MiB 00:07:31.056 associated memzone info: size: 0.015991 MiB name: RG_ring_3_829740 00:07:31.056 element at address: 0x200027e6d5c0 with size: 0.002441 MiB 00:07:31.056 associated memzone info: size: 0.002319 MiB name: RG_MP_Session_Pool 00:07:31.056 element at address: 0x2000003d5f80 with size: 0.001282 MiB 00:07:31.056 associated memzone info: size: 0.001160 MiB name: QAT_SYM_CAPA_GEN_1 00:07:31.056 element at address: 0x2000003d6a40 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3d:01.0_qat 00:07:31.056 element at address: 0x2000003d2840 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3d:01.1_qat 00:07:31.056 element at address: 0x2000003ced80 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3d:01.2_qat 00:07:31.056 element at address: 0x2000003cb2c0 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3d:01.3_qat 00:07:31.056 element at address: 0x2000003c7800 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3d:01.4_qat 00:07:31.056 element at address: 0x2000003c3d40 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3d:01.5_qat 00:07:31.056 element at address: 0x2000003c0280 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3d:01.6_qat 00:07:31.056 element at address: 0x2000003bc7c0 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3d:01.7_qat 00:07:31.056 element at address: 0x2000003b8d00 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3d:02.0_qat 00:07:31.056 element at address: 0x2000003b5240 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3d:02.1_qat 00:07:31.056 element at address: 0x2000003b1780 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3d:02.2_qat 00:07:31.056 element at address: 0x2000003adcc0 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3d:02.3_qat 00:07:31.056 element at address: 0x2000003aa200 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3d:02.4_qat 00:07:31.056 element at address: 0x2000003a6740 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3d:02.5_qat 00:07:31.056 element at address: 0x2000003a2c80 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3d:02.6_qat 00:07:31.056 element at address: 0x20000039f1c0 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3d:02.7_qat 00:07:31.056 element at address: 0x20000039b700 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3f:01.0_qat 00:07:31.056 element at address: 0x200000397c40 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3f:01.1_qat 00:07:31.056 element at address: 0x200000394180 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3f:01.2_qat 00:07:31.056 element at address: 0x2000003906c0 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3f:01.3_qat 00:07:31.056 element at address: 0x20000038cc00 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3f:01.4_qat 00:07:31.056 element at address: 0x200000389140 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3f:01.5_qat 00:07:31.056 element at address: 0x200000385680 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3f:01.6_qat 00:07:31.056 element at address: 0x200000381bc0 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3f:01.7_qat 00:07:31.056 element at address: 0x20000037e100 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3f:02.0_qat 00:07:31.056 element at address: 0x20000037a640 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3f:02.1_qat 00:07:31.056 element at address: 0x200000376b80 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3f:02.2_qat 00:07:31.056 element at address: 0x2000003730c0 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3f:02.3_qat 00:07:31.056 element at address: 0x20000036f600 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3f:02.4_qat 00:07:31.056 element at address: 0x20000036bb40 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3f:02.5_qat 00:07:31.056 element at address: 0x200000368080 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3f:02.6_qat 00:07:31.056 element at address: 0x2000003645c0 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:3f:02.7_qat 00:07:31.056 element at address: 0x200000360b00 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:da:01.0_qat 00:07:31.056 element at address: 0x20000035d580 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:da:01.1_qat 00:07:31.056 element at address: 0x20000035a000 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:da:01.2_qat 00:07:31.056 element at address: 0x200000356a80 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:da:01.3_qat 00:07:31.056 element at address: 0x200000353500 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:da:01.4_qat 00:07:31.056 element at address: 0x20000034ff80 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:da:01.5_qat 00:07:31.056 element at address: 0x20000034ca00 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:da:01.6_qat 00:07:31.056 element at address: 0x200000349480 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:da:01.7_qat 00:07:31.056 element at address: 0x200000345f00 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:da:02.0_qat 00:07:31.056 element at address: 0x200000342980 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:da:02.1_qat 00:07:31.056 element at address: 0x20000033f400 with size: 0.000427 MiB 00:07:31.056 associated memzone info: size: 0.000305 MiB name: 0000:da:02.2_qat 00:07:31.056 element at address: 0x20000033be80 with size: 0.000427 MiB 00:07:31.057 associated memzone info: size: 0.000305 MiB name: 0000:da:02.3_qat 00:07:31.057 element at address: 0x200000338900 with size: 0.000427 MiB 00:07:31.057 associated memzone info: size: 0.000305 MiB name: 0000:da:02.4_qat 00:07:31.057 element at address: 0x200000335380 with size: 0.000427 MiB 00:07:31.057 associated memzone info: size: 0.000305 MiB name: 0000:da:02.5_qat 00:07:31.057 element at address: 0x200000331e00 with size: 0.000427 MiB 00:07:31.057 associated memzone info: size: 0.000305 MiB name: 0000:da:02.6_qat 00:07:31.057 element at address: 0x20000032e880 with size: 0.000427 MiB 00:07:31.057 associated memzone info: size: 0.000305 MiB name: 0000:da:02.7_qat 00:07:31.057 element at address: 0x2000003d6740 with size: 0.000305 MiB 00:07:31.057 associated memzone info: size: 0.000183 MiB name: QAT_ASYM_CAPA_GEN_1 00:07:31.057 element at address: 0x20000022b7c0 with size: 0.000305 MiB 00:07:31.057 associated memzone info: size: 0.000183 MiB name: MP_msgpool_829740 00:07:31.057 element at address: 0x200000206380 with size: 0.000305 MiB 00:07:31.057 associated memzone info: size: 0.000183 MiB name: MP_bdev_io_829740 00:07:31.057 element at address: 0x200027e6e080 with size: 0.000305 MiB 00:07:31.057 associated memzone info: size: 0.000183 MiB name: MP_Session_Pool 00:07:31.057 element at address: 0x2000003d6940 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_0 00:07:31.057 element at address: 0x2000003d6640 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_1 00:07:31.057 element at address: 0x2000003d5e80 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_0 00:07:31.057 element at address: 0x2000003d2740 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_2 00:07:31.057 element at address: 0x2000003d2580 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_3 00:07:31.057 element at address: 0x2000003d2300 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_1 00:07:31.057 element at address: 0x2000003cec80 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_4 00:07:31.057 element at address: 0x2000003ceac0 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_5 00:07:31.057 element at address: 0x2000003ce840 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_2 00:07:31.057 element at address: 0x2000003cb1c0 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_6 00:07:31.057 element at address: 0x2000003cb000 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_7 00:07:31.057 element at address: 0x2000003cad80 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_3 00:07:31.057 element at address: 0x2000003c7700 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_8 00:07:31.057 element at address: 0x2000003c7540 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_9 00:07:31.057 element at address: 0x2000003c72c0 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_4 00:07:31.057 element at address: 0x2000003c3c40 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_10 00:07:31.057 element at address: 0x2000003c3a80 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_11 00:07:31.057 element at address: 0x2000003c3800 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_5 00:07:31.057 element at address: 0x2000003c0180 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_12 00:07:31.057 element at address: 0x2000003bffc0 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_13 00:07:31.057 element at address: 0x2000003bfd40 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_6 00:07:31.057 element at address: 0x2000003bc6c0 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_14 00:07:31.057 element at address: 0x2000003bc500 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_15 00:07:31.057 element at address: 0x2000003bc280 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_7 00:07:31.057 element at address: 0x2000003b8c00 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_16 00:07:31.057 element at address: 0x2000003b8a40 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_17 00:07:31.057 element at address: 0x2000003b87c0 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_8 00:07:31.057 element at address: 0x2000003b5140 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_18 00:07:31.057 element at address: 0x2000003b4f80 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_19 00:07:31.057 element at address: 0x2000003b4d00 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_9 00:07:31.057 element at address: 0x2000003b1680 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_20 00:07:31.057 element at address: 0x2000003b14c0 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_21 00:07:31.057 element at address: 0x2000003b1240 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_10 00:07:31.057 element at address: 0x2000003adbc0 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_22 00:07:31.057 element at address: 0x2000003ada00 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_23 00:07:31.057 element at address: 0x2000003ad780 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_11 00:07:31.057 element at address: 0x2000003aa100 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_24 00:07:31.057 element at address: 0x2000003a9f40 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_25 00:07:31.057 element at address: 0x2000003a9cc0 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_12 00:07:31.057 element at address: 0x2000003a6640 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_26 00:07:31.057 element at address: 0x2000003a6480 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_27 00:07:31.057 element at address: 0x2000003a6200 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_13 00:07:31.057 element at address: 0x2000003a2b80 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_28 00:07:31.057 element at address: 0x2000003a29c0 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_29 00:07:31.057 element at address: 0x2000003a2740 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_14 00:07:31.057 element at address: 0x20000039f0c0 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_30 00:07:31.057 element at address: 0x20000039ef00 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_31 00:07:31.057 element at address: 0x20000039ec80 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_15 00:07:31.057 element at address: 0x20000039b600 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_32 00:07:31.057 element at address: 0x20000039b440 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_33 00:07:31.057 element at address: 0x20000039b1c0 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_16 00:07:31.057 element at address: 0x200000397b40 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_34 00:07:31.057 element at address: 0x200000397980 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_35 00:07:31.057 element at address: 0x200000397700 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_17 00:07:31.057 element at address: 0x200000394080 with size: 0.000244 MiB 00:07:31.057 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_36 00:07:31.057 element at address: 0x200000393ec0 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_37 00:07:31.058 element at address: 0x200000393c40 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_18 00:07:31.058 element at address: 0x2000003905c0 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_38 00:07:31.058 element at address: 0x200000390400 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_39 00:07:31.058 element at address: 0x200000390180 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_19 00:07:31.058 element at address: 0x20000038cb00 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_40 00:07:31.058 element at address: 0x20000038c940 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_41 00:07:31.058 element at address: 0x20000038c6c0 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_20 00:07:31.058 element at address: 0x200000389040 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_42 00:07:31.058 element at address: 0x200000388e80 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_43 00:07:31.058 element at address: 0x200000388c00 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_21 00:07:31.058 element at address: 0x200000385580 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_44 00:07:31.058 element at address: 0x2000003853c0 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_45 00:07:31.058 element at address: 0x200000385140 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_22 00:07:31.058 element at address: 0x200000381ac0 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_46 00:07:31.058 element at address: 0x200000381900 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_47 00:07:31.058 element at address: 0x200000381680 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_23 00:07:31.058 element at address: 0x20000037e000 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_48 00:07:31.058 element at address: 0x20000037de40 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_49 00:07:31.058 element at address: 0x20000037dbc0 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_24 00:07:31.058 element at address: 0x20000037a540 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_50 00:07:31.058 element at address: 0x20000037a380 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_51 00:07:31.058 element at address: 0x20000037a100 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_25 00:07:31.058 element at address: 0x200000376a80 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_52 00:07:31.058 element at address: 0x2000003768c0 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_53 00:07:31.058 element at address: 0x200000376640 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_26 00:07:31.058 element at address: 0x200000372fc0 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_54 00:07:31.058 element at address: 0x200000372e00 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_55 00:07:31.058 element at address: 0x200000372b80 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_27 00:07:31.058 element at address: 0x20000036f500 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_56 00:07:31.058 element at address: 0x20000036f340 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_57 00:07:31.058 element at address: 0x20000036f0c0 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_28 00:07:31.058 element at address: 0x20000036ba40 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_58 00:07:31.058 element at address: 0x20000036b880 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_59 00:07:31.058 element at address: 0x20000036b600 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_29 00:07:31.058 element at address: 0x200000367f80 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_60 00:07:31.058 element at address: 0x200000367dc0 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_61 00:07:31.058 element at address: 0x200000367b40 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_30 00:07:31.058 element at address: 0x2000003644c0 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_62 00:07:31.058 element at address: 0x200000364300 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_cryptodev_data_63 00:07:31.058 element at address: 0x200000364080 with size: 0.000244 MiB 00:07:31.058 associated memzone info: size: 0.000122 MiB name: rte_compressdev_data_31 00:07:31.058 element at address: 0x2000003d5d00 with size: 0.000183 MiB 00:07:31.058 associated memzone info: size: 0.000061 MiB name: QAT_COMP_CAPA_GEN_1 00:07:31.058 11:18:14 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@25 -- # trap - SIGINT SIGTERM EXIT 00:07:31.058 11:18:14 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@26 -- # killprocess 829740 00:07:31.058 11:18:14 dpdk_mem_utility -- common/autotest_common.sh@948 -- # '[' -z 829740 ']' 00:07:31.058 11:18:14 dpdk_mem_utility -- common/autotest_common.sh@952 -- # kill -0 829740 00:07:31.058 11:18:14 dpdk_mem_utility -- common/autotest_common.sh@953 -- # uname 00:07:31.058 11:18:14 dpdk_mem_utility -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:07:31.058 11:18:14 dpdk_mem_utility -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 829740 00:07:31.315 11:18:14 dpdk_mem_utility -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:07:31.315 11:18:14 dpdk_mem_utility -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:07:31.315 11:18:14 dpdk_mem_utility -- common/autotest_common.sh@966 -- # echo 'killing process with pid 829740' 00:07:31.315 killing process with pid 829740 00:07:31.315 11:18:14 dpdk_mem_utility -- common/autotest_common.sh@967 -- # kill 829740 00:07:31.315 11:18:14 dpdk_mem_utility -- common/autotest_common.sh@972 -- # wait 829740 00:07:31.572 00:07:31.572 real 0m2.126s 00:07:31.572 user 0m2.507s 00:07:31.572 sys 0m0.691s 00:07:31.572 11:18:15 dpdk_mem_utility -- common/autotest_common.sh@1124 -- # xtrace_disable 00:07:31.572 11:18:15 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:07:31.572 ************************************ 00:07:31.572 END TEST dpdk_mem_utility 00:07:31.572 ************************************ 00:07:31.572 11:18:15 -- common/autotest_common.sh@1142 -- # return 0 00:07:31.572 11:18:15 -- spdk/autotest.sh@181 -- # run_test event /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/event.sh 00:07:31.572 11:18:15 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:07:31.572 11:18:15 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:07:31.572 11:18:15 -- common/autotest_common.sh@10 -- # set +x 00:07:31.572 ************************************ 00:07:31.572 START TEST event 00:07:31.572 ************************************ 00:07:31.572 11:18:15 event -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/event.sh 00:07:31.829 * Looking for test storage... 00:07:31.829 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event 00:07:31.829 11:18:15 event -- event/event.sh@9 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbd_common.sh 00:07:31.829 11:18:15 event -- bdev/nbd_common.sh@6 -- # set -e 00:07:31.829 11:18:15 event -- event/event.sh@45 -- # run_test event_perf /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:07:31.829 11:18:15 event -- common/autotest_common.sh@1099 -- # '[' 6 -le 1 ']' 00:07:31.829 11:18:15 event -- common/autotest_common.sh@1105 -- # xtrace_disable 00:07:31.829 11:18:15 event -- common/autotest_common.sh@10 -- # set +x 00:07:31.829 ************************************ 00:07:31.829 START TEST event_perf 00:07:31.829 ************************************ 00:07:31.829 11:18:15 event.event_perf -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:07:31.829 Running I/O for 1 seconds...[2024-07-15 11:18:15.297097] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:07:31.829 [2024-07-15 11:18:15.297163] [ DPDK EAL parameters: event_perf --no-shconf -c 0xF --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid830141 ] 00:07:32.086 [2024-07-15 11:18:15.427860] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:07:32.086 [2024-07-15 11:18:15.531170] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:07:32.086 [2024-07-15 11:18:15.531255] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:07:32.086 [2024-07-15 11:18:15.531331] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:07:32.086 [2024-07-15 11:18:15.531335] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:07:33.458 Running I/O for 1 seconds... 00:07:33.458 lcore 0: 175290 00:07:33.458 lcore 1: 175288 00:07:33.458 lcore 2: 175288 00:07:33.458 lcore 3: 175290 00:07:33.458 done. 00:07:33.458 00:07:33.458 real 0m1.363s 00:07:33.458 user 0m4.206s 00:07:33.458 sys 0m0.149s 00:07:33.458 11:18:16 event.event_perf -- common/autotest_common.sh@1124 -- # xtrace_disable 00:07:33.458 11:18:16 event.event_perf -- common/autotest_common.sh@10 -- # set +x 00:07:33.458 ************************************ 00:07:33.458 END TEST event_perf 00:07:33.458 ************************************ 00:07:33.458 11:18:16 event -- common/autotest_common.sh@1142 -- # return 0 00:07:33.458 11:18:16 event -- event/event.sh@46 -- # run_test event_reactor /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/reactor/reactor -t 1 00:07:33.458 11:18:16 event -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:07:33.459 11:18:16 event -- common/autotest_common.sh@1105 -- # xtrace_disable 00:07:33.459 11:18:16 event -- common/autotest_common.sh@10 -- # set +x 00:07:33.459 ************************************ 00:07:33.459 START TEST event_reactor 00:07:33.459 ************************************ 00:07:33.459 11:18:16 event.event_reactor -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/reactor/reactor -t 1 00:07:33.459 [2024-07-15 11:18:16.743871] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:07:33.459 [2024-07-15 11:18:16.743946] [ DPDK EAL parameters: reactor --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid830347 ] 00:07:33.459 [2024-07-15 11:18:16.874152] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:33.459 [2024-07-15 11:18:16.977295] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:07:34.830 test_start 00:07:34.830 oneshot 00:07:34.830 tick 100 00:07:34.830 tick 100 00:07:34.830 tick 250 00:07:34.830 tick 100 00:07:34.830 tick 100 00:07:34.830 tick 250 00:07:34.830 tick 100 00:07:34.830 tick 500 00:07:34.830 tick 100 00:07:34.830 tick 100 00:07:34.830 tick 250 00:07:34.830 tick 100 00:07:34.830 tick 100 00:07:34.830 test_end 00:07:34.830 00:07:34.830 real 0m1.357s 00:07:34.830 user 0m1.200s 00:07:34.830 sys 0m0.150s 00:07:34.830 11:18:18 event.event_reactor -- common/autotest_common.sh@1124 -- # xtrace_disable 00:07:34.830 11:18:18 event.event_reactor -- common/autotest_common.sh@10 -- # set +x 00:07:34.830 ************************************ 00:07:34.830 END TEST event_reactor 00:07:34.830 ************************************ 00:07:34.830 11:18:18 event -- common/autotest_common.sh@1142 -- # return 0 00:07:34.830 11:18:18 event -- event/event.sh@47 -- # run_test event_reactor_perf /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/reactor_perf/reactor_perf -t 1 00:07:34.830 11:18:18 event -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:07:34.830 11:18:18 event -- common/autotest_common.sh@1105 -- # xtrace_disable 00:07:34.830 11:18:18 event -- common/autotest_common.sh@10 -- # set +x 00:07:34.830 ************************************ 00:07:34.830 START TEST event_reactor_perf 00:07:34.830 ************************************ 00:07:34.830 11:18:18 event.event_reactor_perf -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/reactor_perf/reactor_perf -t 1 00:07:34.830 [2024-07-15 11:18:18.190943] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:07:34.830 [2024-07-15 11:18:18.191014] [ DPDK EAL parameters: reactor_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid830539 ] 00:07:34.830 [2024-07-15 11:18:18.324060] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:35.087 [2024-07-15 11:18:18.427708] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:07:36.018 test_start 00:07:36.018 test_end 00:07:36.018 Performance: 328554 events per second 00:07:36.018 00:07:36.018 real 0m1.361s 00:07:36.018 user 0m1.220s 00:07:36.018 sys 0m0.135s 00:07:36.018 11:18:19 event.event_reactor_perf -- common/autotest_common.sh@1124 -- # xtrace_disable 00:07:36.018 11:18:19 event.event_reactor_perf -- common/autotest_common.sh@10 -- # set +x 00:07:36.018 ************************************ 00:07:36.018 END TEST event_reactor_perf 00:07:36.018 ************************************ 00:07:36.018 11:18:19 event -- common/autotest_common.sh@1142 -- # return 0 00:07:36.018 11:18:19 event -- event/event.sh@49 -- # uname -s 00:07:36.018 11:18:19 event -- event/event.sh@49 -- # '[' Linux = Linux ']' 00:07:36.018 11:18:19 event -- event/event.sh@50 -- # run_test event_scheduler /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/scheduler/scheduler.sh 00:07:36.018 11:18:19 event -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:07:36.018 11:18:19 event -- common/autotest_common.sh@1105 -- # xtrace_disable 00:07:36.018 11:18:19 event -- common/autotest_common.sh@10 -- # set +x 00:07:36.276 ************************************ 00:07:36.276 START TEST event_scheduler 00:07:36.276 ************************************ 00:07:36.276 11:18:19 event.event_scheduler -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/scheduler/scheduler.sh 00:07:36.276 * Looking for test storage... 00:07:36.276 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/scheduler 00:07:36.276 11:18:19 event.event_scheduler -- scheduler/scheduler.sh@29 -- # rpc=rpc_cmd 00:07:36.276 11:18:19 event.event_scheduler -- scheduler/scheduler.sh@35 -- # scheduler_pid=830767 00:07:36.276 11:18:19 event.event_scheduler -- scheduler/scheduler.sh@36 -- # trap 'killprocess $scheduler_pid; exit 1' SIGINT SIGTERM EXIT 00:07:36.276 11:18:19 event.event_scheduler -- scheduler/scheduler.sh@37 -- # waitforlisten 830767 00:07:36.276 11:18:19 event.event_scheduler -- common/autotest_common.sh@829 -- # '[' -z 830767 ']' 00:07:36.276 11:18:19 event.event_scheduler -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:36.276 11:18:19 event.event_scheduler -- common/autotest_common.sh@834 -- # local max_retries=100 00:07:36.276 11:18:19 event.event_scheduler -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:36.276 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:36.276 11:18:19 event.event_scheduler -- scheduler/scheduler.sh@34 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/scheduler/scheduler -m 0xF -p 0x2 --wait-for-rpc -f 00:07:36.276 11:18:19 event.event_scheduler -- common/autotest_common.sh@838 -- # xtrace_disable 00:07:36.276 11:18:19 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:07:36.276 [2024-07-15 11:18:19.831424] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:07:36.276 [2024-07-15 11:18:19.831563] [ DPDK EAL parameters: scheduler --no-shconf -c 0xF --main-lcore=2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid830767 ] 00:07:36.534 [2024-07-15 11:18:20.003340] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:07:36.534 [2024-07-15 11:18:20.103949] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:07:36.534 [2024-07-15 11:18:20.103986] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:07:36.534 [2024-07-15 11:18:20.104008] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:07:36.534 [2024-07-15 11:18:20.104011] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:07:37.466 11:18:20 event.event_scheduler -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:07:37.466 11:18:20 event.event_scheduler -- common/autotest_common.sh@862 -- # return 0 00:07:37.466 11:18:20 event.event_scheduler -- scheduler/scheduler.sh@39 -- # rpc_cmd framework_set_scheduler dynamic 00:07:37.466 11:18:20 event.event_scheduler -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:37.466 11:18:20 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:07:37.466 [2024-07-15 11:18:20.738459] dpdk_governor.c: 173:_init: *ERROR*: App core mask contains some but not all of a set of SMT siblings 00:07:37.466 [2024-07-15 11:18:20.738482] scheduler_dynamic.c: 270:init: *NOTICE*: Unable to initialize dpdk governor 00:07:37.466 [2024-07-15 11:18:20.738494] scheduler_dynamic.c: 416:set_opts: *NOTICE*: Setting scheduler load limit to 20 00:07:37.466 [2024-07-15 11:18:20.738502] scheduler_dynamic.c: 418:set_opts: *NOTICE*: Setting scheduler core limit to 80 00:07:37.466 [2024-07-15 11:18:20.738514] scheduler_dynamic.c: 420:set_opts: *NOTICE*: Setting scheduler core busy to 95 00:07:37.466 11:18:20 event.event_scheduler -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:37.466 11:18:20 event.event_scheduler -- scheduler/scheduler.sh@40 -- # rpc_cmd framework_start_init 00:07:37.466 11:18:20 event.event_scheduler -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:37.466 11:18:20 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:07:37.466 [2024-07-15 11:18:20.825319] scheduler.c: 382:test_start: *NOTICE*: Scheduler test application started. 00:07:37.466 11:18:20 event.event_scheduler -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:37.466 11:18:20 event.event_scheduler -- scheduler/scheduler.sh@43 -- # run_test scheduler_create_thread scheduler_create_thread 00:07:37.466 11:18:20 event.event_scheduler -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:07:37.466 11:18:20 event.event_scheduler -- common/autotest_common.sh@1105 -- # xtrace_disable 00:07:37.466 11:18:20 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:07:37.466 ************************************ 00:07:37.466 START TEST scheduler_create_thread 00:07:37.466 ************************************ 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1123 -- # scheduler_create_thread 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@12 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x1 -a 100 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:37.466 2 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@13 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x2 -a 100 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:37.466 3 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@14 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x4 -a 100 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:37.466 4 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@15 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x8 -a 100 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:37.466 5 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@16 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x1 -a 0 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:37.466 6 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@17 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x2 -a 0 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:37.466 7 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@18 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x4 -a 0 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:37.466 8 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@19 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x8 -a 0 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:37.466 9 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@21 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n one_third_active -a 30 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:37.466 10 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n half_active -a 0 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # thread_id=11 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@23 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_set_active 11 50 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:37.466 11:18:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:38.029 11:18:21 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:38.029 11:18:21 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n deleted -a 100 00:07:38.029 11:18:21 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:38.029 11:18:21 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:39.450 11:18:22 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:39.450 11:18:22 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # thread_id=12 00:07:39.450 11:18:22 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@26 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_delete 12 00:07:39.450 11:18:22 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:39.450 11:18:22 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:40.380 11:18:23 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:40.380 00:07:40.380 real 0m3.103s 00:07:40.380 user 0m0.027s 00:07:40.380 sys 0m0.004s 00:07:40.380 11:18:23 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1124 -- # xtrace_disable 00:07:40.380 11:18:23 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:40.380 ************************************ 00:07:40.380 END TEST scheduler_create_thread 00:07:40.380 ************************************ 00:07:40.637 11:18:24 event.event_scheduler -- common/autotest_common.sh@1142 -- # return 0 00:07:40.637 11:18:24 event.event_scheduler -- scheduler/scheduler.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:07:40.637 11:18:24 event.event_scheduler -- scheduler/scheduler.sh@46 -- # killprocess 830767 00:07:40.637 11:18:24 event.event_scheduler -- common/autotest_common.sh@948 -- # '[' -z 830767 ']' 00:07:40.637 11:18:24 event.event_scheduler -- common/autotest_common.sh@952 -- # kill -0 830767 00:07:40.637 11:18:24 event.event_scheduler -- common/autotest_common.sh@953 -- # uname 00:07:40.637 11:18:24 event.event_scheduler -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:07:40.637 11:18:24 event.event_scheduler -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 830767 00:07:40.637 11:18:24 event.event_scheduler -- common/autotest_common.sh@954 -- # process_name=reactor_2 00:07:40.637 11:18:24 event.event_scheduler -- common/autotest_common.sh@958 -- # '[' reactor_2 = sudo ']' 00:07:40.637 11:18:24 event.event_scheduler -- common/autotest_common.sh@966 -- # echo 'killing process with pid 830767' 00:07:40.637 killing process with pid 830767 00:07:40.637 11:18:24 event.event_scheduler -- common/autotest_common.sh@967 -- # kill 830767 00:07:40.637 11:18:24 event.event_scheduler -- common/autotest_common.sh@972 -- # wait 830767 00:07:40.894 [2024-07-15 11:18:24.352835] scheduler.c: 360:test_shutdown: *NOTICE*: Scheduler test application stopped. 00:07:41.152 00:07:41.152 real 0m4.977s 00:07:41.152 user 0m9.402s 00:07:41.152 sys 0m0.536s 00:07:41.152 11:18:24 event.event_scheduler -- common/autotest_common.sh@1124 -- # xtrace_disable 00:07:41.152 11:18:24 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:07:41.152 ************************************ 00:07:41.152 END TEST event_scheduler 00:07:41.152 ************************************ 00:07:41.152 11:18:24 event -- common/autotest_common.sh@1142 -- # return 0 00:07:41.152 11:18:24 event -- event/event.sh@51 -- # modprobe -n nbd 00:07:41.152 11:18:24 event -- event/event.sh@52 -- # run_test app_repeat app_repeat_test 00:07:41.152 11:18:24 event -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:07:41.152 11:18:24 event -- common/autotest_common.sh@1105 -- # xtrace_disable 00:07:41.152 11:18:24 event -- common/autotest_common.sh@10 -- # set +x 00:07:41.152 ************************************ 00:07:41.152 START TEST app_repeat 00:07:41.152 ************************************ 00:07:41.152 11:18:24 event.app_repeat -- common/autotest_common.sh@1123 -- # app_repeat_test 00:07:41.152 11:18:24 event.app_repeat -- event/event.sh@12 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:41.152 11:18:24 event.app_repeat -- event/event.sh@13 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:41.152 11:18:24 event.app_repeat -- event/event.sh@13 -- # local nbd_list 00:07:41.152 11:18:24 event.app_repeat -- event/event.sh@14 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:41.152 11:18:24 event.app_repeat -- event/event.sh@14 -- # local bdev_list 00:07:41.152 11:18:24 event.app_repeat -- event/event.sh@15 -- # local repeat_times=4 00:07:41.152 11:18:24 event.app_repeat -- event/event.sh@17 -- # modprobe nbd 00:07:41.152 11:18:24 event.app_repeat -- event/event.sh@19 -- # repeat_pid=831513 00:07:41.152 11:18:24 event.app_repeat -- event/event.sh@20 -- # trap 'killprocess $repeat_pid; exit 1' SIGINT SIGTERM EXIT 00:07:41.152 11:18:24 event.app_repeat -- event/event.sh@21 -- # echo 'Process app_repeat pid: 831513' 00:07:41.152 Process app_repeat pid: 831513 00:07:41.152 11:18:24 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:07:41.152 11:18:24 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 0' 00:07:41.152 spdk_app_start Round 0 00:07:41.152 11:18:24 event.app_repeat -- event/event.sh@25 -- # waitforlisten 831513 /var/tmp/spdk-nbd.sock 00:07:41.152 11:18:24 event.app_repeat -- common/autotest_common.sh@829 -- # '[' -z 831513 ']' 00:07:41.152 11:18:24 event.app_repeat -- event/event.sh@18 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/app_repeat/app_repeat -r /var/tmp/spdk-nbd.sock -m 0x3 -t 4 00:07:41.152 11:18:24 event.app_repeat -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:07:41.152 11:18:24 event.app_repeat -- common/autotest_common.sh@834 -- # local max_retries=100 00:07:41.152 11:18:24 event.app_repeat -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:07:41.152 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:07:41.152 11:18:24 event.app_repeat -- common/autotest_common.sh@838 -- # xtrace_disable 00:07:41.152 11:18:24 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:07:41.152 [2024-07-15 11:18:24.722719] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:07:41.152 [2024-07-15 11:18:24.722779] [ DPDK EAL parameters: app_repeat --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid831513 ] 00:07:41.410 [2024-07-15 11:18:24.851557] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:07:41.410 [2024-07-15 11:18:24.960552] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:07:41.410 [2024-07-15 11:18:24.960558] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:07:42.780 11:18:25 event.app_repeat -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:07:42.780 11:18:25 event.app_repeat -- common/autotest_common.sh@862 -- # return 0 00:07:42.780 11:18:25 event.app_repeat -- event/event.sh@27 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:07:42.780 Malloc0 00:07:42.780 11:18:26 event.app_repeat -- event/event.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:07:43.038 Malloc1 00:07:43.038 11:18:26 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:07:43.038 11:18:26 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:43.038 11:18:26 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:43.038 11:18:26 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:07:43.038 11:18:26 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:43.038 11:18:26 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:07:43.038 11:18:26 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:07:43.038 11:18:26 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:43.038 11:18:26 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:43.038 11:18:26 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:07:43.038 11:18:26 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:43.038 11:18:26 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:07:43.038 11:18:26 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:07:43.038 11:18:26 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:07:43.038 11:18:26 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:43.038 11:18:26 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:07:43.296 /dev/nbd0 00:07:43.296 11:18:26 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:07:43.296 11:18:26 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:07:43.296 11:18:26 event.app_repeat -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:07:43.296 11:18:26 event.app_repeat -- common/autotest_common.sh@867 -- # local i 00:07:43.296 11:18:26 event.app_repeat -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:07:43.296 11:18:26 event.app_repeat -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:07:43.296 11:18:26 event.app_repeat -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:07:43.296 11:18:26 event.app_repeat -- common/autotest_common.sh@871 -- # break 00:07:43.296 11:18:26 event.app_repeat -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:07:43.296 11:18:26 event.app_repeat -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:07:43.296 11:18:26 event.app_repeat -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:07:43.296 1+0 records in 00:07:43.296 1+0 records out 00:07:43.296 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000284602 s, 14.4 MB/s 00:07:43.296 11:18:26 event.app_repeat -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdtest 00:07:43.296 11:18:26 event.app_repeat -- common/autotest_common.sh@884 -- # size=4096 00:07:43.296 11:18:26 event.app_repeat -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdtest 00:07:43.296 11:18:26 event.app_repeat -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:07:43.296 11:18:26 event.app_repeat -- common/autotest_common.sh@887 -- # return 0 00:07:43.296 11:18:26 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:43.296 11:18:26 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:43.296 11:18:26 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:07:43.554 /dev/nbd1 00:07:43.554 11:18:27 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:07:43.554 11:18:27 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:07:43.554 11:18:27 event.app_repeat -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:07:43.554 11:18:27 event.app_repeat -- common/autotest_common.sh@867 -- # local i 00:07:43.554 11:18:27 event.app_repeat -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:07:43.554 11:18:27 event.app_repeat -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:07:43.554 11:18:27 event.app_repeat -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:07:43.554 11:18:27 event.app_repeat -- common/autotest_common.sh@871 -- # break 00:07:43.554 11:18:27 event.app_repeat -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:07:43.554 11:18:27 event.app_repeat -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:07:43.554 11:18:27 event.app_repeat -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:07:43.554 1+0 records in 00:07:43.554 1+0 records out 00:07:43.554 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000266228 s, 15.4 MB/s 00:07:43.554 11:18:27 event.app_repeat -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdtest 00:07:43.554 11:18:27 event.app_repeat -- common/autotest_common.sh@884 -- # size=4096 00:07:43.554 11:18:27 event.app_repeat -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdtest 00:07:43.554 11:18:27 event.app_repeat -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:07:43.554 11:18:27 event.app_repeat -- common/autotest_common.sh@887 -- # return 0 00:07:43.554 11:18:27 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:43.554 11:18:27 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:43.554 11:18:27 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:07:43.554 11:18:27 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:43.554 11:18:27 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:07:43.812 { 00:07:43.812 "nbd_device": "/dev/nbd0", 00:07:43.812 "bdev_name": "Malloc0" 00:07:43.812 }, 00:07:43.812 { 00:07:43.812 "nbd_device": "/dev/nbd1", 00:07:43.812 "bdev_name": "Malloc1" 00:07:43.812 } 00:07:43.812 ]' 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:07:43.812 { 00:07:43.812 "nbd_device": "/dev/nbd0", 00:07:43.812 "bdev_name": "Malloc0" 00:07:43.812 }, 00:07:43.812 { 00:07:43.812 "nbd_device": "/dev/nbd1", 00:07:43.812 "bdev_name": "Malloc1" 00:07:43.812 } 00:07:43.812 ]' 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:07:43.812 /dev/nbd1' 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:07:43.812 /dev/nbd1' 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:07:43.812 256+0 records in 00:07:43.812 256+0 records out 00:07:43.812 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0106177 s, 98.8 MB/s 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:07:43.812 256+0 records in 00:07:43.812 256+0 records out 00:07:43.812 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0208175 s, 50.4 MB/s 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:07:43.812 11:18:27 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:07:44.070 256+0 records in 00:07:44.070 256+0 records out 00:07:44.070 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0202653 s, 51.7 MB/s 00:07:44.070 11:18:27 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:07:44.070 11:18:27 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:44.070 11:18:27 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:07:44.070 11:18:27 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:07:44.070 11:18:27 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest 00:07:44.070 11:18:27 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:07:44.070 11:18:27 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:07:44.070 11:18:27 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:07:44.070 11:18:27 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:07:44.070 11:18:27 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:07:44.070 11:18:27 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:07:44.070 11:18:27 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest 00:07:44.070 11:18:27 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:07:44.070 11:18:27 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:44.070 11:18:27 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:44.070 11:18:27 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:07:44.070 11:18:27 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:07:44.070 11:18:27 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:44.070 11:18:27 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:07:44.328 11:18:27 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:07:44.328 11:18:27 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:07:44.328 11:18:27 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:07:44.328 11:18:27 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:44.328 11:18:27 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:44.328 11:18:27 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:07:44.328 11:18:27 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:07:44.328 11:18:27 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:07:44.328 11:18:27 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:44.328 11:18:27 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:07:44.586 11:18:28 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:07:44.586 11:18:28 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:07:44.586 11:18:28 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:07:44.586 11:18:28 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:44.586 11:18:28 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:44.586 11:18:28 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:07:44.586 11:18:28 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:07:44.586 11:18:28 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:07:44.586 11:18:28 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:07:44.586 11:18:28 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:44.586 11:18:28 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:07:44.844 11:18:28 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:07:44.844 11:18:28 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:07:44.844 11:18:28 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:44.844 11:18:28 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:07:44.844 11:18:28 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:07:44.844 11:18:28 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:44.844 11:18:28 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:07:44.844 11:18:28 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:07:44.844 11:18:28 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:07:44.844 11:18:28 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:07:44.844 11:18:28 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:07:44.844 11:18:28 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:07:44.844 11:18:28 event.app_repeat -- event/event.sh@34 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:07:45.102 11:18:28 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:07:45.360 [2024-07-15 11:18:28.763583] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:07:45.360 [2024-07-15 11:18:28.867489] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:07:45.360 [2024-07-15 11:18:28.867494] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:07:45.360 [2024-07-15 11:18:28.919704] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:07:45.360 [2024-07-15 11:18:28.919751] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:07:48.636 11:18:31 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:07:48.636 11:18:31 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 1' 00:07:48.636 spdk_app_start Round 1 00:07:48.636 11:18:31 event.app_repeat -- event/event.sh@25 -- # waitforlisten 831513 /var/tmp/spdk-nbd.sock 00:07:48.636 11:18:31 event.app_repeat -- common/autotest_common.sh@829 -- # '[' -z 831513 ']' 00:07:48.636 11:18:31 event.app_repeat -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:07:48.636 11:18:31 event.app_repeat -- common/autotest_common.sh@834 -- # local max_retries=100 00:07:48.636 11:18:31 event.app_repeat -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:07:48.636 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:07:48.636 11:18:31 event.app_repeat -- common/autotest_common.sh@838 -- # xtrace_disable 00:07:48.636 11:18:31 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:07:48.636 11:18:31 event.app_repeat -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:07:48.636 11:18:31 event.app_repeat -- common/autotest_common.sh@862 -- # return 0 00:07:48.636 11:18:31 event.app_repeat -- event/event.sh@27 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:07:48.636 Malloc0 00:07:48.636 11:18:32 event.app_repeat -- event/event.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:07:48.893 Malloc1 00:07:48.893 11:18:32 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:07:48.893 11:18:32 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:48.893 11:18:32 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:48.893 11:18:32 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:07:48.893 11:18:32 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:48.893 11:18:32 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:07:48.893 11:18:32 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:07:48.893 11:18:32 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:48.893 11:18:32 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:48.893 11:18:32 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:07:48.893 11:18:32 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:48.893 11:18:32 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:07:48.893 11:18:32 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:07:48.893 11:18:32 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:07:48.893 11:18:32 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:48.893 11:18:32 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:07:49.151 /dev/nbd0 00:07:49.151 11:18:32 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:07:49.151 11:18:32 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:07:49.151 11:18:32 event.app_repeat -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:07:49.151 11:18:32 event.app_repeat -- common/autotest_common.sh@867 -- # local i 00:07:49.151 11:18:32 event.app_repeat -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:07:49.151 11:18:32 event.app_repeat -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:07:49.151 11:18:32 event.app_repeat -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:07:49.151 11:18:32 event.app_repeat -- common/autotest_common.sh@871 -- # break 00:07:49.151 11:18:32 event.app_repeat -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:07:49.151 11:18:32 event.app_repeat -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:07:49.151 11:18:32 event.app_repeat -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:07:49.151 1+0 records in 00:07:49.151 1+0 records out 00:07:49.151 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000139956 s, 29.3 MB/s 00:07:49.151 11:18:32 event.app_repeat -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdtest 00:07:49.151 11:18:32 event.app_repeat -- common/autotest_common.sh@884 -- # size=4096 00:07:49.151 11:18:32 event.app_repeat -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdtest 00:07:49.151 11:18:32 event.app_repeat -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:07:49.151 11:18:32 event.app_repeat -- common/autotest_common.sh@887 -- # return 0 00:07:49.151 11:18:32 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:49.151 11:18:32 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:49.151 11:18:32 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:07:49.409 /dev/nbd1 00:07:49.409 11:18:32 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:07:49.409 11:18:32 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:07:49.409 11:18:32 event.app_repeat -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:07:49.409 11:18:32 event.app_repeat -- common/autotest_common.sh@867 -- # local i 00:07:49.409 11:18:32 event.app_repeat -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:07:49.409 11:18:32 event.app_repeat -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:07:49.409 11:18:32 event.app_repeat -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:07:49.409 11:18:32 event.app_repeat -- common/autotest_common.sh@871 -- # break 00:07:49.409 11:18:32 event.app_repeat -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:07:49.409 11:18:32 event.app_repeat -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:07:49.409 11:18:32 event.app_repeat -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:07:49.409 1+0 records in 00:07:49.409 1+0 records out 00:07:49.409 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000228634 s, 17.9 MB/s 00:07:49.409 11:18:32 event.app_repeat -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdtest 00:07:49.409 11:18:32 event.app_repeat -- common/autotest_common.sh@884 -- # size=4096 00:07:49.409 11:18:32 event.app_repeat -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdtest 00:07:49.409 11:18:32 event.app_repeat -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:07:49.409 11:18:32 event.app_repeat -- common/autotest_common.sh@887 -- # return 0 00:07:49.409 11:18:32 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:49.409 11:18:32 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:49.409 11:18:32 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:07:49.409 11:18:32 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:49.409 11:18:32 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:07:49.666 { 00:07:49.666 "nbd_device": "/dev/nbd0", 00:07:49.666 "bdev_name": "Malloc0" 00:07:49.666 }, 00:07:49.666 { 00:07:49.666 "nbd_device": "/dev/nbd1", 00:07:49.666 "bdev_name": "Malloc1" 00:07:49.666 } 00:07:49.666 ]' 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:07:49.666 { 00:07:49.666 "nbd_device": "/dev/nbd0", 00:07:49.666 "bdev_name": "Malloc0" 00:07:49.666 }, 00:07:49.666 { 00:07:49.666 "nbd_device": "/dev/nbd1", 00:07:49.666 "bdev_name": "Malloc1" 00:07:49.666 } 00:07:49.666 ]' 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:07:49.666 /dev/nbd1' 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:07:49.666 /dev/nbd1' 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:07:49.666 256+0 records in 00:07:49.666 256+0 records out 00:07:49.666 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0113639 s, 92.3 MB/s 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:07:49.666 256+0 records in 00:07:49.666 256+0 records out 00:07:49.666 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0180871 s, 58.0 MB/s 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:07:49.666 256+0 records in 00:07:49.666 256+0 records out 00:07:49.666 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0214183 s, 49.0 MB/s 00:07:49.666 11:18:33 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:07:49.667 11:18:33 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:49.667 11:18:33 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:07:49.667 11:18:33 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:07:49.667 11:18:33 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest 00:07:49.667 11:18:33 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:07:49.667 11:18:33 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:07:49.667 11:18:33 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:07:49.667 11:18:33 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:07:49.667 11:18:33 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:07:49.667 11:18:33 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:07:49.667 11:18:33 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest 00:07:49.667 11:18:33 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:07:49.667 11:18:33 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:49.667 11:18:33 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:49.667 11:18:33 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:07:49.667 11:18:33 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:07:49.667 11:18:33 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:49.667 11:18:33 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:07:49.923 11:18:33 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:07:49.923 11:18:33 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:07:49.923 11:18:33 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:07:49.923 11:18:33 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:49.923 11:18:33 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:49.923 11:18:33 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:07:49.923 11:18:33 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:07:49.923 11:18:33 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:07:49.923 11:18:33 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:49.923 11:18:33 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:07:50.179 11:18:33 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:07:50.179 11:18:33 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:07:50.179 11:18:33 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:07:50.179 11:18:33 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:50.179 11:18:33 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:50.179 11:18:33 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:07:50.179 11:18:33 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:07:50.179 11:18:33 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:07:50.436 11:18:33 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:07:50.436 11:18:33 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:50.436 11:18:33 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:07:50.436 11:18:34 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:07:50.436 11:18:34 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:07:50.436 11:18:34 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:50.693 11:18:34 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:07:50.693 11:18:34 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:07:50.693 11:18:34 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:50.693 11:18:34 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:07:50.693 11:18:34 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:07:50.693 11:18:34 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:07:50.693 11:18:34 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:07:50.693 11:18:34 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:07:50.693 11:18:34 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:07:50.693 11:18:34 event.app_repeat -- event/event.sh@34 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:07:50.951 11:18:34 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:07:51.208 [2024-07-15 11:18:34.581590] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:07:51.208 [2024-07-15 11:18:34.680564] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:07:51.208 [2024-07-15 11:18:34.680569] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:07:51.208 [2024-07-15 11:18:34.734059] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:07:51.208 [2024-07-15 11:18:34.734112] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:07:54.484 11:18:37 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:07:54.484 11:18:37 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 2' 00:07:54.484 spdk_app_start Round 2 00:07:54.484 11:18:37 event.app_repeat -- event/event.sh@25 -- # waitforlisten 831513 /var/tmp/spdk-nbd.sock 00:07:54.484 11:18:37 event.app_repeat -- common/autotest_common.sh@829 -- # '[' -z 831513 ']' 00:07:54.484 11:18:37 event.app_repeat -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:07:54.484 11:18:37 event.app_repeat -- common/autotest_common.sh@834 -- # local max_retries=100 00:07:54.484 11:18:37 event.app_repeat -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:07:54.484 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:07:54.484 11:18:37 event.app_repeat -- common/autotest_common.sh@838 -- # xtrace_disable 00:07:54.484 11:18:37 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:07:54.484 11:18:37 event.app_repeat -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:07:54.484 11:18:37 event.app_repeat -- common/autotest_common.sh@862 -- # return 0 00:07:54.484 11:18:37 event.app_repeat -- event/event.sh@27 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:07:54.484 Malloc0 00:07:54.484 11:18:37 event.app_repeat -- event/event.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:07:54.484 Malloc1 00:07:54.484 11:18:38 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:07:54.484 11:18:38 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:54.484 11:18:38 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:54.484 11:18:38 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:07:54.484 11:18:38 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:54.484 11:18:38 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:07:54.484 11:18:38 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:07:54.484 11:18:38 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:54.484 11:18:38 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:54.484 11:18:38 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:07:54.484 11:18:38 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:54.484 11:18:38 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:07:54.484 11:18:38 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:07:54.484 11:18:38 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:07:54.484 11:18:38 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:54.484 11:18:38 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:07:54.742 /dev/nbd0 00:07:54.742 11:18:38 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:07:54.742 11:18:38 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:07:54.742 11:18:38 event.app_repeat -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:07:54.742 11:18:38 event.app_repeat -- common/autotest_common.sh@867 -- # local i 00:07:54.742 11:18:38 event.app_repeat -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:07:54.742 11:18:38 event.app_repeat -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:07:54.742 11:18:38 event.app_repeat -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:07:54.742 11:18:38 event.app_repeat -- common/autotest_common.sh@871 -- # break 00:07:54.742 11:18:38 event.app_repeat -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:07:54.742 11:18:38 event.app_repeat -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:07:54.742 11:18:38 event.app_repeat -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:07:54.742 1+0 records in 00:07:54.742 1+0 records out 00:07:54.742 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000252647 s, 16.2 MB/s 00:07:54.742 11:18:38 event.app_repeat -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdtest 00:07:54.742 11:18:38 event.app_repeat -- common/autotest_common.sh@884 -- # size=4096 00:07:54.742 11:18:38 event.app_repeat -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdtest 00:07:54.742 11:18:38 event.app_repeat -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:07:54.742 11:18:38 event.app_repeat -- common/autotest_common.sh@887 -- # return 0 00:07:54.742 11:18:38 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:54.742 11:18:38 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:54.742 11:18:38 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:07:55.000 /dev/nbd1 00:07:55.000 11:18:38 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:07:55.000 11:18:38 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:07:55.000 11:18:38 event.app_repeat -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:07:55.000 11:18:38 event.app_repeat -- common/autotest_common.sh@867 -- # local i 00:07:55.000 11:18:38 event.app_repeat -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:07:55.000 11:18:38 event.app_repeat -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:07:55.000 11:18:38 event.app_repeat -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:07:55.000 11:18:38 event.app_repeat -- common/autotest_common.sh@871 -- # break 00:07:55.000 11:18:38 event.app_repeat -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:07:55.000 11:18:38 event.app_repeat -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:07:55.000 11:18:38 event.app_repeat -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:07:55.000 1+0 records in 00:07:55.000 1+0 records out 00:07:55.000 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000257122 s, 15.9 MB/s 00:07:55.000 11:18:38 event.app_repeat -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdtest 00:07:55.000 11:18:38 event.app_repeat -- common/autotest_common.sh@884 -- # size=4096 00:07:55.000 11:18:38 event.app_repeat -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdtest 00:07:55.000 11:18:38 event.app_repeat -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:07:55.000 11:18:38 event.app_repeat -- common/autotest_common.sh@887 -- # return 0 00:07:55.000 11:18:38 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:55.000 11:18:38 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:55.000 11:18:38 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:07:55.000 11:18:38 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:55.000 11:18:38 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:07:55.258 11:18:38 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:07:55.258 { 00:07:55.258 "nbd_device": "/dev/nbd0", 00:07:55.258 "bdev_name": "Malloc0" 00:07:55.258 }, 00:07:55.258 { 00:07:55.258 "nbd_device": "/dev/nbd1", 00:07:55.258 "bdev_name": "Malloc1" 00:07:55.258 } 00:07:55.258 ]' 00:07:55.258 11:18:38 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:07:55.258 { 00:07:55.258 "nbd_device": "/dev/nbd0", 00:07:55.258 "bdev_name": "Malloc0" 00:07:55.258 }, 00:07:55.258 { 00:07:55.258 "nbd_device": "/dev/nbd1", 00:07:55.258 "bdev_name": "Malloc1" 00:07:55.258 } 00:07:55.258 ]' 00:07:55.258 11:18:38 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:55.258 11:18:38 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:07:55.258 /dev/nbd1' 00:07:55.258 11:18:38 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:07:55.258 /dev/nbd1' 00:07:55.258 11:18:38 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:55.258 11:18:38 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:07:55.258 11:18:38 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:07:55.258 11:18:38 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:07:55.258 11:18:38 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:07:55.258 11:18:38 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:07:55.258 11:18:38 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:55.258 11:18:38 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:07:55.258 11:18:38 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:07:55.258 11:18:38 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest 00:07:55.258 11:18:38 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:07:55.258 11:18:38 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:07:55.258 256+0 records in 00:07:55.258 256+0 records out 00:07:55.258 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0114948 s, 91.2 MB/s 00:07:55.258 11:18:38 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:07:55.258 11:18:38 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:07:55.515 256+0 records in 00:07:55.515 256+0 records out 00:07:55.515 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0299364 s, 35.0 MB/s 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:07:55.515 256+0 records in 00:07:55.515 256+0 records out 00:07:55.515 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0194517 s, 53.9 MB/s 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/crypto-phy-autotest/spdk/test/event/nbdrandtest 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:55.515 11:18:38 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:07:55.772 11:18:39 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:07:55.772 11:18:39 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:07:55.772 11:18:39 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:07:55.772 11:18:39 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:55.772 11:18:39 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:55.772 11:18:39 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:07:55.772 11:18:39 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:07:55.772 11:18:39 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:07:55.772 11:18:39 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:55.772 11:18:39 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:07:56.029 11:18:39 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:07:56.029 11:18:39 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:07:56.029 11:18:39 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:07:56.029 11:18:39 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:56.029 11:18:39 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:56.029 11:18:39 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:07:56.029 11:18:39 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:07:56.029 11:18:39 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:07:56.029 11:18:39 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:07:56.029 11:18:39 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:56.029 11:18:39 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:07:56.285 11:18:39 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:07:56.285 11:18:39 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:07:56.285 11:18:39 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:56.285 11:18:39 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:07:56.285 11:18:39 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:07:56.285 11:18:39 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:56.285 11:18:39 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:07:56.285 11:18:39 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:07:56.285 11:18:39 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:07:56.285 11:18:39 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:07:56.285 11:18:39 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:07:56.285 11:18:39 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:07:56.285 11:18:39 event.app_repeat -- event/event.sh@34 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:07:56.580 11:18:40 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:07:56.842 [2024-07-15 11:18:40.266069] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:07:56.842 [2024-07-15 11:18:40.364695] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:07:56.842 [2024-07-15 11:18:40.364699] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:07:56.842 [2024-07-15 11:18:40.417035] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:07:56.842 [2024-07-15 11:18:40.417087] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:08:00.112 11:18:43 event.app_repeat -- event/event.sh@38 -- # waitforlisten 831513 /var/tmp/spdk-nbd.sock 00:08:00.112 11:18:43 event.app_repeat -- common/autotest_common.sh@829 -- # '[' -z 831513 ']' 00:08:00.112 11:18:43 event.app_repeat -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:08:00.112 11:18:43 event.app_repeat -- common/autotest_common.sh@834 -- # local max_retries=100 00:08:00.112 11:18:43 event.app_repeat -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:08:00.112 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:08:00.112 11:18:43 event.app_repeat -- common/autotest_common.sh@838 -- # xtrace_disable 00:08:00.112 11:18:43 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:08:00.112 11:18:43 event.app_repeat -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:08:00.112 11:18:43 event.app_repeat -- common/autotest_common.sh@862 -- # return 0 00:08:00.112 11:18:43 event.app_repeat -- event/event.sh@39 -- # killprocess 831513 00:08:00.112 11:18:43 event.app_repeat -- common/autotest_common.sh@948 -- # '[' -z 831513 ']' 00:08:00.112 11:18:43 event.app_repeat -- common/autotest_common.sh@952 -- # kill -0 831513 00:08:00.112 11:18:43 event.app_repeat -- common/autotest_common.sh@953 -- # uname 00:08:00.112 11:18:43 event.app_repeat -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:08:00.112 11:18:43 event.app_repeat -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 831513 00:08:00.112 11:18:43 event.app_repeat -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:08:00.112 11:18:43 event.app_repeat -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:08:00.112 11:18:43 event.app_repeat -- common/autotest_common.sh@966 -- # echo 'killing process with pid 831513' 00:08:00.112 killing process with pid 831513 00:08:00.112 11:18:43 event.app_repeat -- common/autotest_common.sh@967 -- # kill 831513 00:08:00.112 11:18:43 event.app_repeat -- common/autotest_common.sh@972 -- # wait 831513 00:08:00.112 spdk_app_start is called in Round 0. 00:08:00.112 Shutdown signal received, stop current app iteration 00:08:00.112 Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 reinitialization... 00:08:00.112 spdk_app_start is called in Round 1. 00:08:00.112 Shutdown signal received, stop current app iteration 00:08:00.112 Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 reinitialization... 00:08:00.112 spdk_app_start is called in Round 2. 00:08:00.112 Shutdown signal received, stop current app iteration 00:08:00.112 Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 reinitialization... 00:08:00.112 spdk_app_start is called in Round 3. 00:08:00.112 Shutdown signal received, stop current app iteration 00:08:00.112 11:18:43 event.app_repeat -- event/event.sh@40 -- # trap - SIGINT SIGTERM EXIT 00:08:00.112 11:18:43 event.app_repeat -- event/event.sh@42 -- # return 0 00:08:00.112 00:08:00.112 real 0m18.841s 00:08:00.112 user 0m40.902s 00:08:00.112 sys 0m3.867s 00:08:00.112 11:18:43 event.app_repeat -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:00.112 11:18:43 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:08:00.112 ************************************ 00:08:00.112 END TEST app_repeat 00:08:00.112 ************************************ 00:08:00.112 11:18:43 event -- common/autotest_common.sh@1142 -- # return 0 00:08:00.112 11:18:43 event -- event/event.sh@54 -- # (( SPDK_TEST_CRYPTO == 0 )) 00:08:00.112 00:08:00.112 real 0m28.443s 00:08:00.112 user 0m57.117s 00:08:00.112 sys 0m5.238s 00:08:00.112 11:18:43 event -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:00.112 11:18:43 event -- common/autotest_common.sh@10 -- # set +x 00:08:00.112 ************************************ 00:08:00.112 END TEST event 00:08:00.112 ************************************ 00:08:00.112 11:18:43 -- common/autotest_common.sh@1142 -- # return 0 00:08:00.112 11:18:43 -- spdk/autotest.sh@182 -- # run_test thread /var/jenkins/workspace/crypto-phy-autotest/spdk/test/thread/thread.sh 00:08:00.112 11:18:43 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:08:00.112 11:18:43 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:00.112 11:18:43 -- common/autotest_common.sh@10 -- # set +x 00:08:00.112 ************************************ 00:08:00.112 START TEST thread 00:08:00.112 ************************************ 00:08:00.112 11:18:43 thread -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/thread/thread.sh 00:08:00.370 * Looking for test storage... 00:08:00.370 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/thread 00:08:00.370 11:18:43 thread -- thread/thread.sh@11 -- # run_test thread_poller_perf /var/jenkins/workspace/crypto-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:08:00.370 11:18:43 thread -- common/autotest_common.sh@1099 -- # '[' 8 -le 1 ']' 00:08:00.370 11:18:43 thread -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:00.370 11:18:43 thread -- common/autotest_common.sh@10 -- # set +x 00:08:00.370 ************************************ 00:08:00.370 START TEST thread_poller_perf 00:08:00.370 ************************************ 00:08:00.370 11:18:43 thread.thread_poller_perf -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:08:00.370 [2024-07-15 11:18:43.821397] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:00.370 [2024-07-15 11:18:43.821467] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid834215 ] 00:08:00.370 [2024-07-15 11:18:43.928891] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:00.628 [2024-07-15 11:18:44.031255] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:00.628 Running 1000 pollers for 1 seconds with 1 microseconds period. 00:08:01.560 ====================================== 00:08:01.560 busy:2312762096 (cyc) 00:08:01.560 total_run_count: 267000 00:08:01.560 tsc_hz: 2300000000 (cyc) 00:08:01.560 ====================================== 00:08:01.560 poller_cost: 8662 (cyc), 3766 (nsec) 00:08:01.560 00:08:01.560 real 0m1.341s 00:08:01.560 user 0m1.208s 00:08:01.560 sys 0m0.123s 00:08:01.560 11:18:45 thread.thread_poller_perf -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:01.560 11:18:45 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:08:01.560 ************************************ 00:08:01.560 END TEST thread_poller_perf 00:08:01.560 ************************************ 00:08:01.818 11:18:45 thread -- common/autotest_common.sh@1142 -- # return 0 00:08:01.818 11:18:45 thread -- thread/thread.sh@12 -- # run_test thread_poller_perf /var/jenkins/workspace/crypto-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:08:01.818 11:18:45 thread -- common/autotest_common.sh@1099 -- # '[' 8 -le 1 ']' 00:08:01.818 11:18:45 thread -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:01.818 11:18:45 thread -- common/autotest_common.sh@10 -- # set +x 00:08:01.818 ************************************ 00:08:01.818 START TEST thread_poller_perf 00:08:01.818 ************************************ 00:08:01.818 11:18:45 thread.thread_poller_perf -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:08:01.818 [2024-07-15 11:18:45.250315] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:01.818 [2024-07-15 11:18:45.250380] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid834414 ] 00:08:01.818 [2024-07-15 11:18:45.382689] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:02.076 [2024-07-15 11:18:45.489733] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:02.076 Running 1000 pollers for 1 seconds with 0 microseconds period. 00:08:03.009 ====================================== 00:08:03.009 busy:2302713046 (cyc) 00:08:03.009 total_run_count: 3475000 00:08:03.009 tsc_hz: 2300000000 (cyc) 00:08:03.009 ====================================== 00:08:03.009 poller_cost: 662 (cyc), 287 (nsec) 00:08:03.009 00:08:03.009 real 0m1.366s 00:08:03.009 user 0m1.205s 00:08:03.009 sys 0m0.155s 00:08:03.009 11:18:46 thread.thread_poller_perf -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:03.009 11:18:46 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:08:03.009 ************************************ 00:08:03.009 END TEST thread_poller_perf 00:08:03.009 ************************************ 00:08:03.267 11:18:46 thread -- common/autotest_common.sh@1142 -- # return 0 00:08:03.267 11:18:46 thread -- thread/thread.sh@17 -- # [[ y != \y ]] 00:08:03.267 00:08:03.267 real 0m2.981s 00:08:03.267 user 0m2.513s 00:08:03.267 sys 0m0.475s 00:08:03.267 11:18:46 thread -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:03.267 11:18:46 thread -- common/autotest_common.sh@10 -- # set +x 00:08:03.267 ************************************ 00:08:03.267 END TEST thread 00:08:03.267 ************************************ 00:08:03.267 11:18:46 -- common/autotest_common.sh@1142 -- # return 0 00:08:03.267 11:18:46 -- spdk/autotest.sh@183 -- # run_test accel /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/accel.sh 00:08:03.267 11:18:46 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:08:03.267 11:18:46 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:03.267 11:18:46 -- common/autotest_common.sh@10 -- # set +x 00:08:03.267 ************************************ 00:08:03.267 START TEST accel 00:08:03.267 ************************************ 00:08:03.267 11:18:46 accel -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/accel.sh 00:08:03.267 * Looking for test storage... 00:08:03.267 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel 00:08:03.267 11:18:46 accel -- accel/accel.sh@81 -- # declare -A expected_opcs 00:08:03.267 11:18:46 accel -- accel/accel.sh@82 -- # get_expected_opcs 00:08:03.267 11:18:46 accel -- accel/accel.sh@60 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:08:03.267 11:18:46 accel -- accel/accel.sh@62 -- # spdk_tgt_pid=834672 00:08:03.267 11:18:46 accel -- accel/accel.sh@63 -- # waitforlisten 834672 00:08:03.267 11:18:46 accel -- accel/accel.sh@61 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt -c /dev/fd/63 00:08:03.267 11:18:46 accel -- common/autotest_common.sh@829 -- # '[' -z 834672 ']' 00:08:03.267 11:18:46 accel -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:03.267 11:18:46 accel -- common/autotest_common.sh@834 -- # local max_retries=100 00:08:03.267 11:18:46 accel -- accel/accel.sh@61 -- # build_accel_config 00:08:03.267 11:18:46 accel -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:03.267 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:03.267 11:18:46 accel -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:03.267 11:18:46 accel -- common/autotest_common.sh@838 -- # xtrace_disable 00:08:03.267 11:18:46 accel -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:03.267 11:18:46 accel -- common/autotest_common.sh@10 -- # set +x 00:08:03.267 11:18:46 accel -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:03.267 11:18:46 accel -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:03.267 11:18:46 accel -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:03.267 11:18:46 accel -- accel/accel.sh@40 -- # local IFS=, 00:08:03.267 11:18:46 accel -- accel/accel.sh@41 -- # jq -r . 00:08:03.525 [2024-07-15 11:18:46.892959] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:03.525 [2024-07-15 11:18:46.893040] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid834672 ] 00:08:03.525 [2024-07-15 11:18:47.014736] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:03.525 [2024-07-15 11:18:47.117524] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:04.089 11:18:47 accel -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:08:04.089 11:18:47 accel -- common/autotest_common.sh@862 -- # return 0 00:08:04.346 11:18:47 accel -- accel/accel.sh@65 -- # [[ 0 -gt 0 ]] 00:08:04.346 11:18:47 accel -- accel/accel.sh@66 -- # [[ 0 -gt 0 ]] 00:08:04.346 11:18:47 accel -- accel/accel.sh@67 -- # [[ 0 -gt 0 ]] 00:08:04.346 11:18:47 accel -- accel/accel.sh@68 -- # [[ -n '' ]] 00:08:04.346 11:18:47 accel -- accel/accel.sh@70 -- # exp_opcs=($($rpc_py accel_get_opc_assignments | jq -r ". | to_entries | map(\"\(.key)=\(.value)\") | .[]")) 00:08:04.346 11:18:47 accel -- accel/accel.sh@70 -- # rpc_cmd accel_get_opc_assignments 00:08:04.346 11:18:47 accel -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:04.346 11:18:47 accel -- common/autotest_common.sh@10 -- # set +x 00:08:04.346 11:18:47 accel -- accel/accel.sh@70 -- # jq -r '. | to_entries | map("\(.key)=\(.value)") | .[]' 00:08:04.346 11:18:47 accel -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:04.346 11:18:47 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # IFS== 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:04.346 11:18:47 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:04.346 11:18:47 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # IFS== 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:04.346 11:18:47 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:04.346 11:18:47 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # IFS== 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:04.346 11:18:47 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:04.346 11:18:47 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # IFS== 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:04.346 11:18:47 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:04.346 11:18:47 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # IFS== 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:04.346 11:18:47 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:04.346 11:18:47 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # IFS== 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:04.346 11:18:47 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:04.346 11:18:47 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # IFS== 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:04.346 11:18:47 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:04.346 11:18:47 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # IFS== 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:04.346 11:18:47 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:04.346 11:18:47 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # IFS== 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:04.346 11:18:47 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:04.346 11:18:47 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # IFS== 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:04.346 11:18:47 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:04.346 11:18:47 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # IFS== 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:04.346 11:18:47 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:04.346 11:18:47 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # IFS== 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:04.346 11:18:47 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:04.346 11:18:47 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # IFS== 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:04.346 11:18:47 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:04.346 11:18:47 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # IFS== 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:04.346 11:18:47 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:04.346 11:18:47 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # IFS== 00:08:04.346 11:18:47 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:04.346 11:18:47 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:04.346 11:18:47 accel -- accel/accel.sh@75 -- # killprocess 834672 00:08:04.346 11:18:47 accel -- common/autotest_common.sh@948 -- # '[' -z 834672 ']' 00:08:04.346 11:18:47 accel -- common/autotest_common.sh@952 -- # kill -0 834672 00:08:04.346 11:18:47 accel -- common/autotest_common.sh@953 -- # uname 00:08:04.346 11:18:47 accel -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:08:04.346 11:18:47 accel -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 834672 00:08:04.346 11:18:47 accel -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:08:04.346 11:18:47 accel -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:08:04.346 11:18:47 accel -- common/autotest_common.sh@966 -- # echo 'killing process with pid 834672' 00:08:04.346 killing process with pid 834672 00:08:04.346 11:18:47 accel -- common/autotest_common.sh@967 -- # kill 834672 00:08:04.346 11:18:47 accel -- common/autotest_common.sh@972 -- # wait 834672 00:08:04.602 11:18:48 accel -- accel/accel.sh@76 -- # trap - ERR 00:08:04.602 11:18:48 accel -- accel/accel.sh@89 -- # run_test accel_help accel_perf -h 00:08:04.602 11:18:48 accel -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:08:04.602 11:18:48 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:04.602 11:18:48 accel -- common/autotest_common.sh@10 -- # set +x 00:08:04.860 11:18:48 accel.accel_help -- common/autotest_common.sh@1123 -- # accel_perf -h 00:08:04.860 11:18:48 accel.accel_help -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -h 00:08:04.860 11:18:48 accel.accel_help -- accel/accel.sh@12 -- # build_accel_config 00:08:04.860 11:18:48 accel.accel_help -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:04.860 11:18:48 accel.accel_help -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:04.860 11:18:48 accel.accel_help -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:04.860 11:18:48 accel.accel_help -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:04.860 11:18:48 accel.accel_help -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:04.860 11:18:48 accel.accel_help -- accel/accel.sh@40 -- # local IFS=, 00:08:04.860 11:18:48 accel.accel_help -- accel/accel.sh@41 -- # jq -r . 00:08:04.860 11:18:48 accel.accel_help -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:04.860 11:18:48 accel.accel_help -- common/autotest_common.sh@10 -- # set +x 00:08:04.860 11:18:48 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:04.860 11:18:48 accel -- accel/accel.sh@91 -- # run_test accel_missing_filename NOT accel_perf -t 1 -w compress 00:08:04.860 11:18:48 accel -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:08:04.860 11:18:48 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:04.860 11:18:48 accel -- common/autotest_common.sh@10 -- # set +x 00:08:04.860 ************************************ 00:08:04.860 START TEST accel_missing_filename 00:08:04.860 ************************************ 00:08:04.860 11:18:48 accel.accel_missing_filename -- common/autotest_common.sh@1123 -- # NOT accel_perf -t 1 -w compress 00:08:04.860 11:18:48 accel.accel_missing_filename -- common/autotest_common.sh@648 -- # local es=0 00:08:04.860 11:18:48 accel.accel_missing_filename -- common/autotest_common.sh@650 -- # valid_exec_arg accel_perf -t 1 -w compress 00:08:04.860 11:18:48 accel.accel_missing_filename -- common/autotest_common.sh@636 -- # local arg=accel_perf 00:08:04.860 11:18:48 accel.accel_missing_filename -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:08:04.860 11:18:48 accel.accel_missing_filename -- common/autotest_common.sh@640 -- # type -t accel_perf 00:08:04.860 11:18:48 accel.accel_missing_filename -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:08:04.860 11:18:48 accel.accel_missing_filename -- common/autotest_common.sh@651 -- # accel_perf -t 1 -w compress 00:08:04.860 11:18:48 accel.accel_missing_filename -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w compress 00:08:04.860 11:18:48 accel.accel_missing_filename -- accel/accel.sh@12 -- # build_accel_config 00:08:04.860 11:18:48 accel.accel_missing_filename -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:04.860 11:18:48 accel.accel_missing_filename -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:04.860 11:18:48 accel.accel_missing_filename -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:04.860 11:18:48 accel.accel_missing_filename -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:04.860 11:18:48 accel.accel_missing_filename -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:04.860 11:18:48 accel.accel_missing_filename -- accel/accel.sh@40 -- # local IFS=, 00:08:04.860 11:18:48 accel.accel_missing_filename -- accel/accel.sh@41 -- # jq -r . 00:08:04.860 [2024-07-15 11:18:48.361922] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:04.860 [2024-07-15 11:18:48.362000] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid834984 ] 00:08:05.117 [2024-07-15 11:18:48.491270] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:05.117 [2024-07-15 11:18:48.592270] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:05.117 [2024-07-15 11:18:48.665420] app.c:1053:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:08:05.374 [2024-07-15 11:18:48.740418] accel_perf.c:1463:main: *ERROR*: ERROR starting application 00:08:05.374 A filename is required. 00:08:05.374 11:18:48 accel.accel_missing_filename -- common/autotest_common.sh@651 -- # es=234 00:08:05.374 11:18:48 accel.accel_missing_filename -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:08:05.374 11:18:48 accel.accel_missing_filename -- common/autotest_common.sh@660 -- # es=106 00:08:05.374 11:18:48 accel.accel_missing_filename -- common/autotest_common.sh@661 -- # case "$es" in 00:08:05.374 11:18:48 accel.accel_missing_filename -- common/autotest_common.sh@668 -- # es=1 00:08:05.374 11:18:48 accel.accel_missing_filename -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:08:05.374 00:08:05.374 real 0m0.512s 00:08:05.374 user 0m0.336s 00:08:05.374 sys 0m0.204s 00:08:05.374 11:18:48 accel.accel_missing_filename -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:05.374 11:18:48 accel.accel_missing_filename -- common/autotest_common.sh@10 -- # set +x 00:08:05.374 ************************************ 00:08:05.374 END TEST accel_missing_filename 00:08:05.374 ************************************ 00:08:05.374 11:18:48 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:05.374 11:18:48 accel -- accel/accel.sh@93 -- # run_test accel_compress_verify NOT accel_perf -t 1 -w compress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y 00:08:05.374 11:18:48 accel -- common/autotest_common.sh@1099 -- # '[' 10 -le 1 ']' 00:08:05.374 11:18:48 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:05.374 11:18:48 accel -- common/autotest_common.sh@10 -- # set +x 00:08:05.374 ************************************ 00:08:05.374 START TEST accel_compress_verify 00:08:05.374 ************************************ 00:08:05.374 11:18:48 accel.accel_compress_verify -- common/autotest_common.sh@1123 -- # NOT accel_perf -t 1 -w compress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y 00:08:05.374 11:18:48 accel.accel_compress_verify -- common/autotest_common.sh@648 -- # local es=0 00:08:05.374 11:18:48 accel.accel_compress_verify -- common/autotest_common.sh@650 -- # valid_exec_arg accel_perf -t 1 -w compress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y 00:08:05.374 11:18:48 accel.accel_compress_verify -- common/autotest_common.sh@636 -- # local arg=accel_perf 00:08:05.374 11:18:48 accel.accel_compress_verify -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:08:05.374 11:18:48 accel.accel_compress_verify -- common/autotest_common.sh@640 -- # type -t accel_perf 00:08:05.374 11:18:48 accel.accel_compress_verify -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:08:05.374 11:18:48 accel.accel_compress_verify -- common/autotest_common.sh@651 -- # accel_perf -t 1 -w compress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y 00:08:05.374 11:18:48 accel.accel_compress_verify -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w compress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y 00:08:05.374 11:18:48 accel.accel_compress_verify -- accel/accel.sh@12 -- # build_accel_config 00:08:05.374 11:18:48 accel.accel_compress_verify -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:05.374 11:18:48 accel.accel_compress_verify -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:05.374 11:18:48 accel.accel_compress_verify -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:05.374 11:18:48 accel.accel_compress_verify -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:05.374 11:18:48 accel.accel_compress_verify -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:05.374 11:18:48 accel.accel_compress_verify -- accel/accel.sh@40 -- # local IFS=, 00:08:05.374 11:18:48 accel.accel_compress_verify -- accel/accel.sh@41 -- # jq -r . 00:08:05.374 [2024-07-15 11:18:48.953377] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:05.374 [2024-07-15 11:18:48.953439] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid835053 ] 00:08:05.631 [2024-07-15 11:18:49.081825] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:05.631 [2024-07-15 11:18:49.181989] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:05.889 [2024-07-15 11:18:49.254488] app.c:1053:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:08:05.889 [2024-07-15 11:18:49.328580] accel_perf.c:1463:main: *ERROR*: ERROR starting application 00:08:05.889 00:08:05.889 Compression does not support the verify option, aborting. 00:08:05.889 11:18:49 accel.accel_compress_verify -- common/autotest_common.sh@651 -- # es=161 00:08:05.889 11:18:49 accel.accel_compress_verify -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:08:05.889 11:18:49 accel.accel_compress_verify -- common/autotest_common.sh@660 -- # es=33 00:08:05.889 11:18:49 accel.accel_compress_verify -- common/autotest_common.sh@661 -- # case "$es" in 00:08:05.889 11:18:49 accel.accel_compress_verify -- common/autotest_common.sh@668 -- # es=1 00:08:05.889 11:18:49 accel.accel_compress_verify -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:08:05.889 00:08:05.889 real 0m0.507s 00:08:05.889 user 0m0.340s 00:08:05.889 sys 0m0.194s 00:08:05.889 11:18:49 accel.accel_compress_verify -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:05.889 11:18:49 accel.accel_compress_verify -- common/autotest_common.sh@10 -- # set +x 00:08:05.889 ************************************ 00:08:05.889 END TEST accel_compress_verify 00:08:05.889 ************************************ 00:08:05.889 11:18:49 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:05.889 11:18:49 accel -- accel/accel.sh@95 -- # run_test accel_wrong_workload NOT accel_perf -t 1 -w foobar 00:08:05.889 11:18:49 accel -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:08:05.889 11:18:49 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:05.889 11:18:49 accel -- common/autotest_common.sh@10 -- # set +x 00:08:06.146 ************************************ 00:08:06.146 START TEST accel_wrong_workload 00:08:06.146 ************************************ 00:08:06.146 11:18:49 accel.accel_wrong_workload -- common/autotest_common.sh@1123 -- # NOT accel_perf -t 1 -w foobar 00:08:06.146 11:18:49 accel.accel_wrong_workload -- common/autotest_common.sh@648 -- # local es=0 00:08:06.146 11:18:49 accel.accel_wrong_workload -- common/autotest_common.sh@650 -- # valid_exec_arg accel_perf -t 1 -w foobar 00:08:06.146 11:18:49 accel.accel_wrong_workload -- common/autotest_common.sh@636 -- # local arg=accel_perf 00:08:06.146 11:18:49 accel.accel_wrong_workload -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:08:06.146 11:18:49 accel.accel_wrong_workload -- common/autotest_common.sh@640 -- # type -t accel_perf 00:08:06.146 11:18:49 accel.accel_wrong_workload -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:08:06.146 11:18:49 accel.accel_wrong_workload -- common/autotest_common.sh@651 -- # accel_perf -t 1 -w foobar 00:08:06.146 11:18:49 accel.accel_wrong_workload -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w foobar 00:08:06.146 11:18:49 accel.accel_wrong_workload -- accel/accel.sh@12 -- # build_accel_config 00:08:06.146 11:18:49 accel.accel_wrong_workload -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:06.146 11:18:49 accel.accel_wrong_workload -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:06.146 11:18:49 accel.accel_wrong_workload -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:06.146 11:18:49 accel.accel_wrong_workload -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:06.146 11:18:49 accel.accel_wrong_workload -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:06.146 11:18:49 accel.accel_wrong_workload -- accel/accel.sh@40 -- # local IFS=, 00:08:06.146 11:18:49 accel.accel_wrong_workload -- accel/accel.sh@41 -- # jq -r . 00:08:06.146 Unsupported workload type: foobar 00:08:06.146 [2024-07-15 11:18:49.541983] app.c:1451:spdk_app_parse_args: *ERROR*: Parsing app-specific command line parameter 'w' failed: 1 00:08:06.146 accel_perf options: 00:08:06.146 [-h help message] 00:08:06.146 [-q queue depth per core] 00:08:06.146 [-C for supported workloads, use this value to configure the io vector size to test (default 1) 00:08:06.146 [-T number of threads per core 00:08:06.146 [-o transfer size in bytes (default: 4KiB. For compress/decompress, 0 means the input file size)] 00:08:06.146 [-t time in seconds] 00:08:06.146 [-w workload type must be one of these: copy, fill, crc32c, copy_crc32c, compare, compress, decompress, dualcast, xor, 00:08:06.146 [ dif_verify, dif_verify_copy, dif_generate, dif_generate_copy 00:08:06.146 [-M assign module to the operation, not compatible with accel_assign_opc RPC 00:08:06.146 [-l for compress/decompress workloads, name of uncompressed input file 00:08:06.146 [-S for crc32c workload, use this seed value (default 0) 00:08:06.146 [-P for compare workload, percentage of operations that should miscompare (percent, default 0) 00:08:06.146 [-f for fill workload, use this BYTE value (default 255) 00:08:06.146 [-x for xor workload, use this number of source buffers (default, minimum: 2)] 00:08:06.146 [-y verify result if this switch is on] 00:08:06.146 [-a tasks to allocate per core (default: same value as -q)] 00:08:06.146 Can be used to spread operations across a wider range of memory. 00:08:06.146 11:18:49 accel.accel_wrong_workload -- common/autotest_common.sh@651 -- # es=1 00:08:06.146 11:18:49 accel.accel_wrong_workload -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:08:06.146 11:18:49 accel.accel_wrong_workload -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:08:06.146 11:18:49 accel.accel_wrong_workload -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:08:06.146 00:08:06.146 real 0m0.043s 00:08:06.146 user 0m0.021s 00:08:06.146 sys 0m0.022s 00:08:06.146 11:18:49 accel.accel_wrong_workload -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:06.146 11:18:49 accel.accel_wrong_workload -- common/autotest_common.sh@10 -- # set +x 00:08:06.146 ************************************ 00:08:06.146 END TEST accel_wrong_workload 00:08:06.146 ************************************ 00:08:06.146 Error: writing output failed: Broken pipe 00:08:06.146 11:18:49 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:06.146 11:18:49 accel -- accel/accel.sh@97 -- # run_test accel_negative_buffers NOT accel_perf -t 1 -w xor -y -x -1 00:08:06.147 11:18:49 accel -- common/autotest_common.sh@1099 -- # '[' 10 -le 1 ']' 00:08:06.147 11:18:49 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:06.147 11:18:49 accel -- common/autotest_common.sh@10 -- # set +x 00:08:06.147 ************************************ 00:08:06.147 START TEST accel_negative_buffers 00:08:06.147 ************************************ 00:08:06.147 11:18:49 accel.accel_negative_buffers -- common/autotest_common.sh@1123 -- # NOT accel_perf -t 1 -w xor -y -x -1 00:08:06.147 11:18:49 accel.accel_negative_buffers -- common/autotest_common.sh@648 -- # local es=0 00:08:06.147 11:18:49 accel.accel_negative_buffers -- common/autotest_common.sh@650 -- # valid_exec_arg accel_perf -t 1 -w xor -y -x -1 00:08:06.147 11:18:49 accel.accel_negative_buffers -- common/autotest_common.sh@636 -- # local arg=accel_perf 00:08:06.147 11:18:49 accel.accel_negative_buffers -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:08:06.147 11:18:49 accel.accel_negative_buffers -- common/autotest_common.sh@640 -- # type -t accel_perf 00:08:06.147 11:18:49 accel.accel_negative_buffers -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:08:06.147 11:18:49 accel.accel_negative_buffers -- common/autotest_common.sh@651 -- # accel_perf -t 1 -w xor -y -x -1 00:08:06.147 11:18:49 accel.accel_negative_buffers -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w xor -y -x -1 00:08:06.147 11:18:49 accel.accel_negative_buffers -- accel/accel.sh@12 -- # build_accel_config 00:08:06.147 11:18:49 accel.accel_negative_buffers -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:06.147 11:18:49 accel.accel_negative_buffers -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:06.147 11:18:49 accel.accel_negative_buffers -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:06.147 11:18:49 accel.accel_negative_buffers -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:06.147 11:18:49 accel.accel_negative_buffers -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:06.147 11:18:49 accel.accel_negative_buffers -- accel/accel.sh@40 -- # local IFS=, 00:08:06.147 11:18:49 accel.accel_negative_buffers -- accel/accel.sh@41 -- # jq -r . 00:08:06.147 -x option must be non-negative. 00:08:06.147 [2024-07-15 11:18:49.663287] app.c:1451:spdk_app_parse_args: *ERROR*: Parsing app-specific command line parameter 'x' failed: 1 00:08:06.147 accel_perf options: 00:08:06.147 [-h help message] 00:08:06.147 [-q queue depth per core] 00:08:06.147 [-C for supported workloads, use this value to configure the io vector size to test (default 1) 00:08:06.147 [-T number of threads per core 00:08:06.147 [-o transfer size in bytes (default: 4KiB. For compress/decompress, 0 means the input file size)] 00:08:06.147 [-t time in seconds] 00:08:06.147 [-w workload type must be one of these: copy, fill, crc32c, copy_crc32c, compare, compress, decompress, dualcast, xor, 00:08:06.147 [ dif_verify, dif_verify_copy, dif_generate, dif_generate_copy 00:08:06.147 [-M assign module to the operation, not compatible with accel_assign_opc RPC 00:08:06.147 [-l for compress/decompress workloads, name of uncompressed input file 00:08:06.147 [-S for crc32c workload, use this seed value (default 0) 00:08:06.147 [-P for compare workload, percentage of operations that should miscompare (percent, default 0) 00:08:06.147 [-f for fill workload, use this BYTE value (default 255) 00:08:06.147 [-x for xor workload, use this number of source buffers (default, minimum: 2)] 00:08:06.147 [-y verify result if this switch is on] 00:08:06.147 [-a tasks to allocate per core (default: same value as -q)] 00:08:06.147 Can be used to spread operations across a wider range of memory. 00:08:06.147 11:18:49 accel.accel_negative_buffers -- common/autotest_common.sh@651 -- # es=1 00:08:06.147 11:18:49 accel.accel_negative_buffers -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:08:06.147 11:18:49 accel.accel_negative_buffers -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:08:06.147 11:18:49 accel.accel_negative_buffers -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:08:06.147 00:08:06.147 real 0m0.042s 00:08:06.147 user 0m0.020s 00:08:06.147 sys 0m0.021s 00:08:06.147 11:18:49 accel.accel_negative_buffers -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:06.147 11:18:49 accel.accel_negative_buffers -- common/autotest_common.sh@10 -- # set +x 00:08:06.147 ************************************ 00:08:06.147 END TEST accel_negative_buffers 00:08:06.147 ************************************ 00:08:06.147 Error: writing output failed: Broken pipe 00:08:06.147 11:18:49 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:06.147 11:18:49 accel -- accel/accel.sh@101 -- # run_test accel_crc32c accel_test -t 1 -w crc32c -S 32 -y 00:08:06.147 11:18:49 accel -- common/autotest_common.sh@1099 -- # '[' 9 -le 1 ']' 00:08:06.147 11:18:49 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:06.147 11:18:49 accel -- common/autotest_common.sh@10 -- # set +x 00:08:06.404 ************************************ 00:08:06.404 START TEST accel_crc32c 00:08:06.404 ************************************ 00:08:06.404 11:18:49 accel.accel_crc32c -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w crc32c -S 32 -y 00:08:06.404 11:18:49 accel.accel_crc32c -- accel/accel.sh@16 -- # local accel_opc 00:08:06.404 11:18:49 accel.accel_crc32c -- accel/accel.sh@17 -- # local accel_module 00:08:06.404 11:18:49 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:06.404 11:18:49 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:06.404 11:18:49 accel.accel_crc32c -- accel/accel.sh@15 -- # accel_perf -t 1 -w crc32c -S 32 -y 00:08:06.404 11:18:49 accel.accel_crc32c -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w crc32c -S 32 -y 00:08:06.404 11:18:49 accel.accel_crc32c -- accel/accel.sh@12 -- # build_accel_config 00:08:06.404 11:18:49 accel.accel_crc32c -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:06.404 11:18:49 accel.accel_crc32c -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:06.404 11:18:49 accel.accel_crc32c -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:06.404 11:18:49 accel.accel_crc32c -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:06.404 11:18:49 accel.accel_crc32c -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:06.404 11:18:49 accel.accel_crc32c -- accel/accel.sh@40 -- # local IFS=, 00:08:06.404 11:18:49 accel.accel_crc32c -- accel/accel.sh@41 -- # jq -r . 00:08:06.404 [2024-07-15 11:18:49.784911] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:06.404 [2024-07-15 11:18:49.784985] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid835280 ] 00:08:06.404 [2024-07-15 11:18:49.915625] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:06.661 [2024-07-15 11:18:50.021807] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:06.661 11:18:50 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:08:06.661 11:18:50 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:06.661 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:06.661 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@20 -- # val=0x1 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@20 -- # val=crc32c 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@23 -- # accel_opc=crc32c 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@20 -- # val=32 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@20 -- # val=software 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@22 -- # accel_module=software 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@20 -- # val=32 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@20 -- # val=32 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@20 -- # val=1 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@20 -- # val='1 seconds' 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@20 -- # val=Yes 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:06.662 11:18:50 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@27 -- # [[ -n software ]] 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@27 -- # [[ -n crc32c ]] 00:08:08.033 11:18:51 accel.accel_crc32c -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:08:08.033 00:08:08.033 real 0m1.506s 00:08:08.033 user 0m0.011s 00:08:08.033 sys 0m0.002s 00:08:08.033 11:18:51 accel.accel_crc32c -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:08.033 11:18:51 accel.accel_crc32c -- common/autotest_common.sh@10 -- # set +x 00:08:08.033 ************************************ 00:08:08.033 END TEST accel_crc32c 00:08:08.033 ************************************ 00:08:08.033 11:18:51 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:08.033 11:18:51 accel -- accel/accel.sh@102 -- # run_test accel_crc32c_C2 accel_test -t 1 -w crc32c -y -C 2 00:08:08.033 11:18:51 accel -- common/autotest_common.sh@1099 -- # '[' 9 -le 1 ']' 00:08:08.033 11:18:51 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:08.033 11:18:51 accel -- common/autotest_common.sh@10 -- # set +x 00:08:08.033 ************************************ 00:08:08.033 START TEST accel_crc32c_C2 00:08:08.033 ************************************ 00:08:08.033 11:18:51 accel.accel_crc32c_C2 -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w crc32c -y -C 2 00:08:08.033 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@16 -- # local accel_opc 00:08:08.033 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@17 -- # local accel_module 00:08:08.033 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:08.033 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:08.033 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@15 -- # accel_perf -t 1 -w crc32c -y -C 2 00:08:08.033 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w crc32c -y -C 2 00:08:08.033 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@12 -- # build_accel_config 00:08:08.033 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:08.033 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:08.033 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:08.033 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:08.033 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:08.033 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@40 -- # local IFS=, 00:08:08.033 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@41 -- # jq -r . 00:08:08.033 [2024-07-15 11:18:51.374012] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:08.033 [2024-07-15 11:18:51.374068] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid835478 ] 00:08:08.033 [2024-07-15 11:18:51.488955] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:08.033 [2024-07-15 11:18:51.592489] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val=0x1 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val=crc32c 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@23 -- # accel_opc=crc32c 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val=0 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val=software 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@22 -- # accel_module=software 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val=32 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val=32 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val=1 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val='1 seconds' 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val=Yes 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:08.292 11:18:51 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@27 -- # [[ -n software ]] 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@27 -- # [[ -n crc32c ]] 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:08:09.666 00:08:09.666 real 0m1.494s 00:08:09.666 user 0m0.011s 00:08:09.666 sys 0m0.003s 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:09.666 11:18:52 accel.accel_crc32c_C2 -- common/autotest_common.sh@10 -- # set +x 00:08:09.666 ************************************ 00:08:09.666 END TEST accel_crc32c_C2 00:08:09.666 ************************************ 00:08:09.666 11:18:52 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:09.666 11:18:52 accel -- accel/accel.sh@103 -- # run_test accel_copy accel_test -t 1 -w copy -y 00:08:09.666 11:18:52 accel -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:08:09.666 11:18:52 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:09.666 11:18:52 accel -- common/autotest_common.sh@10 -- # set +x 00:08:09.666 ************************************ 00:08:09.666 START TEST accel_copy 00:08:09.666 ************************************ 00:08:09.666 11:18:52 accel.accel_copy -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w copy -y 00:08:09.666 11:18:52 accel.accel_copy -- accel/accel.sh@16 -- # local accel_opc 00:08:09.666 11:18:52 accel.accel_copy -- accel/accel.sh@17 -- # local accel_module 00:08:09.666 11:18:52 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:09.666 11:18:52 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:09.666 11:18:52 accel.accel_copy -- accel/accel.sh@15 -- # accel_perf -t 1 -w copy -y 00:08:09.666 11:18:52 accel.accel_copy -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w copy -y 00:08:09.666 11:18:52 accel.accel_copy -- accel/accel.sh@12 -- # build_accel_config 00:08:09.666 11:18:52 accel.accel_copy -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:09.666 11:18:52 accel.accel_copy -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:09.666 11:18:52 accel.accel_copy -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:09.666 11:18:52 accel.accel_copy -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:09.666 11:18:52 accel.accel_copy -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:09.666 11:18:52 accel.accel_copy -- accel/accel.sh@40 -- # local IFS=, 00:08:09.666 11:18:52 accel.accel_copy -- accel/accel.sh@41 -- # jq -r . 00:08:09.666 [2024-07-15 11:18:52.938374] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:09.666 [2024-07-15 11:18:52.938437] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid835674 ] 00:08:09.666 [2024-07-15 11:18:53.068395] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:09.666 [2024-07-15 11:18:53.169520] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:09.666 11:18:53 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:08:09.666 11:18:53 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:09.666 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:09.666 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:09.666 11:18:53 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:08:09.666 11:18:53 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:09.666 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:09.666 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@20 -- # val=0x1 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@20 -- # val=copy 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@23 -- # accel_opc=copy 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@20 -- # val=software 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@22 -- # accel_module=software 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@20 -- # val=32 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@20 -- # val=32 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@20 -- # val=1 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@20 -- # val='1 seconds' 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@20 -- # val=Yes 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:09.667 11:18:53 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@27 -- # [[ -n software ]] 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@27 -- # [[ -n copy ]] 00:08:11.040 11:18:54 accel.accel_copy -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:08:11.040 00:08:11.040 real 0m1.499s 00:08:11.040 user 0m0.010s 00:08:11.040 sys 0m0.001s 00:08:11.040 11:18:54 accel.accel_copy -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:11.040 11:18:54 accel.accel_copy -- common/autotest_common.sh@10 -- # set +x 00:08:11.040 ************************************ 00:08:11.040 END TEST accel_copy 00:08:11.040 ************************************ 00:08:11.040 11:18:54 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:11.040 11:18:54 accel -- accel/accel.sh@104 -- # run_test accel_fill accel_test -t 1 -w fill -f 128 -q 64 -a 64 -y 00:08:11.040 11:18:54 accel -- common/autotest_common.sh@1099 -- # '[' 13 -le 1 ']' 00:08:11.040 11:18:54 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:11.040 11:18:54 accel -- common/autotest_common.sh@10 -- # set +x 00:08:11.040 ************************************ 00:08:11.040 START TEST accel_fill 00:08:11.040 ************************************ 00:08:11.040 11:18:54 accel.accel_fill -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w fill -f 128 -q 64 -a 64 -y 00:08:11.040 11:18:54 accel.accel_fill -- accel/accel.sh@16 -- # local accel_opc 00:08:11.040 11:18:54 accel.accel_fill -- accel/accel.sh@17 -- # local accel_module 00:08:11.040 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:11.040 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:11.040 11:18:54 accel.accel_fill -- accel/accel.sh@15 -- # accel_perf -t 1 -w fill -f 128 -q 64 -a 64 -y 00:08:11.040 11:18:54 accel.accel_fill -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w fill -f 128 -q 64 -a 64 -y 00:08:11.040 11:18:54 accel.accel_fill -- accel/accel.sh@12 -- # build_accel_config 00:08:11.041 11:18:54 accel.accel_fill -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:11.041 11:18:54 accel.accel_fill -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:11.041 11:18:54 accel.accel_fill -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:11.041 11:18:54 accel.accel_fill -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:11.041 11:18:54 accel.accel_fill -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:11.041 11:18:54 accel.accel_fill -- accel/accel.sh@40 -- # local IFS=, 00:08:11.041 11:18:54 accel.accel_fill -- accel/accel.sh@41 -- # jq -r . 00:08:11.041 [2024-07-15 11:18:54.515851] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:11.041 [2024-07-15 11:18:54.515914] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid835874 ] 00:08:11.041 [2024-07-15 11:18:54.628534] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:11.299 [2024-07-15 11:18:54.736226] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:11.299 11:18:54 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@20 -- # val=0x1 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@20 -- # val=fill 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@23 -- # accel_opc=fill 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@20 -- # val=0x80 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@20 -- # val=software 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@22 -- # accel_module=software 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@20 -- # val=64 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@20 -- # val=64 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@20 -- # val=1 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@20 -- # val='1 seconds' 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@20 -- # val=Yes 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:11.300 11:18:54 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@27 -- # [[ -n software ]] 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@27 -- # [[ -n fill ]] 00:08:12.674 11:18:55 accel.accel_fill -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:08:12.674 00:08:12.674 real 0m1.493s 00:08:12.674 user 0m0.011s 00:08:12.674 sys 0m0.002s 00:08:12.674 11:18:55 accel.accel_fill -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:12.674 11:18:55 accel.accel_fill -- common/autotest_common.sh@10 -- # set +x 00:08:12.674 ************************************ 00:08:12.674 END TEST accel_fill 00:08:12.674 ************************************ 00:08:12.674 11:18:56 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:12.674 11:18:56 accel -- accel/accel.sh@105 -- # run_test accel_copy_crc32c accel_test -t 1 -w copy_crc32c -y 00:08:12.674 11:18:56 accel -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:08:12.674 11:18:56 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:12.674 11:18:56 accel -- common/autotest_common.sh@10 -- # set +x 00:08:12.674 ************************************ 00:08:12.674 START TEST accel_copy_crc32c 00:08:12.674 ************************************ 00:08:12.674 11:18:56 accel.accel_copy_crc32c -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w copy_crc32c -y 00:08:12.674 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@16 -- # local accel_opc 00:08:12.674 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@17 -- # local accel_module 00:08:12.674 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:12.674 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:12.674 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@15 -- # accel_perf -t 1 -w copy_crc32c -y 00:08:12.674 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w copy_crc32c -y 00:08:12.674 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@12 -- # build_accel_config 00:08:12.674 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:12.674 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:12.674 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:12.674 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:12.674 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:12.674 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@40 -- # local IFS=, 00:08:12.674 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@41 -- # jq -r . 00:08:12.674 [2024-07-15 11:18:56.080289] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:12.674 [2024-07-15 11:18:56.080348] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid836080 ] 00:08:12.674 [2024-07-15 11:18:56.208872] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:12.930 [2024-07-15 11:18:56.307109] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val=0x1 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val=copy_crc32c 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@23 -- # accel_opc=copy_crc32c 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val=0 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val=software 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@22 -- # accel_module=software 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val=32 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val=32 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val=1 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val='1 seconds' 00:08:12.930 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:12.931 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:12.931 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:12.931 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val=Yes 00:08:12.931 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:12.931 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:12.931 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:12.931 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:08:12.931 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:12.931 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:12.931 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:12.931 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:08:12.931 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:12.931 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:12.931 11:18:56 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:14.330 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:08:14.330 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:14.330 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:14.330 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:14.330 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:08:14.330 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@27 -- # [[ -n software ]] 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@27 -- # [[ -n copy_crc32c ]] 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:08:14.331 00:08:14.331 real 0m1.479s 00:08:14.331 user 0m0.011s 00:08:14.331 sys 0m0.001s 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:14.331 11:18:57 accel.accel_copy_crc32c -- common/autotest_common.sh@10 -- # set +x 00:08:14.331 ************************************ 00:08:14.331 END TEST accel_copy_crc32c 00:08:14.331 ************************************ 00:08:14.331 11:18:57 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:14.331 11:18:57 accel -- accel/accel.sh@106 -- # run_test accel_copy_crc32c_C2 accel_test -t 1 -w copy_crc32c -y -C 2 00:08:14.331 11:18:57 accel -- common/autotest_common.sh@1099 -- # '[' 9 -le 1 ']' 00:08:14.331 11:18:57 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:14.331 11:18:57 accel -- common/autotest_common.sh@10 -- # set +x 00:08:14.331 ************************************ 00:08:14.331 START TEST accel_copy_crc32c_C2 00:08:14.331 ************************************ 00:08:14.331 11:18:57 accel.accel_copy_crc32c_C2 -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w copy_crc32c -y -C 2 00:08:14.331 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@16 -- # local accel_opc 00:08:14.331 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@17 -- # local accel_module 00:08:14.331 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:14.331 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:14.331 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@15 -- # accel_perf -t 1 -w copy_crc32c -y -C 2 00:08:14.331 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w copy_crc32c -y -C 2 00:08:14.331 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@12 -- # build_accel_config 00:08:14.331 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:14.331 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:14.331 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:14.331 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:14.331 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:14.331 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@40 -- # local IFS=, 00:08:14.331 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@41 -- # jq -r . 00:08:14.331 [2024-07-15 11:18:57.634214] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:14.331 [2024-07-15 11:18:57.634274] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid836398 ] 00:08:14.331 [2024-07-15 11:18:57.762819] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:14.331 [2024-07-15 11:18:57.863225] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:14.588 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:14.588 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:14.588 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:14.588 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:14.588 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:14.588 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:14.588 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:14.588 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:14.588 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val=0x1 00:08:14.588 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:14.588 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:14.588 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:14.588 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val=copy_crc32c 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@23 -- # accel_opc=copy_crc32c 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val=0 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val='8192 bytes' 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val=software 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@22 -- # accel_module=software 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val=32 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val=32 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val=1 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val='1 seconds' 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val=Yes 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:14.589 11:18:57 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@27 -- # [[ -n software ]] 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@27 -- # [[ -n copy_crc32c ]] 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:08:15.522 00:08:15.522 real 0m1.497s 00:08:15.522 user 0m0.010s 00:08:15.522 sys 0m0.003s 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:15.522 11:18:59 accel.accel_copy_crc32c_C2 -- common/autotest_common.sh@10 -- # set +x 00:08:15.522 ************************************ 00:08:15.522 END TEST accel_copy_crc32c_C2 00:08:15.522 ************************************ 00:08:15.780 11:18:59 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:15.780 11:18:59 accel -- accel/accel.sh@107 -- # run_test accel_dualcast accel_test -t 1 -w dualcast -y 00:08:15.780 11:18:59 accel -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:08:15.780 11:18:59 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:15.780 11:18:59 accel -- common/autotest_common.sh@10 -- # set +x 00:08:15.780 ************************************ 00:08:15.780 START TEST accel_dualcast 00:08:15.780 ************************************ 00:08:15.780 11:18:59 accel.accel_dualcast -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w dualcast -y 00:08:15.780 11:18:59 accel.accel_dualcast -- accel/accel.sh@16 -- # local accel_opc 00:08:15.780 11:18:59 accel.accel_dualcast -- accel/accel.sh@17 -- # local accel_module 00:08:15.780 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:15.780 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:15.780 11:18:59 accel.accel_dualcast -- accel/accel.sh@15 -- # accel_perf -t 1 -w dualcast -y 00:08:15.780 11:18:59 accel.accel_dualcast -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w dualcast -y 00:08:15.780 11:18:59 accel.accel_dualcast -- accel/accel.sh@12 -- # build_accel_config 00:08:15.780 11:18:59 accel.accel_dualcast -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:15.780 11:18:59 accel.accel_dualcast -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:15.780 11:18:59 accel.accel_dualcast -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:15.780 11:18:59 accel.accel_dualcast -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:15.780 11:18:59 accel.accel_dualcast -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:15.780 11:18:59 accel.accel_dualcast -- accel/accel.sh@40 -- # local IFS=, 00:08:15.780 11:18:59 accel.accel_dualcast -- accel/accel.sh@41 -- # jq -r . 00:08:15.780 [2024-07-15 11:18:59.214042] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:15.780 [2024-07-15 11:18:59.214107] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid836624 ] 00:08:15.780 [2024-07-15 11:18:59.341698] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:16.038 [2024-07-15 11:18:59.443322] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:16.038 11:18:59 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:08:16.038 11:18:59 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:16.038 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:16.038 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:16.038 11:18:59 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:08:16.038 11:18:59 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:16.038 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:16.038 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:16.038 11:18:59 accel.accel_dualcast -- accel/accel.sh@20 -- # val=0x1 00:08:16.038 11:18:59 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:16.038 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:16.038 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:16.038 11:18:59 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:08:16.038 11:18:59 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:16.038 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:16.038 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@20 -- # val=dualcast 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@23 -- # accel_opc=dualcast 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@20 -- # val=software 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@22 -- # accel_module=software 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@20 -- # val=32 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@20 -- # val=32 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@20 -- # val=1 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@20 -- # val='1 seconds' 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@20 -- # val=Yes 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:16.039 11:18:59 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@27 -- # [[ -n software ]] 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@27 -- # [[ -n dualcast ]] 00:08:17.412 11:19:00 accel.accel_dualcast -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:08:17.412 00:08:17.412 real 0m1.508s 00:08:17.412 user 0m0.009s 00:08:17.412 sys 0m0.003s 00:08:17.412 11:19:00 accel.accel_dualcast -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:17.412 11:19:00 accel.accel_dualcast -- common/autotest_common.sh@10 -- # set +x 00:08:17.412 ************************************ 00:08:17.412 END TEST accel_dualcast 00:08:17.412 ************************************ 00:08:17.412 11:19:00 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:17.412 11:19:00 accel -- accel/accel.sh@108 -- # run_test accel_compare accel_test -t 1 -w compare -y 00:08:17.412 11:19:00 accel -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:08:17.412 11:19:00 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:17.412 11:19:00 accel -- common/autotest_common.sh@10 -- # set +x 00:08:17.412 ************************************ 00:08:17.412 START TEST accel_compare 00:08:17.412 ************************************ 00:08:17.412 11:19:00 accel.accel_compare -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w compare -y 00:08:17.412 11:19:00 accel.accel_compare -- accel/accel.sh@16 -- # local accel_opc 00:08:17.412 11:19:00 accel.accel_compare -- accel/accel.sh@17 -- # local accel_module 00:08:17.412 11:19:00 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:17.412 11:19:00 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:17.412 11:19:00 accel.accel_compare -- accel/accel.sh@15 -- # accel_perf -t 1 -w compare -y 00:08:17.412 11:19:00 accel.accel_compare -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w compare -y 00:08:17.412 11:19:00 accel.accel_compare -- accel/accel.sh@12 -- # build_accel_config 00:08:17.412 11:19:00 accel.accel_compare -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:17.412 11:19:00 accel.accel_compare -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:17.412 11:19:00 accel.accel_compare -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:17.413 11:19:00 accel.accel_compare -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:17.413 11:19:00 accel.accel_compare -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:17.413 11:19:00 accel.accel_compare -- accel/accel.sh@40 -- # local IFS=, 00:08:17.413 11:19:00 accel.accel_compare -- accel/accel.sh@41 -- # jq -r . 00:08:17.413 [2024-07-15 11:19:00.796341] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:17.413 [2024-07-15 11:19:00.796406] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid836823 ] 00:08:17.413 [2024-07-15 11:19:00.922981] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:17.670 [2024-07-15 11:19:01.024414] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@20 -- # val=0x1 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@20 -- # val=compare 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@23 -- # accel_opc=compare 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:17.670 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@20 -- # val=software 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@22 -- # accel_module=software 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@20 -- # val=32 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@20 -- # val=32 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@20 -- # val=1 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@20 -- # val='1 seconds' 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@20 -- # val=Yes 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:17.671 11:19:01 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@27 -- # [[ -n software ]] 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@27 -- # [[ -n compare ]] 00:08:19.043 11:19:02 accel.accel_compare -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:08:19.043 00:08:19.043 real 0m1.499s 00:08:19.043 user 0m0.011s 00:08:19.043 sys 0m0.001s 00:08:19.043 11:19:02 accel.accel_compare -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:19.043 11:19:02 accel.accel_compare -- common/autotest_common.sh@10 -- # set +x 00:08:19.043 ************************************ 00:08:19.043 END TEST accel_compare 00:08:19.043 ************************************ 00:08:19.043 11:19:02 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:19.043 11:19:02 accel -- accel/accel.sh@109 -- # run_test accel_xor accel_test -t 1 -w xor -y 00:08:19.043 11:19:02 accel -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:08:19.043 11:19:02 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:19.043 11:19:02 accel -- common/autotest_common.sh@10 -- # set +x 00:08:19.043 ************************************ 00:08:19.043 START TEST accel_xor 00:08:19.043 ************************************ 00:08:19.043 11:19:02 accel.accel_xor -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w xor -y 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@16 -- # local accel_opc 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@17 -- # local accel_module 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@15 -- # accel_perf -t 1 -w xor -y 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w xor -y 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@12 -- # build_accel_config 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@40 -- # local IFS=, 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@41 -- # jq -r . 00:08:19.043 [2024-07-15 11:19:02.349467] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:19.043 [2024-07-15 11:19:02.349509] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid837017 ] 00:08:19.043 [2024-07-15 11:19:02.462761] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:19.043 [2024-07-15 11:19:02.564471] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@20 -- # val=0x1 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:19.043 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@20 -- # val=xor 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@23 -- # accel_opc=xor 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@20 -- # val=2 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@20 -- # val=software 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@22 -- # accel_module=software 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@20 -- # val=32 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@20 -- # val=32 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@20 -- # val=1 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@20 -- # val='1 seconds' 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@20 -- # val=Yes 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:19.044 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:19.301 11:19:02 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@27 -- # [[ -n software ]] 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@27 -- # [[ -n xor ]] 00:08:20.230 11:19:03 accel.accel_xor -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:08:20.230 00:08:20.230 real 0m1.468s 00:08:20.230 user 0m0.011s 00:08:20.230 sys 0m0.001s 00:08:20.230 11:19:03 accel.accel_xor -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:20.230 11:19:03 accel.accel_xor -- common/autotest_common.sh@10 -- # set +x 00:08:20.230 ************************************ 00:08:20.230 END TEST accel_xor 00:08:20.230 ************************************ 00:08:20.486 11:19:03 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:20.486 11:19:03 accel -- accel/accel.sh@110 -- # run_test accel_xor accel_test -t 1 -w xor -y -x 3 00:08:20.486 11:19:03 accel -- common/autotest_common.sh@1099 -- # '[' 9 -le 1 ']' 00:08:20.486 11:19:03 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:20.486 11:19:03 accel -- common/autotest_common.sh@10 -- # set +x 00:08:20.486 ************************************ 00:08:20.486 START TEST accel_xor 00:08:20.486 ************************************ 00:08:20.486 11:19:03 accel.accel_xor -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w xor -y -x 3 00:08:20.486 11:19:03 accel.accel_xor -- accel/accel.sh@16 -- # local accel_opc 00:08:20.486 11:19:03 accel.accel_xor -- accel/accel.sh@17 -- # local accel_module 00:08:20.486 11:19:03 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.486 11:19:03 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.486 11:19:03 accel.accel_xor -- accel/accel.sh@15 -- # accel_perf -t 1 -w xor -y -x 3 00:08:20.486 11:19:03 accel.accel_xor -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w xor -y -x 3 00:08:20.486 11:19:03 accel.accel_xor -- accel/accel.sh@12 -- # build_accel_config 00:08:20.486 11:19:03 accel.accel_xor -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:20.486 11:19:03 accel.accel_xor -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:20.486 11:19:03 accel.accel_xor -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:20.486 11:19:03 accel.accel_xor -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:20.486 11:19:03 accel.accel_xor -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:20.486 11:19:03 accel.accel_xor -- accel/accel.sh@40 -- # local IFS=, 00:08:20.486 11:19:03 accel.accel_xor -- accel/accel.sh@41 -- # jq -r . 00:08:20.486 [2024-07-15 11:19:03.913745] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:20.486 [2024-07-15 11:19:03.913814] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid837217 ] 00:08:20.486 [2024-07-15 11:19:04.047312] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:20.742 [2024-07-15 11:19:04.153151] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@20 -- # val=0x1 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@20 -- # val=xor 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@23 -- # accel_opc=xor 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@20 -- # val=3 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.742 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@20 -- # val=software 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@22 -- # accel_module=software 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@20 -- # val=32 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@20 -- # val=32 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@20 -- # val=1 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@20 -- # val='1 seconds' 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@20 -- # val=Yes 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:20.743 11:19:04 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@27 -- # [[ -n software ]] 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@27 -- # [[ -n xor ]] 00:08:22.109 11:19:05 accel.accel_xor -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:08:22.109 00:08:22.109 real 0m1.507s 00:08:22.109 user 0m0.013s 00:08:22.109 sys 0m0.000s 00:08:22.109 11:19:05 accel.accel_xor -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:22.109 11:19:05 accel.accel_xor -- common/autotest_common.sh@10 -- # set +x 00:08:22.109 ************************************ 00:08:22.109 END TEST accel_xor 00:08:22.109 ************************************ 00:08:22.109 11:19:05 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:22.109 11:19:05 accel -- accel/accel.sh@111 -- # run_test accel_dif_verify accel_test -t 1 -w dif_verify 00:08:22.109 11:19:05 accel -- common/autotest_common.sh@1099 -- # '[' 6 -le 1 ']' 00:08:22.109 11:19:05 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:22.109 11:19:05 accel -- common/autotest_common.sh@10 -- # set +x 00:08:22.109 ************************************ 00:08:22.109 START TEST accel_dif_verify 00:08:22.109 ************************************ 00:08:22.109 11:19:05 accel.accel_dif_verify -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w dif_verify 00:08:22.109 11:19:05 accel.accel_dif_verify -- accel/accel.sh@16 -- # local accel_opc 00:08:22.109 11:19:05 accel.accel_dif_verify -- accel/accel.sh@17 -- # local accel_module 00:08:22.109 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:22.109 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:22.109 11:19:05 accel.accel_dif_verify -- accel/accel.sh@15 -- # accel_perf -t 1 -w dif_verify 00:08:22.109 11:19:05 accel.accel_dif_verify -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w dif_verify 00:08:22.109 11:19:05 accel.accel_dif_verify -- accel/accel.sh@12 -- # build_accel_config 00:08:22.109 11:19:05 accel.accel_dif_verify -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:22.109 11:19:05 accel.accel_dif_verify -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:22.109 11:19:05 accel.accel_dif_verify -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:22.109 11:19:05 accel.accel_dif_verify -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:22.109 11:19:05 accel.accel_dif_verify -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:22.109 11:19:05 accel.accel_dif_verify -- accel/accel.sh@40 -- # local IFS=, 00:08:22.109 11:19:05 accel.accel_dif_verify -- accel/accel.sh@41 -- # jq -r . 00:08:22.109 [2024-07-15 11:19:05.488993] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:22.109 [2024-07-15 11:19:05.489052] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid837418 ] 00:08:22.109 [2024-07-15 11:19:05.617403] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:22.366 [2024-07-15 11:19:05.719078] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:22.366 11:19:05 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:08:22.366 11:19:05 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:22.366 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:22.366 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:22.366 11:19:05 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:08:22.366 11:19:05 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:22.366 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:22.366 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:22.366 11:19:05 accel.accel_dif_verify -- accel/accel.sh@20 -- # val=0x1 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@20 -- # val=dif_verify 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@23 -- # accel_opc=dif_verify 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@20 -- # val='512 bytes' 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@20 -- # val='8 bytes' 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@20 -- # val=software 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@22 -- # accel_module=software 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@20 -- # val=32 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@20 -- # val=32 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@20 -- # val=1 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@20 -- # val='1 seconds' 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@20 -- # val=No 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:22.367 11:19:05 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@27 -- # [[ -n software ]] 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@27 -- # [[ -n dif_verify ]] 00:08:23.739 11:19:06 accel.accel_dif_verify -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:08:23.739 00:08:23.739 real 0m1.494s 00:08:23.739 user 0m0.010s 00:08:23.739 sys 0m0.003s 00:08:23.739 11:19:06 accel.accel_dif_verify -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:23.739 11:19:06 accel.accel_dif_verify -- common/autotest_common.sh@10 -- # set +x 00:08:23.739 ************************************ 00:08:23.739 END TEST accel_dif_verify 00:08:23.739 ************************************ 00:08:23.739 11:19:06 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:23.739 11:19:06 accel -- accel/accel.sh@112 -- # run_test accel_dif_generate accel_test -t 1 -w dif_generate 00:08:23.739 11:19:06 accel -- common/autotest_common.sh@1099 -- # '[' 6 -le 1 ']' 00:08:23.739 11:19:06 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:23.739 11:19:06 accel -- common/autotest_common.sh@10 -- # set +x 00:08:23.739 ************************************ 00:08:23.739 START TEST accel_dif_generate 00:08:23.739 ************************************ 00:08:23.739 11:19:07 accel.accel_dif_generate -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w dif_generate 00:08:23.739 11:19:07 accel.accel_dif_generate -- accel/accel.sh@16 -- # local accel_opc 00:08:23.739 11:19:07 accel.accel_dif_generate -- accel/accel.sh@17 -- # local accel_module 00:08:23.739 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:23.739 11:19:07 accel.accel_dif_generate -- accel/accel.sh@15 -- # accel_perf -t 1 -w dif_generate 00:08:23.739 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:23.739 11:19:07 accel.accel_dif_generate -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w dif_generate 00:08:23.739 11:19:07 accel.accel_dif_generate -- accel/accel.sh@12 -- # build_accel_config 00:08:23.739 11:19:07 accel.accel_dif_generate -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:23.739 11:19:07 accel.accel_dif_generate -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:23.739 11:19:07 accel.accel_dif_generate -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:23.739 11:19:07 accel.accel_dif_generate -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:23.739 11:19:07 accel.accel_dif_generate -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:23.739 11:19:07 accel.accel_dif_generate -- accel/accel.sh@40 -- # local IFS=, 00:08:23.739 11:19:07 accel.accel_dif_generate -- accel/accel.sh@41 -- # jq -r . 00:08:23.739 [2024-07-15 11:19:07.041183] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:23.739 [2024-07-15 11:19:07.041227] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid837716 ] 00:08:23.739 [2024-07-15 11:19:07.153592] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:23.739 [2024-07-15 11:19:07.254303] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:23.739 11:19:07 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:08:23.739 11:19:07 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:23.739 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@20 -- # val=0x1 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@20 -- # val=dif_generate 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@23 -- # accel_opc=dif_generate 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@20 -- # val='512 bytes' 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@20 -- # val='8 bytes' 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@20 -- # val=software 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@22 -- # accel_module=software 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@20 -- # val=32 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@20 -- # val=32 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@20 -- # val=1 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@20 -- # val='1 seconds' 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@20 -- # val=No 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:23.740 11:19:07 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@27 -- # [[ -n software ]] 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@27 -- # [[ -n dif_generate ]] 00:08:25.113 11:19:08 accel.accel_dif_generate -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:08:25.113 00:08:25.113 real 0m1.471s 00:08:25.113 user 0m0.013s 00:08:25.113 sys 0m0.000s 00:08:25.113 11:19:08 accel.accel_dif_generate -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:25.113 11:19:08 accel.accel_dif_generate -- common/autotest_common.sh@10 -- # set +x 00:08:25.113 ************************************ 00:08:25.113 END TEST accel_dif_generate 00:08:25.113 ************************************ 00:08:25.113 11:19:08 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:25.113 11:19:08 accel -- accel/accel.sh@113 -- # run_test accel_dif_generate_copy accel_test -t 1 -w dif_generate_copy 00:08:25.113 11:19:08 accel -- common/autotest_common.sh@1099 -- # '[' 6 -le 1 ']' 00:08:25.113 11:19:08 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:25.113 11:19:08 accel -- common/autotest_common.sh@10 -- # set +x 00:08:25.113 ************************************ 00:08:25.113 START TEST accel_dif_generate_copy 00:08:25.113 ************************************ 00:08:25.113 11:19:08 accel.accel_dif_generate_copy -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w dif_generate_copy 00:08:25.113 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@16 -- # local accel_opc 00:08:25.113 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@17 -- # local accel_module 00:08:25.113 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:25.113 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:25.113 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@15 -- # accel_perf -t 1 -w dif_generate_copy 00:08:25.113 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@12 -- # build_accel_config 00:08:25.113 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w dif_generate_copy 00:08:25.113 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:25.113 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:25.113 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:25.113 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:25.113 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:25.113 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@40 -- # local IFS=, 00:08:25.113 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@41 -- # jq -r . 00:08:25.113 [2024-07-15 11:19:08.597791] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:25.113 [2024-07-15 11:19:08.597848] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid837966 ] 00:08:25.371 [2024-07-15 11:19:08.727413] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:25.371 [2024-07-15 11:19:08.828382] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val=0x1 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val=dif_generate_copy 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@23 -- # accel_opc=dif_generate_copy 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:25.371 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val=software 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@22 -- # accel_module=software 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val=32 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val=32 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val=1 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val='1 seconds' 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val=No 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:25.372 11:19:08 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@27 -- # [[ -n software ]] 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@27 -- # [[ -n dif_generate_copy ]] 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:08:26.745 00:08:26.745 real 0m1.505s 00:08:26.745 user 0m0.010s 00:08:26.745 sys 0m0.002s 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:26.745 11:19:10 accel.accel_dif_generate_copy -- common/autotest_common.sh@10 -- # set +x 00:08:26.745 ************************************ 00:08:26.745 END TEST accel_dif_generate_copy 00:08:26.745 ************************************ 00:08:26.745 11:19:10 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:26.745 11:19:10 accel -- accel/accel.sh@115 -- # [[ y == y ]] 00:08:26.745 11:19:10 accel -- accel/accel.sh@116 -- # run_test accel_comp accel_test -t 1 -w compress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:26.745 11:19:10 accel -- common/autotest_common.sh@1099 -- # '[' 8 -le 1 ']' 00:08:26.745 11:19:10 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:26.745 11:19:10 accel -- common/autotest_common.sh@10 -- # set +x 00:08:26.745 ************************************ 00:08:26.745 START TEST accel_comp 00:08:26.745 ************************************ 00:08:26.745 11:19:10 accel.accel_comp -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w compress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:26.745 11:19:10 accel.accel_comp -- accel/accel.sh@16 -- # local accel_opc 00:08:26.745 11:19:10 accel.accel_comp -- accel/accel.sh@17 -- # local accel_module 00:08:26.745 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:26.745 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:26.745 11:19:10 accel.accel_comp -- accel/accel.sh@15 -- # accel_perf -t 1 -w compress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:26.745 11:19:10 accel.accel_comp -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w compress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:26.745 11:19:10 accel.accel_comp -- accel/accel.sh@12 -- # build_accel_config 00:08:26.745 11:19:10 accel.accel_comp -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:26.745 11:19:10 accel.accel_comp -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:26.745 11:19:10 accel.accel_comp -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:26.745 11:19:10 accel.accel_comp -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:26.745 11:19:10 accel.accel_comp -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:26.745 11:19:10 accel.accel_comp -- accel/accel.sh@40 -- # local IFS=, 00:08:26.745 11:19:10 accel.accel_comp -- accel/accel.sh@41 -- # jq -r . 00:08:26.745 [2024-07-15 11:19:10.187725] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:26.745 [2024-07-15 11:19:10.187789] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid838165 ] 00:08:26.745 [2024-07-15 11:19:10.303372] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:27.003 [2024-07-15 11:19:10.409815] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@20 -- # val=0x1 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@20 -- # val=compress 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@23 -- # accel_opc=compress 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@20 -- # val=software 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@22 -- # accel_module=software 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@20 -- # val=32 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@20 -- # val=32 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@20 -- # val=1 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@20 -- # val='1 seconds' 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@20 -- # val=No 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:27.003 11:19:10 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:28.375 11:19:11 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:08:28.375 11:19:11 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.375 11:19:11 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:28.375 11:19:11 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:28.375 11:19:11 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:08:28.375 11:19:11 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.375 11:19:11 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:28.376 11:19:11 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:28.376 11:19:11 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:08:28.376 11:19:11 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.376 11:19:11 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:28.376 11:19:11 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:28.376 11:19:11 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:08:28.376 11:19:11 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.376 11:19:11 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:28.376 11:19:11 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:28.376 11:19:11 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:08:28.376 11:19:11 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.376 11:19:11 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:28.376 11:19:11 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:28.376 11:19:11 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:08:28.376 11:19:11 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.376 11:19:11 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:08:28.376 11:19:11 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:08:28.376 11:19:11 accel.accel_comp -- accel/accel.sh@27 -- # [[ -n software ]] 00:08:28.376 11:19:11 accel.accel_comp -- accel/accel.sh@27 -- # [[ -n compress ]] 00:08:28.376 11:19:11 accel.accel_comp -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:08:28.376 00:08:28.376 real 0m1.506s 00:08:28.376 user 0m0.008s 00:08:28.376 sys 0m0.006s 00:08:28.376 11:19:11 accel.accel_comp -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:28.376 11:19:11 accel.accel_comp -- common/autotest_common.sh@10 -- # set +x 00:08:28.376 ************************************ 00:08:28.376 END TEST accel_comp 00:08:28.376 ************************************ 00:08:28.376 11:19:11 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:28.376 11:19:11 accel -- accel/accel.sh@117 -- # run_test accel_decomp accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y 00:08:28.376 11:19:11 accel -- common/autotest_common.sh@1099 -- # '[' 9 -le 1 ']' 00:08:28.376 11:19:11 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:28.376 11:19:11 accel -- common/autotest_common.sh@10 -- # set +x 00:08:28.376 ************************************ 00:08:28.376 START TEST accel_decomp 00:08:28.376 ************************************ 00:08:28.376 11:19:11 accel.accel_decomp -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y 00:08:28.376 11:19:11 accel.accel_decomp -- accel/accel.sh@16 -- # local accel_opc 00:08:28.376 11:19:11 accel.accel_decomp -- accel/accel.sh@17 -- # local accel_module 00:08:28.376 11:19:11 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:28.376 11:19:11 accel.accel_decomp -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y 00:08:28.376 11:19:11 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:28.376 11:19:11 accel.accel_decomp -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y 00:08:28.376 11:19:11 accel.accel_decomp -- accel/accel.sh@12 -- # build_accel_config 00:08:28.376 11:19:11 accel.accel_decomp -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:28.376 11:19:11 accel.accel_decomp -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:28.376 11:19:11 accel.accel_decomp -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:28.376 11:19:11 accel.accel_decomp -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:28.376 11:19:11 accel.accel_decomp -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:28.376 11:19:11 accel.accel_decomp -- accel/accel.sh@40 -- # local IFS=, 00:08:28.376 11:19:11 accel.accel_decomp -- accel/accel.sh@41 -- # jq -r . 00:08:28.376 [2024-07-15 11:19:11.733112] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:28.376 [2024-07-15 11:19:11.733156] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid838359 ] 00:08:28.376 [2024-07-15 11:19:11.846392] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:28.376 [2024-07-15 11:19:11.944700] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:28.633 11:19:12 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:08:28.633 11:19:12 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.633 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@20 -- # val=0x1 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@20 -- # val=decompress 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@23 -- # accel_opc=decompress 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@20 -- # val=software 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@22 -- # accel_module=software 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@20 -- # val=32 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@20 -- # val=32 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@20 -- # val=1 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@20 -- # val='1 seconds' 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@20 -- # val=Yes 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:28.634 11:19:12 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@27 -- # [[ -n software ]] 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:08:30.005 11:19:13 accel.accel_decomp -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:08:30.005 00:08:30.005 real 0m1.466s 00:08:30.005 user 0m0.008s 00:08:30.005 sys 0m0.004s 00:08:30.005 11:19:13 accel.accel_decomp -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:30.005 11:19:13 accel.accel_decomp -- common/autotest_common.sh@10 -- # set +x 00:08:30.005 ************************************ 00:08:30.005 END TEST accel_decomp 00:08:30.005 ************************************ 00:08:30.005 11:19:13 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:30.005 11:19:13 accel -- accel/accel.sh@118 -- # run_test accel_decomp_full accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 00:08:30.005 11:19:13 accel -- common/autotest_common.sh@1099 -- # '[' 11 -le 1 ']' 00:08:30.005 11:19:13 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:30.005 11:19:13 accel -- common/autotest_common.sh@10 -- # set +x 00:08:30.005 ************************************ 00:08:30.005 START TEST accel_decomp_full 00:08:30.005 ************************************ 00:08:30.005 11:19:13 accel.accel_decomp_full -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 00:08:30.005 11:19:13 accel.accel_decomp_full -- accel/accel.sh@16 -- # local accel_opc 00:08:30.005 11:19:13 accel.accel_decomp_full -- accel/accel.sh@17 -- # local accel_module 00:08:30.005 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:30.005 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:30.005 11:19:13 accel.accel_decomp_full -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 00:08:30.005 11:19:13 accel.accel_decomp_full -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 00:08:30.005 11:19:13 accel.accel_decomp_full -- accel/accel.sh@12 -- # build_accel_config 00:08:30.005 11:19:13 accel.accel_decomp_full -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:30.005 11:19:13 accel.accel_decomp_full -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:30.005 11:19:13 accel.accel_decomp_full -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:30.005 11:19:13 accel.accel_decomp_full -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:30.005 11:19:13 accel.accel_decomp_full -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:30.005 11:19:13 accel.accel_decomp_full -- accel/accel.sh@40 -- # local IFS=, 00:08:30.005 11:19:13 accel.accel_decomp_full -- accel/accel.sh@41 -- # jq -r . 00:08:30.005 [2024-07-15 11:19:13.278746] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:30.005 [2024-07-15 11:19:13.278804] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid838563 ] 00:08:30.005 [2024-07-15 11:19:13.407077] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:30.005 [2024-07-15 11:19:13.504522] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:30.005 11:19:13 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@20 -- # val=0x1 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@20 -- # val=decompress 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@23 -- # accel_opc=decompress 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@20 -- # val='111250 bytes' 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@20 -- # val=software 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@22 -- # accel_module=software 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@20 -- # val=32 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@20 -- # val=32 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@20 -- # val=1 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@20 -- # val='1 seconds' 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@20 -- # val=Yes 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:30.006 11:19:13 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@27 -- # [[ -n software ]] 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:08:31.458 11:19:14 accel.accel_decomp_full -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:08:31.458 00:08:31.458 real 0m1.487s 00:08:31.458 user 0m0.010s 00:08:31.458 sys 0m0.003s 00:08:31.458 11:19:14 accel.accel_decomp_full -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:31.458 11:19:14 accel.accel_decomp_full -- common/autotest_common.sh@10 -- # set +x 00:08:31.458 ************************************ 00:08:31.458 END TEST accel_decomp_full 00:08:31.458 ************************************ 00:08:31.458 11:19:14 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:31.458 11:19:14 accel -- accel/accel.sh@119 -- # run_test accel_decomp_mcore accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -m 0xf 00:08:31.458 11:19:14 accel -- common/autotest_common.sh@1099 -- # '[' 11 -le 1 ']' 00:08:31.458 11:19:14 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:31.458 11:19:14 accel -- common/autotest_common.sh@10 -- # set +x 00:08:31.458 ************************************ 00:08:31.458 START TEST accel_decomp_mcore 00:08:31.458 ************************************ 00:08:31.458 11:19:14 accel.accel_decomp_mcore -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -m 0xf 00:08:31.458 11:19:14 accel.accel_decomp_mcore -- accel/accel.sh@16 -- # local accel_opc 00:08:31.458 11:19:14 accel.accel_decomp_mcore -- accel/accel.sh@17 -- # local accel_module 00:08:31.458 11:19:14 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:31.458 11:19:14 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:31.458 11:19:14 accel.accel_decomp_mcore -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -m 0xf 00:08:31.458 11:19:14 accel.accel_decomp_mcore -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -m 0xf 00:08:31.458 11:19:14 accel.accel_decomp_mcore -- accel/accel.sh@12 -- # build_accel_config 00:08:31.458 11:19:14 accel.accel_decomp_mcore -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:31.458 11:19:14 accel.accel_decomp_mcore -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:31.458 11:19:14 accel.accel_decomp_mcore -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:31.458 11:19:14 accel.accel_decomp_mcore -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:31.458 11:19:14 accel.accel_decomp_mcore -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:31.458 11:19:14 accel.accel_decomp_mcore -- accel/accel.sh@40 -- # local IFS=, 00:08:31.458 11:19:14 accel.accel_decomp_mcore -- accel/accel.sh@41 -- # jq -r . 00:08:31.458 [2024-07-15 11:19:14.853406] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:31.458 [2024-07-15 11:19:14.853466] [ DPDK EAL parameters: accel_perf --no-shconf -c 0xf --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid838757 ] 00:08:31.458 [2024-07-15 11:19:14.985214] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:08:31.716 [2024-07-15 11:19:15.091831] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:08:31.716 [2024-07-15 11:19:15.091917] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:08:31.716 [2024-07-15 11:19:15.091998] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:08:31.716 [2024-07-15 11:19:15.092003] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val=0xf 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val=decompress 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@23 -- # accel_opc=decompress 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val=software 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@22 -- # accel_module=software 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val=32 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val=32 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val=1 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val='1 seconds' 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val=Yes 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:31.716 11:19:15 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.085 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:33.085 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.085 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.085 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.085 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:33.085 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.085 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.085 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.085 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:33.085 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.085 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.085 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.085 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:33.085 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.085 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.085 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.085 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@27 -- # [[ -n software ]] 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:08:33.086 00:08:33.086 real 0m1.527s 00:08:33.086 user 0m4.773s 00:08:33.086 sys 0m0.206s 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:33.086 11:19:16 accel.accel_decomp_mcore -- common/autotest_common.sh@10 -- # set +x 00:08:33.086 ************************************ 00:08:33.086 END TEST accel_decomp_mcore 00:08:33.086 ************************************ 00:08:33.086 11:19:16 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:33.086 11:19:16 accel -- accel/accel.sh@120 -- # run_test accel_decomp_full_mcore accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 -m 0xf 00:08:33.086 11:19:16 accel -- common/autotest_common.sh@1099 -- # '[' 13 -le 1 ']' 00:08:33.086 11:19:16 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:33.086 11:19:16 accel -- common/autotest_common.sh@10 -- # set +x 00:08:33.086 ************************************ 00:08:33.086 START TEST accel_decomp_full_mcore 00:08:33.086 ************************************ 00:08:33.086 11:19:16 accel.accel_decomp_full_mcore -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 -m 0xf 00:08:33.086 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@16 -- # local accel_opc 00:08:33.086 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@17 -- # local accel_module 00:08:33.086 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.086 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.086 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 -m 0xf 00:08:33.086 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@12 -- # build_accel_config 00:08:33.086 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:33.086 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 -m 0xf 00:08:33.086 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:33.086 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:33.086 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:33.086 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:33.086 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@40 -- # local IFS=, 00:08:33.086 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@41 -- # jq -r . 00:08:33.086 [2024-07-15 11:19:16.460944] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:33.086 [2024-07-15 11:19:16.461005] [ DPDK EAL parameters: accel_perf --no-shconf -c 0xf --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid839016 ] 00:08:33.086 [2024-07-15 11:19:16.589592] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:08:33.344 [2024-07-15 11:19:16.695853] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:08:33.344 [2024-07-15 11:19:16.695948] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:08:33.344 [2024-07-15 11:19:16.696012] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:08:33.344 [2024-07-15 11:19:16.696016] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val=0xf 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val=decompress 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@23 -- # accel_opc=decompress 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val='111250 bytes' 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val=software 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@22 -- # accel_module=software 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val=32 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val=32 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val=1 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val='1 seconds' 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val=Yes 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:33.344 11:19:16 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:34.717 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:34.718 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:34.718 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:34.718 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:34.718 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:34.718 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:34.718 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:34.718 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@27 -- # [[ -n software ]] 00:08:34.718 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:08:34.718 11:19:17 accel.accel_decomp_full_mcore -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:08:34.718 00:08:34.718 real 0m1.535s 00:08:34.718 user 0m4.812s 00:08:34.718 sys 0m0.210s 00:08:34.718 11:19:17 accel.accel_decomp_full_mcore -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:34.718 11:19:17 accel.accel_decomp_full_mcore -- common/autotest_common.sh@10 -- # set +x 00:08:34.718 ************************************ 00:08:34.718 END TEST accel_decomp_full_mcore 00:08:34.718 ************************************ 00:08:34.718 11:19:18 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:34.718 11:19:18 accel -- accel/accel.sh@121 -- # run_test accel_decomp_mthread accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -T 2 00:08:34.718 11:19:18 accel -- common/autotest_common.sh@1099 -- # '[' 11 -le 1 ']' 00:08:34.718 11:19:18 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:34.718 11:19:18 accel -- common/autotest_common.sh@10 -- # set +x 00:08:34.718 ************************************ 00:08:34.718 START TEST accel_decomp_mthread 00:08:34.718 ************************************ 00:08:34.718 11:19:18 accel.accel_decomp_mthread -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -T 2 00:08:34.718 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@16 -- # local accel_opc 00:08:34.718 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@17 -- # local accel_module 00:08:34.718 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:34.718 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:34.718 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -T 2 00:08:34.718 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -T 2 00:08:34.718 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@12 -- # build_accel_config 00:08:34.718 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:34.718 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:34.718 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:34.718 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:34.718 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:34.718 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@40 -- # local IFS=, 00:08:34.718 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@41 -- # jq -r . 00:08:34.718 [2024-07-15 11:19:18.077397] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:34.718 [2024-07-15 11:19:18.077459] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid839319 ] 00:08:34.718 [2024-07-15 11:19:18.205281] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:34.718 [2024-07-15 11:19:18.305864] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val=0x1 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val=decompress 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@23 -- # accel_opc=decompress 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val=software 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@22 -- # accel_module=software 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val=32 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val=32 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val=2 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val='1 seconds' 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val=Yes 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:34.976 11:19:18 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@27 -- # [[ -n software ]] 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:08:36.349 00:08:36.349 real 0m1.510s 00:08:36.349 user 0m1.309s 00:08:36.349 sys 0m0.209s 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:36.349 11:19:19 accel.accel_decomp_mthread -- common/autotest_common.sh@10 -- # set +x 00:08:36.349 ************************************ 00:08:36.349 END TEST accel_decomp_mthread 00:08:36.349 ************************************ 00:08:36.349 11:19:19 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:36.349 11:19:19 accel -- accel/accel.sh@122 -- # run_test accel_decomp_full_mthread accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 -T 2 00:08:36.349 11:19:19 accel -- common/autotest_common.sh@1099 -- # '[' 13 -le 1 ']' 00:08:36.349 11:19:19 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:36.349 11:19:19 accel -- common/autotest_common.sh@10 -- # set +x 00:08:36.349 ************************************ 00:08:36.349 START TEST accel_decomp_full_mthread 00:08:36.349 ************************************ 00:08:36.349 11:19:19 accel.accel_decomp_full_mthread -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 -T 2 00:08:36.349 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@16 -- # local accel_opc 00:08:36.349 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@17 -- # local accel_module 00:08:36.349 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.349 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.349 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 -T 2 00:08:36.349 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 -T 2 00:08:36.349 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@12 -- # build_accel_config 00:08:36.349 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:36.349 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:36.349 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:36.349 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:36.349 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:36.349 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@40 -- # local IFS=, 00:08:36.349 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@41 -- # jq -r . 00:08:36.349 [2024-07-15 11:19:19.666476] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:36.350 [2024-07-15 11:19:19.666536] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid839515 ] 00:08:36.350 [2024-07-15 11:19:19.795543] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:36.350 [2024-07-15 11:19:19.893712] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:36.607 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:36.607 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.607 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val=0x1 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val=decompress 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@23 -- # accel_opc=decompress 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val='111250 bytes' 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val=software 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@22 -- # accel_module=software 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val=32 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val=32 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val=2 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val='1 seconds' 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val=Yes 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:36.608 11:19:19 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@27 -- # [[ -n software ]] 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:08:37.980 00:08:37.980 real 0m1.538s 00:08:37.980 user 0m1.357s 00:08:37.980 sys 0m0.182s 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:37.980 11:19:21 accel.accel_decomp_full_mthread -- common/autotest_common.sh@10 -- # set +x 00:08:37.980 ************************************ 00:08:37.980 END TEST accel_decomp_full_mthread 00:08:37.980 ************************************ 00:08:37.980 11:19:21 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:37.980 11:19:21 accel -- accel/accel.sh@124 -- # [[ y == y ]] 00:08:37.980 11:19:21 accel -- accel/accel.sh@125 -- # COMPRESSDEV=1 00:08:37.980 11:19:21 accel -- accel/accel.sh@126 -- # get_expected_opcs 00:08:37.980 11:19:21 accel -- accel/accel.sh@60 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:08:37.980 11:19:21 accel -- accel/accel.sh@62 -- # spdk_tgt_pid=839713 00:08:37.980 11:19:21 accel -- accel/accel.sh@63 -- # waitforlisten 839713 00:08:37.980 11:19:21 accel -- common/autotest_common.sh@829 -- # '[' -z 839713 ']' 00:08:37.980 11:19:21 accel -- accel/accel.sh@61 -- # build_accel_config 00:08:37.980 11:19:21 accel -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:37.980 11:19:21 accel -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:37.980 11:19:21 accel -- accel/accel.sh@61 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt -c /dev/fd/63 00:08:37.980 11:19:21 accel -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:37.980 11:19:21 accel -- common/autotest_common.sh@834 -- # local max_retries=100 00:08:37.980 11:19:21 accel -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:37.980 11:19:21 accel -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:37.980 11:19:21 accel -- accel/accel.sh@36 -- # [[ -n 1 ]] 00:08:37.980 11:19:21 accel -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:37.980 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:37.980 11:19:21 accel -- accel/accel.sh@37 -- # accel_json_cfg+=('{"method": "compressdev_scan_accel_module", "params":{"pmd": 0}}') 00:08:37.980 11:19:21 accel -- accel/accel.sh@40 -- # local IFS=, 00:08:37.980 11:19:21 accel -- common/autotest_common.sh@838 -- # xtrace_disable 00:08:37.981 11:19:21 accel -- accel/accel.sh@41 -- # jq -r . 00:08:37.981 11:19:21 accel -- common/autotest_common.sh@10 -- # set +x 00:08:37.981 [2024-07-15 11:19:21.285164] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:37.981 [2024-07-15 11:19:21.285235] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid839713 ] 00:08:37.981 [2024-07-15 11:19:21.414968] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:37.981 [2024-07-15 11:19:21.519579] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:38.913 [2024-07-15 11:19:22.291016] accel_dpdk_compressdev.c: 296:accel_init_compress_drivers: *NOTICE*: initialized QAT PMD 00:08:38.913 11:19:22 accel -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:08:38.913 11:19:22 accel -- common/autotest_common.sh@862 -- # return 0 00:08:38.913 11:19:22 accel -- accel/accel.sh@65 -- # [[ 0 -gt 0 ]] 00:08:38.913 11:19:22 accel -- accel/accel.sh@66 -- # [[ 0 -gt 0 ]] 00:08:38.913 11:19:22 accel -- accel/accel.sh@67 -- # [[ 0 -gt 0 ]] 00:08:38.913 11:19:22 accel -- accel/accel.sh@68 -- # [[ -n 1 ]] 00:08:38.913 11:19:22 accel -- accel/accel.sh@68 -- # check_save_config compressdev_scan_accel_module 00:08:38.913 11:19:22 accel -- accel/accel.sh@56 -- # rpc_cmd save_config 00:08:38.913 11:19:22 accel -- accel/accel.sh@56 -- # jq -r '.subsystems[] | select(.subsystem=="accel").config[]' 00:08:38.913 11:19:22 accel -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:38.913 11:19:22 accel -- common/autotest_common.sh@10 -- # set +x 00:08:38.913 11:19:22 accel -- accel/accel.sh@56 -- # grep compressdev_scan_accel_module 00:08:39.170 11:19:22 accel -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:39.170 "method": "compressdev_scan_accel_module", 00:08:39.170 11:19:22 accel -- accel/accel.sh@70 -- # exp_opcs=($($rpc_py accel_get_opc_assignments | jq -r ". | to_entries | map(\"\(.key)=\(.value)\") | .[]")) 00:08:39.170 11:19:22 accel -- accel/accel.sh@70 -- # rpc_cmd accel_get_opc_assignments 00:08:39.171 11:19:22 accel -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:39.171 11:19:22 accel -- accel/accel.sh@70 -- # jq -r '. | to_entries | map("\(.key)=\(.value)") | .[]' 00:08:39.171 11:19:22 accel -- common/autotest_common.sh@10 -- # set +x 00:08:39.171 11:19:22 accel -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:39.171 11:19:22 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # IFS== 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:39.171 11:19:22 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:39.171 11:19:22 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # IFS== 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:39.171 11:19:22 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:39.171 11:19:22 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # IFS== 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:39.171 11:19:22 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:39.171 11:19:22 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # IFS== 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:39.171 11:19:22 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:39.171 11:19:22 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # IFS== 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:39.171 11:19:22 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:39.171 11:19:22 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # IFS== 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:39.171 11:19:22 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:39.171 11:19:22 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # IFS== 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:39.171 11:19:22 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=dpdk_compressdev 00:08:39.171 11:19:22 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # IFS== 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:39.171 11:19:22 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=dpdk_compressdev 00:08:39.171 11:19:22 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # IFS== 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:39.171 11:19:22 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:39.171 11:19:22 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # IFS== 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:39.171 11:19:22 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:39.171 11:19:22 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # IFS== 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:39.171 11:19:22 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:39.171 11:19:22 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # IFS== 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:39.171 11:19:22 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:39.171 11:19:22 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # IFS== 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:39.171 11:19:22 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:39.171 11:19:22 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # IFS== 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:39.171 11:19:22 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:39.171 11:19:22 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # IFS== 00:08:39.171 11:19:22 accel -- accel/accel.sh@72 -- # read -r opc module 00:08:39.171 11:19:22 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:08:39.171 11:19:22 accel -- accel/accel.sh@75 -- # killprocess 839713 00:08:39.171 11:19:22 accel -- common/autotest_common.sh@948 -- # '[' -z 839713 ']' 00:08:39.171 11:19:22 accel -- common/autotest_common.sh@952 -- # kill -0 839713 00:08:39.171 11:19:22 accel -- common/autotest_common.sh@953 -- # uname 00:08:39.171 11:19:22 accel -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:08:39.171 11:19:22 accel -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 839713 00:08:39.429 11:19:22 accel -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:08:39.429 11:19:22 accel -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:08:39.429 11:19:22 accel -- common/autotest_common.sh@966 -- # echo 'killing process with pid 839713' 00:08:39.429 killing process with pid 839713 00:08:39.429 11:19:22 accel -- common/autotest_common.sh@967 -- # kill 839713 00:08:39.429 11:19:22 accel -- common/autotest_common.sh@972 -- # wait 839713 00:08:39.687 11:19:23 accel -- accel/accel.sh@76 -- # trap - ERR 00:08:39.687 11:19:23 accel -- accel/accel.sh@127 -- # run_test accel_cdev_comp accel_test -t 1 -w compress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:39.687 11:19:23 accel -- common/autotest_common.sh@1099 -- # '[' 8 -le 1 ']' 00:08:39.687 11:19:23 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:39.687 11:19:23 accel -- common/autotest_common.sh@10 -- # set +x 00:08:39.687 ************************************ 00:08:39.687 START TEST accel_cdev_comp 00:08:39.687 ************************************ 00:08:39.687 11:19:23 accel.accel_cdev_comp -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w compress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:39.687 11:19:23 accel.accel_cdev_comp -- accel/accel.sh@16 -- # local accel_opc 00:08:39.687 11:19:23 accel.accel_cdev_comp -- accel/accel.sh@17 -- # local accel_module 00:08:39.687 11:19:23 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:39.687 11:19:23 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:39.687 11:19:23 accel.accel_cdev_comp -- accel/accel.sh@15 -- # accel_perf -t 1 -w compress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:39.687 11:19:23 accel.accel_cdev_comp -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w compress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:39.687 11:19:23 accel.accel_cdev_comp -- accel/accel.sh@12 -- # build_accel_config 00:08:39.687 11:19:23 accel.accel_cdev_comp -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:39.687 11:19:23 accel.accel_cdev_comp -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:39.687 11:19:23 accel.accel_cdev_comp -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:39.687 11:19:23 accel.accel_cdev_comp -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:39.687 11:19:23 accel.accel_cdev_comp -- accel/accel.sh@36 -- # [[ -n 1 ]] 00:08:39.687 11:19:23 accel.accel_cdev_comp -- accel/accel.sh@37 -- # accel_json_cfg+=('{"method": "compressdev_scan_accel_module", "params":{"pmd": 0}}') 00:08:39.687 11:19:23 accel.accel_cdev_comp -- accel/accel.sh@40 -- # local IFS=, 00:08:39.687 11:19:23 accel.accel_cdev_comp -- accel/accel.sh@41 -- # jq -r . 00:08:39.687 [2024-07-15 11:19:23.237318] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:39.687 [2024-07-15 11:19:23.237384] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid839959 ] 00:08:39.945 [2024-07-15 11:19:23.367530] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:39.945 [2024-07-15 11:19:23.469208] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:40.877 [2024-07-15 11:19:24.231834] accel_dpdk_compressdev.c: 296:accel_init_compress_drivers: *NOTICE*: initialized QAT PMD 00:08:40.877 [2024-07-15 11:19:24.234457] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x20cc080 PMD being used: compress_qat 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val= 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val= 00:08:40.877 [2024-07-15 11:19:24.238578] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x20d0e60 PMD being used: compress_qat 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val= 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val=0x1 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val= 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val= 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val=compress 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@23 -- # accel_opc=compress 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val= 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val=dpdk_compressdev 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@22 -- # accel_module=dpdk_compressdev 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val=32 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val=32 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val=1 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val='1 seconds' 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val=No 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val= 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val= 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:40.877 11:19:24 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:41.809 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val= 00:08:41.809 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:41.809 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:41.809 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:41.809 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val= 00:08:41.809 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:41.809 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:41.809 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:41.809 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val= 00:08:41.809 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:41.809 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:42.089 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:42.089 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val= 00:08:42.089 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:42.089 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:42.089 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:42.089 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val= 00:08:42.089 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:42.089 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:42.089 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:42.089 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@20 -- # val= 00:08:42.089 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@21 -- # case "$var" in 00:08:42.089 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@19 -- # IFS=: 00:08:42.089 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@19 -- # read -r var val 00:08:42.089 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@27 -- # [[ -n dpdk_compressdev ]] 00:08:42.089 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@27 -- # [[ -n compress ]] 00:08:42.089 11:19:25 accel.accel_cdev_comp -- accel/accel.sh@27 -- # [[ dpdk_compressdev == \d\p\d\k\_\c\o\m\p\r\e\s\s\d\e\v ]] 00:08:42.089 00:08:42.089 real 0m2.204s 00:08:42.089 user 0m1.650s 00:08:42.089 sys 0m0.559s 00:08:42.089 11:19:25 accel.accel_cdev_comp -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:42.089 11:19:25 accel.accel_cdev_comp -- common/autotest_common.sh@10 -- # set +x 00:08:42.089 ************************************ 00:08:42.089 END TEST accel_cdev_comp 00:08:42.089 ************************************ 00:08:42.089 11:19:25 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:42.089 11:19:25 accel -- accel/accel.sh@128 -- # run_test accel_cdev_decomp accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y 00:08:42.089 11:19:25 accel -- common/autotest_common.sh@1099 -- # '[' 9 -le 1 ']' 00:08:42.089 11:19:25 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:42.089 11:19:25 accel -- common/autotest_common.sh@10 -- # set +x 00:08:42.089 ************************************ 00:08:42.089 START TEST accel_cdev_decomp 00:08:42.089 ************************************ 00:08:42.089 11:19:25 accel.accel_cdev_decomp -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y 00:08:42.089 11:19:25 accel.accel_cdev_decomp -- accel/accel.sh@16 -- # local accel_opc 00:08:42.089 11:19:25 accel.accel_cdev_decomp -- accel/accel.sh@17 -- # local accel_module 00:08:42.089 11:19:25 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:42.089 11:19:25 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:42.089 11:19:25 accel.accel_cdev_decomp -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y 00:08:42.089 11:19:25 accel.accel_cdev_decomp -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y 00:08:42.089 11:19:25 accel.accel_cdev_decomp -- accel/accel.sh@12 -- # build_accel_config 00:08:42.089 11:19:25 accel.accel_cdev_decomp -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:42.089 11:19:25 accel.accel_cdev_decomp -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:42.089 11:19:25 accel.accel_cdev_decomp -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:42.089 11:19:25 accel.accel_cdev_decomp -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:42.089 11:19:25 accel.accel_cdev_decomp -- accel/accel.sh@36 -- # [[ -n 1 ]] 00:08:42.089 11:19:25 accel.accel_cdev_decomp -- accel/accel.sh@37 -- # accel_json_cfg+=('{"method": "compressdev_scan_accel_module", "params":{"pmd": 0}}') 00:08:42.089 11:19:25 accel.accel_cdev_decomp -- accel/accel.sh@40 -- # local IFS=, 00:08:42.089 11:19:25 accel.accel_cdev_decomp -- accel/accel.sh@41 -- # jq -r . 00:08:42.089 [2024-07-15 11:19:25.522688] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:42.089 [2024-07-15 11:19:25.522748] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid840281 ] 00:08:42.089 [2024-07-15 11:19:25.652411] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:42.346 [2024-07-15 11:19:25.754036] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:43.281 [2024-07-15 11:19:26.522035] accel_dpdk_compressdev.c: 296:accel_init_compress_drivers: *NOTICE*: initialized QAT PMD 00:08:43.281 [2024-07-15 11:19:26.524574] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x1181080 PMD being used: compress_qat 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val= 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val= 00:08:43.281 [2024-07-15 11:19:26.528676] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x1185e60 PMD being used: compress_qat 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val= 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val=0x1 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val= 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val= 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val=decompress 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@23 -- # accel_opc=decompress 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val= 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val=dpdk_compressdev 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@22 -- # accel_module=dpdk_compressdev 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val=32 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val=32 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val=1 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val='1 seconds' 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val=Yes 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val= 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val= 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:43.281 11:19:26 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val= 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val= 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val= 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val= 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val= 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@20 -- # val= 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@21 -- # case "$var" in 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # IFS=: 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@19 -- # read -r var val 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@27 -- # [[ -n dpdk_compressdev ]] 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- accel/accel.sh@27 -- # [[ dpdk_compressdev == \d\p\d\k\_\c\o\m\p\r\e\s\s\d\e\v ]] 00:08:44.215 00:08:44.215 real 0m2.223s 00:08:44.215 user 0m1.633s 00:08:44.215 sys 0m0.584s 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:44.215 11:19:27 accel.accel_cdev_decomp -- common/autotest_common.sh@10 -- # set +x 00:08:44.215 ************************************ 00:08:44.215 END TEST accel_cdev_decomp 00:08:44.215 ************************************ 00:08:44.215 11:19:27 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:44.215 11:19:27 accel -- accel/accel.sh@129 -- # run_test accel_cdev_decomp_full accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 00:08:44.215 11:19:27 accel -- common/autotest_common.sh@1099 -- # '[' 11 -le 1 ']' 00:08:44.215 11:19:27 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:44.215 11:19:27 accel -- common/autotest_common.sh@10 -- # set +x 00:08:44.215 ************************************ 00:08:44.215 START TEST accel_cdev_decomp_full 00:08:44.215 ************************************ 00:08:44.215 11:19:27 accel.accel_cdev_decomp_full -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 00:08:44.215 11:19:27 accel.accel_cdev_decomp_full -- accel/accel.sh@16 -- # local accel_opc 00:08:44.215 11:19:27 accel.accel_cdev_decomp_full -- accel/accel.sh@17 -- # local accel_module 00:08:44.215 11:19:27 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:44.215 11:19:27 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:44.215 11:19:27 accel.accel_cdev_decomp_full -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 00:08:44.215 11:19:27 accel.accel_cdev_decomp_full -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 00:08:44.215 11:19:27 accel.accel_cdev_decomp_full -- accel/accel.sh@12 -- # build_accel_config 00:08:44.215 11:19:27 accel.accel_cdev_decomp_full -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:44.215 11:19:27 accel.accel_cdev_decomp_full -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:44.215 11:19:27 accel.accel_cdev_decomp_full -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:44.215 11:19:27 accel.accel_cdev_decomp_full -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:44.215 11:19:27 accel.accel_cdev_decomp_full -- accel/accel.sh@36 -- # [[ -n 1 ]] 00:08:44.216 11:19:27 accel.accel_cdev_decomp_full -- accel/accel.sh@37 -- # accel_json_cfg+=('{"method": "compressdev_scan_accel_module", "params":{"pmd": 0}}') 00:08:44.216 11:19:27 accel.accel_cdev_decomp_full -- accel/accel.sh@40 -- # local IFS=, 00:08:44.216 11:19:27 accel.accel_cdev_decomp_full -- accel/accel.sh@41 -- # jq -r . 00:08:44.474 [2024-07-15 11:19:27.820648] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:44.474 [2024-07-15 11:19:27.820707] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid840644 ] 00:08:44.474 [2024-07-15 11:19:27.946310] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:44.474 [2024-07-15 11:19:28.045086] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:45.406 [2024-07-15 11:19:28.814587] accel_dpdk_compressdev.c: 296:accel_init_compress_drivers: *NOTICE*: initialized QAT PMD 00:08:45.406 [2024-07-15 11:19:28.817202] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x26e1080 PMD being used: compress_qat 00:08:45.406 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val= 00:08:45.406 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:45.406 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:45.406 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:45.406 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val= 00:08:45.406 [2024-07-15 11:19:28.820581] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x26e0ce0 PMD being used: compress_qat 00:08:45.406 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:45.406 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:45.406 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:45.406 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val= 00:08:45.406 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val=0x1 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val= 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val= 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val=decompress 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@23 -- # accel_opc=decompress 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val='111250 bytes' 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val= 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val=dpdk_compressdev 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@22 -- # accel_module=dpdk_compressdev 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val=32 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val=32 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val=1 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val='1 seconds' 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val=Yes 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val= 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val= 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:45.407 11:19:28 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:46.776 11:19:29 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val= 00:08:46.776 11:19:29 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:46.776 11:19:29 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:46.776 11:19:29 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:46.776 11:19:29 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val= 00:08:46.776 11:19:29 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:46.776 11:19:29 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:46.776 11:19:29 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:46.776 11:19:29 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val= 00:08:46.776 11:19:29 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:46.776 11:19:29 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:46.776 11:19:29 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:46.776 11:19:30 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val= 00:08:46.776 11:19:30 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:46.776 11:19:30 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:46.776 11:19:30 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:46.776 11:19:30 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val= 00:08:46.776 11:19:30 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:46.776 11:19:30 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:46.776 11:19:30 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:46.776 11:19:30 accel.accel_cdev_decomp_full -- accel/accel.sh@20 -- # val= 00:08:46.776 11:19:30 accel.accel_cdev_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:08:46.776 11:19:30 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:08:46.776 11:19:30 accel.accel_cdev_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:08:46.776 11:19:30 accel.accel_cdev_decomp_full -- accel/accel.sh@27 -- # [[ -n dpdk_compressdev ]] 00:08:46.776 11:19:30 accel.accel_cdev_decomp_full -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:08:46.776 11:19:30 accel.accel_cdev_decomp_full -- accel/accel.sh@27 -- # [[ dpdk_compressdev == \d\p\d\k\_\c\o\m\p\r\e\s\s\d\e\v ]] 00:08:46.776 00:08:46.776 real 0m2.221s 00:08:46.776 user 0m1.671s 00:08:46.776 sys 0m0.555s 00:08:46.776 11:19:30 accel.accel_cdev_decomp_full -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:46.776 11:19:30 accel.accel_cdev_decomp_full -- common/autotest_common.sh@10 -- # set +x 00:08:46.776 ************************************ 00:08:46.776 END TEST accel_cdev_decomp_full 00:08:46.776 ************************************ 00:08:46.776 11:19:30 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:46.776 11:19:30 accel -- accel/accel.sh@130 -- # run_test accel_cdev_decomp_mcore accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -m 0xf 00:08:46.776 11:19:30 accel -- common/autotest_common.sh@1099 -- # '[' 11 -le 1 ']' 00:08:46.776 11:19:30 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:46.776 11:19:30 accel -- common/autotest_common.sh@10 -- # set +x 00:08:46.776 ************************************ 00:08:46.776 START TEST accel_cdev_decomp_mcore 00:08:46.776 ************************************ 00:08:46.776 11:19:30 accel.accel_cdev_decomp_mcore -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -m 0xf 00:08:46.776 11:19:30 accel.accel_cdev_decomp_mcore -- accel/accel.sh@16 -- # local accel_opc 00:08:46.776 11:19:30 accel.accel_cdev_decomp_mcore -- accel/accel.sh@17 -- # local accel_module 00:08:46.776 11:19:30 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:46.776 11:19:30 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:46.776 11:19:30 accel.accel_cdev_decomp_mcore -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -m 0xf 00:08:46.776 11:19:30 accel.accel_cdev_decomp_mcore -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -m 0xf 00:08:46.776 11:19:30 accel.accel_cdev_decomp_mcore -- accel/accel.sh@12 -- # build_accel_config 00:08:46.776 11:19:30 accel.accel_cdev_decomp_mcore -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:46.776 11:19:30 accel.accel_cdev_decomp_mcore -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:46.776 11:19:30 accel.accel_cdev_decomp_mcore -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:46.776 11:19:30 accel.accel_cdev_decomp_mcore -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:46.776 11:19:30 accel.accel_cdev_decomp_mcore -- accel/accel.sh@36 -- # [[ -n 1 ]] 00:08:46.776 11:19:30 accel.accel_cdev_decomp_mcore -- accel/accel.sh@37 -- # accel_json_cfg+=('{"method": "compressdev_scan_accel_module", "params":{"pmd": 0}}') 00:08:46.776 11:19:30 accel.accel_cdev_decomp_mcore -- accel/accel.sh@40 -- # local IFS=, 00:08:46.776 11:19:30 accel.accel_cdev_decomp_mcore -- accel/accel.sh@41 -- # jq -r . 00:08:46.776 [2024-07-15 11:19:30.130678] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:46.776 [2024-07-15 11:19:30.130746] [ DPDK EAL parameters: accel_perf --no-shconf -c 0xf --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid840942 ] 00:08:46.776 [2024-07-15 11:19:30.263068] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:08:47.034 [2024-07-15 11:19:30.372249] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:08:47.034 [2024-07-15 11:19:30.372336] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:08:47.034 [2024-07-15 11:19:30.372363] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:08:47.034 [2024-07-15 11:19:30.372367] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:47.631 [2024-07-15 11:19:31.130354] accel_dpdk_compressdev.c: 296:accel_init_compress_drivers: *NOTICE*: initialized QAT PMD 00:08:47.631 [2024-07-15 11:19:31.132946] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x17ee720 PMD being used: compress_qat 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val=0xf 00:08:47.631 [2024-07-15 11:19:31.138639] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x7fb66819b8b0 PMD being used: compress_qat 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:47.631 [2024-07-15 11:19:31.139434] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x7fb66019b8b0 PMD being used: compress_qat 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:47.631 [2024-07-15 11:19:31.140444] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x17f39f0 PMD being used: compress_qat 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:47.631 [2024-07-15 11:19:31.140644] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x7fb65819b8b0 PMD being used: compress_qat 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val=decompress 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@23 -- # accel_opc=decompress 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val=dpdk_compressdev 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@22 -- # accel_module=dpdk_compressdev 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val=32 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val=32 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val=1 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val='1 seconds' 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val=Yes 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:47.631 11:19:31 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@20 -- # val= 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:49.000 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:49.001 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@27 -- # [[ -n dpdk_compressdev ]] 00:08:49.001 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:08:49.001 11:19:32 accel.accel_cdev_decomp_mcore -- accel/accel.sh@27 -- # [[ dpdk_compressdev == \d\p\d\k\_\c\o\m\p\r\e\s\s\d\e\v ]] 00:08:49.001 00:08:49.001 real 0m2.247s 00:08:49.001 user 0m7.241s 00:08:49.001 sys 0m0.587s 00:08:49.001 11:19:32 accel.accel_cdev_decomp_mcore -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:49.001 11:19:32 accel.accel_cdev_decomp_mcore -- common/autotest_common.sh@10 -- # set +x 00:08:49.001 ************************************ 00:08:49.001 END TEST accel_cdev_decomp_mcore 00:08:49.001 ************************************ 00:08:49.001 11:19:32 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:49.001 11:19:32 accel -- accel/accel.sh@131 -- # run_test accel_cdev_decomp_full_mcore accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 -m 0xf 00:08:49.001 11:19:32 accel -- common/autotest_common.sh@1099 -- # '[' 13 -le 1 ']' 00:08:49.001 11:19:32 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:49.001 11:19:32 accel -- common/autotest_common.sh@10 -- # set +x 00:08:49.001 ************************************ 00:08:49.001 START TEST accel_cdev_decomp_full_mcore 00:08:49.001 ************************************ 00:08:49.001 11:19:32 accel.accel_cdev_decomp_full_mcore -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 -m 0xf 00:08:49.001 11:19:32 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@16 -- # local accel_opc 00:08:49.001 11:19:32 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@17 -- # local accel_module 00:08:49.001 11:19:32 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:49.001 11:19:32 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:49.001 11:19:32 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 -m 0xf 00:08:49.001 11:19:32 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@12 -- # build_accel_config 00:08:49.001 11:19:32 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 -m 0xf 00:08:49.001 11:19:32 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:49.001 11:19:32 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:49.001 11:19:32 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:49.001 11:19:32 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:49.001 11:19:32 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@36 -- # [[ -n 1 ]] 00:08:49.001 11:19:32 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@37 -- # accel_json_cfg+=('{"method": "compressdev_scan_accel_module", "params":{"pmd": 0}}') 00:08:49.001 11:19:32 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@40 -- # local IFS=, 00:08:49.001 11:19:32 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@41 -- # jq -r . 00:08:49.001 [2024-07-15 11:19:32.450491] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:49.001 [2024-07-15 11:19:32.450548] [ DPDK EAL parameters: accel_perf --no-shconf -c 0xf --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid841223 ] 00:08:49.001 [2024-07-15 11:19:32.577866] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:08:49.258 [2024-07-15 11:19:32.683955] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:08:49.258 [2024-07-15 11:19:32.684040] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:08:49.258 [2024-07-15 11:19:32.684118] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:08:49.258 [2024-07-15 11:19:32.684123] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:50.187 [2024-07-15 11:19:33.434816] accel_dpdk_compressdev.c: 296:accel_init_compress_drivers: *NOTICE*: initialized QAT PMD 00:08:50.187 [2024-07-15 11:19:33.437448] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x1db6720 PMD being used: compress_qat 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val=0xf 00:08:50.187 [2024-07-15 11:19:33.442223] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x7ff6e019b8b0 PMD being used: compress_qat 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:50.187 [2024-07-15 11:19:33.442947] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x7ff6d819b8b0 PMD being used: compress_qat 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:50.187 [2024-07-15 11:19:33.444157] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x1db9a30 PMD being used: compress_qat 00:08:50.187 [2024-07-15 11:19:33.444334] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x7ff6d019b8b0 PMD being used: compress_qat 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val=decompress 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@23 -- # accel_opc=decompress 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val='111250 bytes' 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:50.187 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val=dpdk_compressdev 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@22 -- # accel_module=dpdk_compressdev 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val=32 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val=32 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val=1 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val='1 seconds' 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val=Yes 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:50.188 11:19:33 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@27 -- # [[ -n dpdk_compressdev ]] 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- accel/accel.sh@27 -- # [[ dpdk_compressdev == \d\p\d\k\_\c\o\m\p\r\e\s\s\d\e\v ]] 00:08:51.119 00:08:51.119 real 0m2.210s 00:08:51.119 user 0m7.135s 00:08:51.119 sys 0m0.595s 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:51.119 11:19:34 accel.accel_cdev_decomp_full_mcore -- common/autotest_common.sh@10 -- # set +x 00:08:51.119 ************************************ 00:08:51.119 END TEST accel_cdev_decomp_full_mcore 00:08:51.119 ************************************ 00:08:51.119 11:19:34 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:51.119 11:19:34 accel -- accel/accel.sh@132 -- # run_test accel_cdev_decomp_mthread accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -T 2 00:08:51.119 11:19:34 accel -- common/autotest_common.sh@1099 -- # '[' 11 -le 1 ']' 00:08:51.119 11:19:34 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:51.119 11:19:34 accel -- common/autotest_common.sh@10 -- # set +x 00:08:51.119 ************************************ 00:08:51.119 START TEST accel_cdev_decomp_mthread 00:08:51.119 ************************************ 00:08:51.119 11:19:34 accel.accel_cdev_decomp_mthread -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -T 2 00:08:51.119 11:19:34 accel.accel_cdev_decomp_mthread -- accel/accel.sh@16 -- # local accel_opc 00:08:51.119 11:19:34 accel.accel_cdev_decomp_mthread -- accel/accel.sh@17 -- # local accel_module 00:08:51.119 11:19:34 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:51.119 11:19:34 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:51.119 11:19:34 accel.accel_cdev_decomp_mthread -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -T 2 00:08:51.119 11:19:34 accel.accel_cdev_decomp_mthread -- accel/accel.sh@12 -- # build_accel_config 00:08:51.119 11:19:34 accel.accel_cdev_decomp_mthread -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:51.120 11:19:34 accel.accel_cdev_decomp_mthread -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:51.120 11:19:34 accel.accel_cdev_decomp_mthread -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -T 2 00:08:51.120 11:19:34 accel.accel_cdev_decomp_mthread -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:51.120 11:19:34 accel.accel_cdev_decomp_mthread -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:51.120 11:19:34 accel.accel_cdev_decomp_mthread -- accel/accel.sh@36 -- # [[ -n 1 ]] 00:08:51.120 11:19:34 accel.accel_cdev_decomp_mthread -- accel/accel.sh@37 -- # accel_json_cfg+=('{"method": "compressdev_scan_accel_module", "params":{"pmd": 0}}') 00:08:51.120 11:19:34 accel.accel_cdev_decomp_mthread -- accel/accel.sh@40 -- # local IFS=, 00:08:51.120 11:19:34 accel.accel_cdev_decomp_mthread -- accel/accel.sh@41 -- # jq -r . 00:08:51.377 [2024-07-15 11:19:34.727870] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:51.377 [2024-07-15 11:19:34.727934] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid841590 ] 00:08:51.377 [2024-07-15 11:19:34.856262] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:51.377 [2024-07-15 11:19:34.956571] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:52.309 [2024-07-15 11:19:35.724414] accel_dpdk_compressdev.c: 296:accel_init_compress_drivers: *NOTICE*: initialized QAT PMD 00:08:52.309 [2024-07-15 11:19:35.726979] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x27ab080 PMD being used: compress_qat 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:52.309 [2024-07-15 11:19:35.731844] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x27b02a0 PMD being used: compress_qat 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val=0x1 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:52.309 [2024-07-15 11:19:35.734395] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x28d30f0 PMD being used: compress_qat 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val=decompress 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@23 -- # accel_opc=decompress 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val='4096 bytes' 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val=dpdk_compressdev 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@22 -- # accel_module=dpdk_compressdev 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:52.309 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val=32 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val=32 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val=2 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val='1 seconds' 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val=Yes 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:52.310 11:19:35 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:53.681 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:53.681 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:53.681 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:53.681 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:53.681 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:53.681 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:53.681 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:53.681 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:53.681 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:53.681 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:53.681 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:53.681 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:53.681 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:53.681 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:53.681 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:53.682 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:53.682 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:53.682 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:53.682 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:53.682 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:53.682 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:53.682 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:53.682 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:53.682 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:53.682 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@20 -- # val= 00:08:53.682 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:53.682 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:53.682 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:53.682 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@27 -- # [[ -n dpdk_compressdev ]] 00:08:53.682 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:08:53.682 11:19:36 accel.accel_cdev_decomp_mthread -- accel/accel.sh@27 -- # [[ dpdk_compressdev == \d\p\d\k\_\c\o\m\p\r\e\s\s\d\e\v ]] 00:08:53.682 00:08:53.682 real 0m2.225s 00:08:53.682 user 0m1.652s 00:08:53.682 sys 0m0.578s 00:08:53.682 11:19:36 accel.accel_cdev_decomp_mthread -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:53.682 11:19:36 accel.accel_cdev_decomp_mthread -- common/autotest_common.sh@10 -- # set +x 00:08:53.682 ************************************ 00:08:53.682 END TEST accel_cdev_decomp_mthread 00:08:53.682 ************************************ 00:08:53.682 11:19:36 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:53.682 11:19:36 accel -- accel/accel.sh@133 -- # run_test accel_cdev_decomp_full_mthread accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 -T 2 00:08:53.682 11:19:36 accel -- common/autotest_common.sh@1099 -- # '[' 13 -le 1 ']' 00:08:53.682 11:19:36 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:53.682 11:19:36 accel -- common/autotest_common.sh@10 -- # set +x 00:08:53.682 ************************************ 00:08:53.682 START TEST accel_cdev_decomp_full_mthread 00:08:53.682 ************************************ 00:08:53.682 11:19:37 accel.accel_cdev_decomp_full_mthread -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 -T 2 00:08:53.682 11:19:37 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@16 -- # local accel_opc 00:08:53.682 11:19:37 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@17 -- # local accel_module 00:08:53.682 11:19:37 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:53.682 11:19:37 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:53.682 11:19:37 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 -T 2 00:08:53.682 11:19:37 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@12 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib -y -o 0 -T 2 00:08:53.682 11:19:37 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@12 -- # build_accel_config 00:08:53.682 11:19:37 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:53.682 11:19:37 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:53.682 11:19:37 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:53.682 11:19:37 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:53.682 11:19:37 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@36 -- # [[ -n 1 ]] 00:08:53.682 11:19:37 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@37 -- # accel_json_cfg+=('{"method": "compressdev_scan_accel_module", "params":{"pmd": 0}}') 00:08:53.682 11:19:37 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@40 -- # local IFS=, 00:08:53.682 11:19:37 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@41 -- # jq -r . 00:08:53.682 [2024-07-15 11:19:37.039103] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:53.682 [2024-07-15 11:19:37.039170] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid841942 ] 00:08:53.682 [2024-07-15 11:19:37.168400] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:53.682 [2024-07-15 11:19:37.274516] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:54.616 [2024-07-15 11:19:38.047194] accel_dpdk_compressdev.c: 296:accel_init_compress_drivers: *NOTICE*: initialized QAT PMD 00:08:54.616 [2024-07-15 11:19:38.049803] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x11f8080 PMD being used: compress_qat 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:54.616 [2024-07-15 11:19:38.054046] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x11fb3b0 PMD being used: compress_qat 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val=0x1 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:54.616 [2024-07-15 11:19:38.056916] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x131fcc0 PMD being used: compress_qat 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val=decompress 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@23 -- # accel_opc=decompress 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val='111250 bytes' 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val=dpdk_compressdev 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@22 -- # accel_module=dpdk_compressdev 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/bib 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val=32 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val=32 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val=2 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val='1 seconds' 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val=Yes 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:54.616 11:19:38 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:55.988 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:55.988 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@27 -- # [[ -n dpdk_compressdev ]] 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- accel/accel.sh@27 -- # [[ dpdk_compressdev == \d\p\d\k\_\c\o\m\p\r\e\s\s\d\e\v ]] 00:08:55.989 00:08:55.989 real 0m2.240s 00:08:55.989 user 0m1.643s 00:08:55.989 sys 0m0.604s 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:55.989 11:19:39 accel.accel_cdev_decomp_full_mthread -- common/autotest_common.sh@10 -- # set +x 00:08:55.989 ************************************ 00:08:55.989 END TEST accel_cdev_decomp_full_mthread 00:08:55.989 ************************************ 00:08:55.989 11:19:39 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:55.989 11:19:39 accel -- accel/accel.sh@134 -- # unset COMPRESSDEV 00:08:55.989 11:19:39 accel -- accel/accel.sh@137 -- # run_test accel_dif_functional_tests /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/dif/dif -c /dev/fd/62 00:08:55.989 11:19:39 accel -- accel/accel.sh@137 -- # build_accel_config 00:08:55.989 11:19:39 accel -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:08:55.989 11:19:39 accel -- accel/accel.sh@31 -- # accel_json_cfg=() 00:08:55.989 11:19:39 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:55.989 11:19:39 accel -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:08:55.989 11:19:39 accel -- common/autotest_common.sh@10 -- # set +x 00:08:55.989 11:19:39 accel -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:08:55.989 11:19:39 accel -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:08:55.989 11:19:39 accel -- accel/accel.sh@36 -- # [[ -n '' ]] 00:08:55.989 11:19:39 accel -- accel/accel.sh@40 -- # local IFS=, 00:08:55.989 11:19:39 accel -- accel/accel.sh@41 -- # jq -r . 00:08:55.989 ************************************ 00:08:55.989 START TEST accel_dif_functional_tests 00:08:55.989 ************************************ 00:08:55.989 11:19:39 accel.accel_dif_functional_tests -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/dif/dif -c /dev/fd/62 00:08:55.989 [2024-07-15 11:19:39.363023] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:55.989 [2024-07-15 11:19:39.363068] [ DPDK EAL parameters: DIF --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid842155 ] 00:08:55.989 [2024-07-15 11:19:39.476910] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:08:55.989 [2024-07-15 11:19:39.580488] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:08:55.989 [2024-07-15 11:19:39.580511] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:08:55.989 [2024-07-15 11:19:39.580515] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:56.247 00:08:56.247 00:08:56.247 CUnit - A unit testing framework for C - Version 2.1-3 00:08:56.247 http://cunit.sourceforge.net/ 00:08:56.247 00:08:56.247 00:08:56.247 Suite: accel_dif 00:08:56.247 Test: verify: DIF generated, GUARD check ...passed 00:08:56.247 Test: verify: DIF generated, APPTAG check ...passed 00:08:56.247 Test: verify: DIF generated, REFTAG check ...passed 00:08:56.247 Test: verify: DIF not generated, GUARD check ...[2024-07-15 11:19:39.686622] dif.c: 826:_dif_verify: *ERROR*: Failed to compare Guard: LBA=10, Expected=5a5a, Actual=7867 00:08:56.247 passed 00:08:56.247 Test: verify: DIF not generated, APPTAG check ...[2024-07-15 11:19:39.686692] dif.c: 841:_dif_verify: *ERROR*: Failed to compare App Tag: LBA=10, Expected=14, Actual=5a5a 00:08:56.247 passed 00:08:56.247 Test: verify: DIF not generated, REFTAG check ...[2024-07-15 11:19:39.686730] dif.c: 776:_dif_reftag_check: *ERROR*: Failed to compare Ref Tag: LBA=10, Expected=a, Actual=5a5a5a5a 00:08:56.247 passed 00:08:56.247 Test: verify: APPTAG correct, APPTAG check ...passed 00:08:56.247 Test: verify: APPTAG incorrect, APPTAG check ...[2024-07-15 11:19:39.686801] dif.c: 841:_dif_verify: *ERROR*: Failed to compare App Tag: LBA=30, Expected=28, Actual=14 00:08:56.247 passed 00:08:56.247 Test: verify: APPTAG incorrect, no APPTAG check ...passed 00:08:56.247 Test: verify: REFTAG incorrect, REFTAG ignore ...passed 00:08:56.247 Test: verify: REFTAG_INIT correct, REFTAG check ...passed 00:08:56.247 Test: verify: REFTAG_INIT incorrect, REFTAG check ...[2024-07-15 11:19:39.686962] dif.c: 776:_dif_reftag_check: *ERROR*: Failed to compare Ref Tag: LBA=10, Expected=a, Actual=10 00:08:56.247 passed 00:08:56.247 Test: verify copy: DIF generated, GUARD check ...passed 00:08:56.247 Test: verify copy: DIF generated, APPTAG check ...passed 00:08:56.247 Test: verify copy: DIF generated, REFTAG check ...passed 00:08:56.247 Test: verify copy: DIF not generated, GUARD check ...[2024-07-15 11:19:39.687126] dif.c: 826:_dif_verify: *ERROR*: Failed to compare Guard: LBA=10, Expected=5a5a, Actual=7867 00:08:56.247 passed 00:08:56.247 Test: verify copy: DIF not generated, APPTAG check ...[2024-07-15 11:19:39.687160] dif.c: 841:_dif_verify: *ERROR*: Failed to compare App Tag: LBA=10, Expected=14, Actual=5a5a 00:08:56.247 passed 00:08:56.247 Test: verify copy: DIF not generated, REFTAG check ...[2024-07-15 11:19:39.687193] dif.c: 776:_dif_reftag_check: *ERROR*: Failed to compare Ref Tag: LBA=10, Expected=a, Actual=5a5a5a5a 00:08:56.247 passed 00:08:56.247 Test: generate copy: DIF generated, GUARD check ...passed 00:08:56.247 Test: generate copy: DIF generated, APTTAG check ...passed 00:08:56.247 Test: generate copy: DIF generated, REFTAG check ...passed 00:08:56.247 Test: generate copy: DIF generated, no GUARD check flag set ...passed 00:08:56.247 Test: generate copy: DIF generated, no APPTAG check flag set ...passed 00:08:56.247 Test: generate copy: DIF generated, no REFTAG check flag set ...passed 00:08:56.247 Test: generate copy: iovecs-len validate ...[2024-07-15 11:19:39.687437] dif.c:1190:spdk_dif_generate_copy: *ERROR*: Size of bounce_iovs arrays are not valid or misaligned with block_size. 00:08:56.247 passed 00:08:56.247 Test: generate copy: buffer alignment validate ...passed 00:08:56.247 00:08:56.247 Run Summary: Type Total Ran Passed Failed Inactive 00:08:56.247 suites 1 1 n/a 0 0 00:08:56.247 tests 26 26 26 0 0 00:08:56.247 asserts 115 115 115 0 n/a 00:08:56.247 00:08:56.247 Elapsed time = 0.003 seconds 00:08:56.505 00:08:56.505 real 0m0.564s 00:08:56.505 user 0m0.799s 00:08:56.505 sys 0m0.205s 00:08:56.505 11:19:39 accel.accel_dif_functional_tests -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:56.505 11:19:39 accel.accel_dif_functional_tests -- common/autotest_common.sh@10 -- # set +x 00:08:56.505 ************************************ 00:08:56.505 END TEST accel_dif_functional_tests 00:08:56.505 ************************************ 00:08:56.505 11:19:39 accel -- common/autotest_common.sh@1142 -- # return 0 00:08:56.505 00:08:56.505 real 0m53.220s 00:08:56.505 user 1m1.314s 00:08:56.505 sys 0m11.806s 00:08:56.505 11:19:39 accel -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:56.505 11:19:39 accel -- common/autotest_common.sh@10 -- # set +x 00:08:56.505 ************************************ 00:08:56.505 END TEST accel 00:08:56.505 ************************************ 00:08:56.505 11:19:39 -- common/autotest_common.sh@1142 -- # return 0 00:08:56.505 11:19:39 -- spdk/autotest.sh@184 -- # run_test accel_rpc /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/accel_rpc.sh 00:08:56.505 11:19:39 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:08:56.505 11:19:39 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:56.505 11:19:39 -- common/autotest_common.sh@10 -- # set +x 00:08:56.505 ************************************ 00:08:56.505 START TEST accel_rpc 00:08:56.506 ************************************ 00:08:56.506 11:19:40 accel_rpc -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel/accel_rpc.sh 00:08:56.763 * Looking for test storage... 00:08:56.763 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/accel 00:08:56.763 11:19:40 accel_rpc -- accel/accel_rpc.sh@11 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:08:56.763 11:19:40 accel_rpc -- accel/accel_rpc.sh@14 -- # spdk_tgt_pid=842389 00:08:56.763 11:19:40 accel_rpc -- accel/accel_rpc.sh@15 -- # waitforlisten 842389 00:08:56.763 11:19:40 accel_rpc -- accel/accel_rpc.sh@13 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt --wait-for-rpc 00:08:56.763 11:19:40 accel_rpc -- common/autotest_common.sh@829 -- # '[' -z 842389 ']' 00:08:56.763 11:19:40 accel_rpc -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:56.763 11:19:40 accel_rpc -- common/autotest_common.sh@834 -- # local max_retries=100 00:08:56.763 11:19:40 accel_rpc -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:56.763 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:56.763 11:19:40 accel_rpc -- common/autotest_common.sh@838 -- # xtrace_disable 00:08:56.763 11:19:40 accel_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:56.763 [2024-07-15 11:19:40.203707] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:56.763 [2024-07-15 11:19:40.203782] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid842389 ] 00:08:56.763 [2024-07-15 11:19:40.334349] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:57.020 [2024-07-15 11:19:40.438213] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:57.583 11:19:41 accel_rpc -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:08:57.583 11:19:41 accel_rpc -- common/autotest_common.sh@862 -- # return 0 00:08:57.583 11:19:41 accel_rpc -- accel/accel_rpc.sh@45 -- # [[ y == y ]] 00:08:57.583 11:19:41 accel_rpc -- accel/accel_rpc.sh@45 -- # [[ 0 -gt 0 ]] 00:08:57.583 11:19:41 accel_rpc -- accel/accel_rpc.sh@49 -- # [[ y == y ]] 00:08:57.583 11:19:41 accel_rpc -- accel/accel_rpc.sh@49 -- # [[ 0 -gt 0 ]] 00:08:57.583 11:19:41 accel_rpc -- accel/accel_rpc.sh@53 -- # run_test accel_assign_opcode accel_assign_opcode_test_suite 00:08:57.583 11:19:41 accel_rpc -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:08:57.583 11:19:41 accel_rpc -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:57.583 11:19:41 accel_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:57.583 ************************************ 00:08:57.583 START TEST accel_assign_opcode 00:08:57.583 ************************************ 00:08:57.583 11:19:41 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@1123 -- # accel_assign_opcode_test_suite 00:08:57.583 11:19:41 accel_rpc.accel_assign_opcode -- accel/accel_rpc.sh@38 -- # rpc_cmd accel_assign_opc -o copy -m incorrect 00:08:57.583 11:19:41 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:57.583 11:19:41 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@10 -- # set +x 00:08:57.583 [2024-07-15 11:19:41.120429] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation copy will be assigned to module incorrect 00:08:57.583 11:19:41 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:57.583 11:19:41 accel_rpc.accel_assign_opcode -- accel/accel_rpc.sh@40 -- # rpc_cmd accel_assign_opc -o copy -m software 00:08:57.583 11:19:41 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:57.583 11:19:41 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@10 -- # set +x 00:08:57.583 [2024-07-15 11:19:41.128436] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation copy will be assigned to module software 00:08:57.583 11:19:41 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:57.583 11:19:41 accel_rpc.accel_assign_opcode -- accel/accel_rpc.sh@41 -- # rpc_cmd framework_start_init 00:08:57.583 11:19:41 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:57.583 11:19:41 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@10 -- # set +x 00:08:57.840 11:19:41 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:57.840 11:19:41 accel_rpc.accel_assign_opcode -- accel/accel_rpc.sh@42 -- # jq -r .copy 00:08:57.840 11:19:41 accel_rpc.accel_assign_opcode -- accel/accel_rpc.sh@42 -- # rpc_cmd accel_get_opc_assignments 00:08:57.840 11:19:41 accel_rpc.accel_assign_opcode -- accel/accel_rpc.sh@42 -- # grep software 00:08:57.840 11:19:41 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:57.840 11:19:41 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@10 -- # set +x 00:08:57.840 11:19:41 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:57.840 software 00:08:57.840 00:08:57.840 real 0m0.279s 00:08:57.840 user 0m0.035s 00:08:57.840 sys 0m0.010s 00:08:57.840 11:19:41 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:57.840 11:19:41 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@10 -- # set +x 00:08:57.840 ************************************ 00:08:57.840 END TEST accel_assign_opcode 00:08:57.840 ************************************ 00:08:57.840 11:19:41 accel_rpc -- common/autotest_common.sh@1142 -- # return 0 00:08:57.840 11:19:41 accel_rpc -- accel/accel_rpc.sh@55 -- # killprocess 842389 00:08:58.097 11:19:41 accel_rpc -- common/autotest_common.sh@948 -- # '[' -z 842389 ']' 00:08:58.097 11:19:41 accel_rpc -- common/autotest_common.sh@952 -- # kill -0 842389 00:08:58.097 11:19:41 accel_rpc -- common/autotest_common.sh@953 -- # uname 00:08:58.097 11:19:41 accel_rpc -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:08:58.097 11:19:41 accel_rpc -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 842389 00:08:58.097 11:19:41 accel_rpc -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:08:58.097 11:19:41 accel_rpc -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:08:58.097 11:19:41 accel_rpc -- common/autotest_common.sh@966 -- # echo 'killing process with pid 842389' 00:08:58.097 killing process with pid 842389 00:08:58.097 11:19:41 accel_rpc -- common/autotest_common.sh@967 -- # kill 842389 00:08:58.097 11:19:41 accel_rpc -- common/autotest_common.sh@972 -- # wait 842389 00:08:58.354 00:08:58.354 real 0m1.854s 00:08:58.354 user 0m1.852s 00:08:58.354 sys 0m0.595s 00:08:58.354 11:19:41 accel_rpc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:58.355 11:19:41 accel_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:58.355 ************************************ 00:08:58.355 END TEST accel_rpc 00:08:58.355 ************************************ 00:08:58.355 11:19:41 -- common/autotest_common.sh@1142 -- # return 0 00:08:58.355 11:19:41 -- spdk/autotest.sh@185 -- # run_test app_cmdline /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/cmdline.sh 00:08:58.355 11:19:41 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:08:58.355 11:19:41 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:58.355 11:19:41 -- common/autotest_common.sh@10 -- # set +x 00:08:58.612 ************************************ 00:08:58.612 START TEST app_cmdline 00:08:58.612 ************************************ 00:08:58.612 11:19:41 app_cmdline -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/cmdline.sh 00:08:58.612 * Looking for test storage... 00:08:58.612 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app 00:08:58.612 11:19:42 app_cmdline -- app/cmdline.sh@14 -- # trap 'killprocess $spdk_tgt_pid' EXIT 00:08:58.612 11:19:42 app_cmdline -- app/cmdline.sh@17 -- # spdk_tgt_pid=842645 00:08:58.612 11:19:42 app_cmdline -- app/cmdline.sh@18 -- # waitforlisten 842645 00:08:58.612 11:19:42 app_cmdline -- app/cmdline.sh@16 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt --rpcs-allowed spdk_get_version,rpc_get_methods 00:08:58.612 11:19:42 app_cmdline -- common/autotest_common.sh@829 -- # '[' -z 842645 ']' 00:08:58.612 11:19:42 app_cmdline -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:58.612 11:19:42 app_cmdline -- common/autotest_common.sh@834 -- # local max_retries=100 00:08:58.612 11:19:42 app_cmdline -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:58.612 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:58.612 11:19:42 app_cmdline -- common/autotest_common.sh@838 -- # xtrace_disable 00:08:58.612 11:19:42 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:08:58.612 [2024-07-15 11:19:42.137099] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:08:58.612 [2024-07-15 11:19:42.137172] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid842645 ] 00:08:58.869 [2024-07-15 11:19:42.260115] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:58.869 [2024-07-15 11:19:42.357347] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:59.803 11:19:43 app_cmdline -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:08:59.803 11:19:43 app_cmdline -- common/autotest_common.sh@862 -- # return 0 00:08:59.803 11:19:43 app_cmdline -- app/cmdline.sh@20 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py spdk_get_version 00:08:59.803 { 00:08:59.803 "version": "SPDK v24.09-pre git sha1 e7cce062d", 00:08:59.803 "fields": { 00:08:59.803 "major": 24, 00:08:59.803 "minor": 9, 00:08:59.803 "patch": 0, 00:08:59.803 "suffix": "-pre", 00:08:59.803 "commit": "e7cce062d" 00:08:59.803 } 00:08:59.803 } 00:08:59.803 11:19:43 app_cmdline -- app/cmdline.sh@22 -- # expected_methods=() 00:08:59.803 11:19:43 app_cmdline -- app/cmdline.sh@23 -- # expected_methods+=("rpc_get_methods") 00:08:59.803 11:19:43 app_cmdline -- app/cmdline.sh@24 -- # expected_methods+=("spdk_get_version") 00:08:59.803 11:19:43 app_cmdline -- app/cmdline.sh@26 -- # methods=($(rpc_cmd rpc_get_methods | jq -r ".[]" | sort)) 00:08:59.803 11:19:43 app_cmdline -- app/cmdline.sh@26 -- # rpc_cmd rpc_get_methods 00:08:59.803 11:19:43 app_cmdline -- app/cmdline.sh@26 -- # sort 00:08:59.803 11:19:43 app_cmdline -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:59.803 11:19:43 app_cmdline -- app/cmdline.sh@26 -- # jq -r '.[]' 00:08:59.803 11:19:43 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:08:59.803 11:19:43 app_cmdline -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:59.803 11:19:43 app_cmdline -- app/cmdline.sh@27 -- # (( 2 == 2 )) 00:08:59.803 11:19:43 app_cmdline -- app/cmdline.sh@28 -- # [[ rpc_get_methods spdk_get_version == \r\p\c\_\g\e\t\_\m\e\t\h\o\d\s\ \s\p\d\k\_\g\e\t\_\v\e\r\s\i\o\n ]] 00:08:59.803 11:19:43 app_cmdline -- app/cmdline.sh@30 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:08:59.803 11:19:43 app_cmdline -- common/autotest_common.sh@648 -- # local es=0 00:08:59.803 11:19:43 app_cmdline -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:08:59.803 11:19:43 app_cmdline -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:08:59.803 11:19:43 app_cmdline -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:08:59.803 11:19:43 app_cmdline -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:08:59.803 11:19:43 app_cmdline -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:08:59.803 11:19:43 app_cmdline -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:08:59.803 11:19:43 app_cmdline -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:08:59.803 11:19:43 app_cmdline -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:08:59.803 11:19:43 app_cmdline -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py ]] 00:08:59.803 11:19:43 app_cmdline -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:09:00.061 request: 00:09:00.061 { 00:09:00.061 "method": "env_dpdk_get_mem_stats", 00:09:00.061 "req_id": 1 00:09:00.062 } 00:09:00.062 Got JSON-RPC error response 00:09:00.062 response: 00:09:00.062 { 00:09:00.062 "code": -32601, 00:09:00.062 "message": "Method not found" 00:09:00.062 } 00:09:00.062 11:19:43 app_cmdline -- common/autotest_common.sh@651 -- # es=1 00:09:00.062 11:19:43 app_cmdline -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:09:00.062 11:19:43 app_cmdline -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:09:00.062 11:19:43 app_cmdline -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:09:00.062 11:19:43 app_cmdline -- app/cmdline.sh@1 -- # killprocess 842645 00:09:00.062 11:19:43 app_cmdline -- common/autotest_common.sh@948 -- # '[' -z 842645 ']' 00:09:00.062 11:19:43 app_cmdline -- common/autotest_common.sh@952 -- # kill -0 842645 00:09:00.062 11:19:43 app_cmdline -- common/autotest_common.sh@953 -- # uname 00:09:00.062 11:19:43 app_cmdline -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:09:00.062 11:19:43 app_cmdline -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 842645 00:09:00.062 11:19:43 app_cmdline -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:09:00.062 11:19:43 app_cmdline -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:09:00.062 11:19:43 app_cmdline -- common/autotest_common.sh@966 -- # echo 'killing process with pid 842645' 00:09:00.062 killing process with pid 842645 00:09:00.062 11:19:43 app_cmdline -- common/autotest_common.sh@967 -- # kill 842645 00:09:00.062 11:19:43 app_cmdline -- common/autotest_common.sh@972 -- # wait 842645 00:09:00.624 00:09:00.624 real 0m2.071s 00:09:00.624 user 0m2.476s 00:09:00.624 sys 0m0.629s 00:09:00.624 11:19:44 app_cmdline -- common/autotest_common.sh@1124 -- # xtrace_disable 00:09:00.624 11:19:44 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:09:00.624 ************************************ 00:09:00.624 END TEST app_cmdline 00:09:00.624 ************************************ 00:09:00.624 11:19:44 -- common/autotest_common.sh@1142 -- # return 0 00:09:00.624 11:19:44 -- spdk/autotest.sh@186 -- # run_test version /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/version.sh 00:09:00.624 11:19:44 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:09:00.624 11:19:44 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:09:00.624 11:19:44 -- common/autotest_common.sh@10 -- # set +x 00:09:00.624 ************************************ 00:09:00.624 START TEST version 00:09:00.624 ************************************ 00:09:00.624 11:19:44 version -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/version.sh 00:09:00.624 * Looking for test storage... 00:09:00.624 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app 00:09:00.624 11:19:44 version -- app/version.sh@17 -- # get_header_version major 00:09:00.625 11:19:44 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MAJOR[[:space:]]+' /var/jenkins/workspace/crypto-phy-autotest/spdk/include/spdk/version.h 00:09:00.625 11:19:44 version -- app/version.sh@14 -- # cut -f2 00:09:00.625 11:19:44 version -- app/version.sh@14 -- # tr -d '"' 00:09:00.625 11:19:44 version -- app/version.sh@17 -- # major=24 00:09:00.625 11:19:44 version -- app/version.sh@18 -- # get_header_version minor 00:09:00.625 11:19:44 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MINOR[[:space:]]+' /var/jenkins/workspace/crypto-phy-autotest/spdk/include/spdk/version.h 00:09:00.625 11:19:44 version -- app/version.sh@14 -- # cut -f2 00:09:00.625 11:19:44 version -- app/version.sh@14 -- # tr -d '"' 00:09:00.882 11:19:44 version -- app/version.sh@18 -- # minor=9 00:09:00.882 11:19:44 version -- app/version.sh@19 -- # get_header_version patch 00:09:00.882 11:19:44 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_PATCH[[:space:]]+' /var/jenkins/workspace/crypto-phy-autotest/spdk/include/spdk/version.h 00:09:00.882 11:19:44 version -- app/version.sh@14 -- # cut -f2 00:09:00.882 11:19:44 version -- app/version.sh@14 -- # tr -d '"' 00:09:00.882 11:19:44 version -- app/version.sh@19 -- # patch=0 00:09:00.882 11:19:44 version -- app/version.sh@20 -- # get_header_version suffix 00:09:00.882 11:19:44 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_SUFFIX[[:space:]]+' /var/jenkins/workspace/crypto-phy-autotest/spdk/include/spdk/version.h 00:09:00.882 11:19:44 version -- app/version.sh@14 -- # cut -f2 00:09:00.882 11:19:44 version -- app/version.sh@14 -- # tr -d '"' 00:09:00.882 11:19:44 version -- app/version.sh@20 -- # suffix=-pre 00:09:00.882 11:19:44 version -- app/version.sh@22 -- # version=24.9 00:09:00.882 11:19:44 version -- app/version.sh@25 -- # (( patch != 0 )) 00:09:00.882 11:19:44 version -- app/version.sh@28 -- # version=24.9rc0 00:09:00.882 11:19:44 version -- app/version.sh@30 -- # PYTHONPATH=:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/python 00:09:00.882 11:19:44 version -- app/version.sh@30 -- # python3 -c 'import spdk; print(spdk.__version__)' 00:09:00.882 11:19:44 version -- app/version.sh@30 -- # py_version=24.9rc0 00:09:00.882 11:19:44 version -- app/version.sh@31 -- # [[ 24.9rc0 == \2\4\.\9\r\c\0 ]] 00:09:00.882 00:09:00.883 real 0m0.187s 00:09:00.883 user 0m0.093s 00:09:00.883 sys 0m0.137s 00:09:00.883 11:19:44 version -- common/autotest_common.sh@1124 -- # xtrace_disable 00:09:00.883 11:19:44 version -- common/autotest_common.sh@10 -- # set +x 00:09:00.883 ************************************ 00:09:00.883 END TEST version 00:09:00.883 ************************************ 00:09:00.883 11:19:44 -- common/autotest_common.sh@1142 -- # return 0 00:09:00.883 11:19:44 -- spdk/autotest.sh@188 -- # '[' 1 -eq 1 ']' 00:09:00.883 11:19:44 -- spdk/autotest.sh@189 -- # run_test blockdev_general /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/blockdev.sh 00:09:00.883 11:19:44 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:09:00.883 11:19:44 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:09:00.883 11:19:44 -- common/autotest_common.sh@10 -- # set +x 00:09:00.883 ************************************ 00:09:00.883 START TEST blockdev_general 00:09:00.883 ************************************ 00:09:00.883 11:19:44 blockdev_general -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/blockdev.sh 00:09:00.883 * Looking for test storage... 00:09:01.141 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@10 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbd_common.sh 00:09:01.141 11:19:44 blockdev_general -- bdev/nbd_common.sh@6 -- # set -e 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@12 -- # rpc_py=rpc_cmd 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@13 -- # conf_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@14 -- # nonenclosed_conf_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonenclosed.json 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@15 -- # nonarray_conf_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonarray.json 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@17 -- # export RPC_PIPE_TIMEOUT=30 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@17 -- # RPC_PIPE_TIMEOUT=30 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@20 -- # : 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@670 -- # QOS_DEV_1=Malloc_0 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@671 -- # QOS_DEV_2=Null_1 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@672 -- # QOS_RUN_TIME=5 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@674 -- # uname -s 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@674 -- # '[' Linux = Linux ']' 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@676 -- # PRE_RESERVED_MEM=0 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@682 -- # test_type=bdev 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@683 -- # crypto_device= 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@684 -- # dek= 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@685 -- # env_ctx= 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@686 -- # wait_for_rpc= 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@687 -- # '[' -n '' ']' 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@690 -- # [[ bdev == bdev ]] 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@691 -- # wait_for_rpc=--wait-for-rpc 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@693 -- # start_spdk_tgt 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@47 -- # spdk_tgt_pid=843115 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@46 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt '' --wait-for-rpc 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@48 -- # trap 'killprocess "$spdk_tgt_pid"; exit 1' SIGINT SIGTERM EXIT 00:09:01.141 11:19:44 blockdev_general -- bdev/blockdev.sh@49 -- # waitforlisten 843115 00:09:01.141 11:19:44 blockdev_general -- common/autotest_common.sh@829 -- # '[' -z 843115 ']' 00:09:01.141 11:19:44 blockdev_general -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:01.141 11:19:44 blockdev_general -- common/autotest_common.sh@834 -- # local max_retries=100 00:09:01.141 11:19:44 blockdev_general -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:01.141 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:01.141 11:19:44 blockdev_general -- common/autotest_common.sh@838 -- # xtrace_disable 00:09:01.141 11:19:44 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:09:01.141 [2024-07-15 11:19:44.563013] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:09:01.141 [2024-07-15 11:19:44.563084] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid843115 ] 00:09:01.141 [2024-07-15 11:19:44.691670] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:01.515 [2024-07-15 11:19:44.790340] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:09:02.078 11:19:45 blockdev_general -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:09:02.078 11:19:45 blockdev_general -- common/autotest_common.sh@862 -- # return 0 00:09:02.078 11:19:45 blockdev_general -- bdev/blockdev.sh@694 -- # case "$test_type" in 00:09:02.078 11:19:45 blockdev_general -- bdev/blockdev.sh@696 -- # setup_bdev_conf 00:09:02.078 11:19:45 blockdev_general -- bdev/blockdev.sh@53 -- # rpc_cmd 00:09:02.078 11:19:45 blockdev_general -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:02.078 11:19:45 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:09:02.334 [2024-07-15 11:19:45.752395] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:09:02.334 [2024-07-15 11:19:45.752449] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:09:02.334 00:09:02.334 [2024-07-15 11:19:45.760381] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:09:02.334 [2024-07-15 11:19:45.760406] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:09:02.334 00:09:02.334 Malloc0 00:09:02.334 Malloc1 00:09:02.334 Malloc2 00:09:02.334 Malloc3 00:09:02.334 Malloc4 00:09:02.334 Malloc5 00:09:02.334 Malloc6 00:09:02.334 Malloc7 00:09:02.334 Malloc8 00:09:02.334 Malloc9 00:09:02.334 [2024-07-15 11:19:45.908829] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc3 00:09:02.335 [2024-07-15 11:19:45.908876] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:02.335 [2024-07-15 11:19:45.908895] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1304350 00:09:02.335 [2024-07-15 11:19:45.908908] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:02.335 [2024-07-15 11:19:45.910242] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:02.335 [2024-07-15 11:19:45.910269] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: TestPT 00:09:02.335 TestPT 00:09:02.591 11:19:45 blockdev_general -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:02.591 11:19:45 blockdev_general -- bdev/blockdev.sh@76 -- # dd if=/dev/zero of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/aiofile bs=2048 count=5000 00:09:02.591 5000+0 records in 00:09:02.591 5000+0 records out 00:09:02.591 10240000 bytes (10 MB, 9.8 MiB) copied, 0.017297 s, 592 MB/s 00:09:02.591 11:19:45 blockdev_general -- bdev/blockdev.sh@77 -- # rpc_cmd bdev_aio_create /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/aiofile AIO0 2048 00:09:02.591 11:19:45 blockdev_general -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:02.591 11:19:45 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:09:02.591 AIO0 00:09:02.591 11:19:45 blockdev_general -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:02.591 11:19:45 blockdev_general -- bdev/blockdev.sh@737 -- # rpc_cmd bdev_wait_for_examine 00:09:02.591 11:19:45 blockdev_general -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:02.591 11:19:45 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:09:02.591 11:19:46 blockdev_general -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:02.591 11:19:46 blockdev_general -- bdev/blockdev.sh@740 -- # cat 00:09:02.591 11:19:46 blockdev_general -- bdev/blockdev.sh@740 -- # rpc_cmd save_subsystem_config -n accel 00:09:02.591 11:19:46 blockdev_general -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:02.591 11:19:46 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:09:02.591 11:19:46 blockdev_general -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:02.591 11:19:46 blockdev_general -- bdev/blockdev.sh@740 -- # rpc_cmd save_subsystem_config -n bdev 00:09:02.591 11:19:46 blockdev_general -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:02.591 11:19:46 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:09:02.591 11:19:46 blockdev_general -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:02.591 11:19:46 blockdev_general -- bdev/blockdev.sh@740 -- # rpc_cmd save_subsystem_config -n iobuf 00:09:02.591 11:19:46 blockdev_general -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:02.591 11:19:46 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:09:02.591 11:19:46 blockdev_general -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:02.591 11:19:46 blockdev_general -- bdev/blockdev.sh@748 -- # mapfile -t bdevs 00:09:02.591 11:19:46 blockdev_general -- bdev/blockdev.sh@748 -- # rpc_cmd bdev_get_bdevs 00:09:02.591 11:19:46 blockdev_general -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:02.591 11:19:46 blockdev_general -- bdev/blockdev.sh@748 -- # jq -r '.[] | select(.claimed == false)' 00:09:02.591 11:19:46 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:09:02.849 11:19:46 blockdev_general -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:02.849 11:19:46 blockdev_general -- bdev/blockdev.sh@749 -- # mapfile -t bdevs_name 00:09:02.849 11:19:46 blockdev_general -- bdev/blockdev.sh@749 -- # jq -r .name 00:09:02.850 11:19:46 blockdev_general -- bdev/blockdev.sh@749 -- # printf '%s\n' '{' ' "name": "Malloc0",' ' "aliases": [' ' "5ce527dc-eb58-413a-827d-f3c0a4fd090b"' ' ],' ' "product_name": "Malloc disk",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "5ce527dc-eb58-413a-827d-f3c0a4fd090b",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 20000,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {}' '}' '{' ' "name": "Malloc1p0",' ' "aliases": [' ' "5fadbd15-2a90-55bd-ae35-f5123bafe8da"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 32768,' ' "uuid": "5fadbd15-2a90-55bd-ae35-f5123bafe8da",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc1",' ' "offset_blocks": 0' ' }' ' }' '}' '{' ' "name": "Malloc1p1",' ' "aliases": [' ' "2d0d2f9e-1b12-5e40-8b76-3029d8b089cc"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 32768,' ' "uuid": "2d0d2f9e-1b12-5e40-8b76-3029d8b089cc",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc1",' ' "offset_blocks": 32768' ' }' ' }' '}' '{' ' "name": "Malloc2p0",' ' "aliases": [' ' "d022053a-fcc7-56c1-af68-8f04876a4109"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "d022053a-fcc7-56c1-af68-8f04876a4109",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 0' ' }' ' }' '}' '{' ' "name": "Malloc2p1",' ' "aliases": [' ' "b6ff7cf9-30d0-53bb-856f-e4fd483be15c"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "b6ff7cf9-30d0-53bb-856f-e4fd483be15c",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 8192' ' }' ' }' '}' '{' ' "name": "Malloc2p2",' ' "aliases": [' ' "6df4933e-5f3c-5662-8ad9-b1cb90107c3c"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "6df4933e-5f3c-5662-8ad9-b1cb90107c3c",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 16384' ' }' ' }' '}' '{' ' "name": "Malloc2p3",' ' "aliases": [' ' "f27852e2-a19f-5771-a71c-11be9d704d5d"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "f27852e2-a19f-5771-a71c-11be9d704d5d",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 24576' ' }' ' }' '}' '{' ' "name": "Malloc2p4",' ' "aliases": [' ' "86f93d9d-194b-5784-8a78-9f5e07868a7e"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "86f93d9d-194b-5784-8a78-9f5e07868a7e",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 32768' ' }' ' }' '}' '{' ' "name": "Malloc2p5",' ' "aliases": [' ' "da83a269-dd19-5e99-a12c-13edcb7db0cd"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "da83a269-dd19-5e99-a12c-13edcb7db0cd",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 40960' ' }' ' }' '}' '{' ' "name": "Malloc2p6",' ' "aliases": [' ' "9d07fa39-831f-5784-b785-279997ba3af0"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "9d07fa39-831f-5784-b785-279997ba3af0",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 49152' ' }' ' }' '}' '{' ' "name": "Malloc2p7",' ' "aliases": [' ' "6c40d456-1f09-5c69-9c20-4523d2ef39e8"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "6c40d456-1f09-5c69-9c20-4523d2ef39e8",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 57344' ' }' ' }' '}' '{' ' "name": "TestPT",' ' "aliases": [' ' "34c56ada-8252-54db-aeb7-e85474d6df8d"' ' ],' ' "product_name": "passthru",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "34c56ada-8252-54db-aeb7-e85474d6df8d",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "passthru": {' ' "name": "TestPT",' ' "base_bdev_name": "Malloc3"' ' }' ' }' '}' '{' ' "name": "raid0",' ' "aliases": [' ' "23b1ca62-1017-4838-909e-76ea1f6cad62"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "23b1ca62-1017-4838-909e-76ea1f6cad62",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "raid": {' ' "uuid": "23b1ca62-1017-4838-909e-76ea1f6cad62",' ' "strip_size_kb": 64,' ' "state": "online",' ' "raid_level": "raid0",' ' "superblock": false,' ' "num_base_bdevs": 2,' ' "num_base_bdevs_discovered": 2,' ' "num_base_bdevs_operational": 2,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc4",' ' "uuid": "6de1596c-08db-4b80-86ea-96d56c0345c7",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc5",' ' "uuid": "05f2ee9c-ccde-4341-ac99-c1ef2c8f5ce7",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' '{' ' "name": "concat0",' ' "aliases": [' ' "e2eb45e4-69ab-4edf-95dd-c81cd873e951"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "e2eb45e4-69ab-4edf-95dd-c81cd873e951",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "raid": {' ' "uuid": "e2eb45e4-69ab-4edf-95dd-c81cd873e951",' ' "strip_size_kb": 64,' ' "state": "online",' ' "raid_level": "concat",' ' "superblock": false,' ' "num_base_bdevs": 2,' ' "num_base_bdevs_discovered": 2,' ' "num_base_bdevs_operational": 2,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc6",' ' "uuid": "e5067475-6ef9-43ae-bcb9-d5ba9d228e5e",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc7",' ' "uuid": "64e1c577-11ea-4bed-84d9-9fee33627e6c",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' '{' ' "name": "raid1",' ' "aliases": [' ' "80122521-8fa1-4582-a8d0-a569befdbe0f"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "80122521-8fa1-4582-a8d0-a569befdbe0f",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "raid": {' ' "uuid": "80122521-8fa1-4582-a8d0-a569befdbe0f",' ' "strip_size_kb": 0,' ' "state": "online",' ' "raid_level": "raid1",' ' "superblock": false,' ' "num_base_bdevs": 2,' ' "num_base_bdevs_discovered": 2,' ' "num_base_bdevs_operational": 2,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc8",' ' "uuid": "91863e0b-b6e3-4ae1-85b3-dbd0ab80231f",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc9",' ' "uuid": "e8e513fa-3b18-46dd-ad2f-1e2ffbb5ae8e",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' '{' ' "name": "AIO0",' ' "aliases": [' ' "0afa1591-1717-40c4-a4dd-e5b16747b803"' ' ],' ' "product_name": "AIO disk",' ' "block_size": 2048,' ' "num_blocks": 5000,' ' "uuid": "0afa1591-1717-40c4-a4dd-e5b16747b803",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "aio": {' ' "filename": "/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/aiofile",' ' "block_size_override": true,' ' "readonly": false,' ' "fallocate": false' ' }' ' }' '}' 00:09:02.850 11:19:46 blockdev_general -- bdev/blockdev.sh@750 -- # bdev_list=("${bdevs_name[@]}") 00:09:02.850 11:19:46 blockdev_general -- bdev/blockdev.sh@752 -- # hello_world_bdev=Malloc0 00:09:02.850 11:19:46 blockdev_general -- bdev/blockdev.sh@753 -- # trap - SIGINT SIGTERM EXIT 00:09:02.850 11:19:46 blockdev_general -- bdev/blockdev.sh@754 -- # killprocess 843115 00:09:02.850 11:19:46 blockdev_general -- common/autotest_common.sh@948 -- # '[' -z 843115 ']' 00:09:02.850 11:19:46 blockdev_general -- common/autotest_common.sh@952 -- # kill -0 843115 00:09:02.850 11:19:46 blockdev_general -- common/autotest_common.sh@953 -- # uname 00:09:02.850 11:19:46 blockdev_general -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:09:02.850 11:19:46 blockdev_general -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 843115 00:09:02.850 11:19:46 blockdev_general -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:09:02.850 11:19:46 blockdev_general -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:09:02.850 11:19:46 blockdev_general -- common/autotest_common.sh@966 -- # echo 'killing process with pid 843115' 00:09:02.850 killing process with pid 843115 00:09:02.850 11:19:46 blockdev_general -- common/autotest_common.sh@967 -- # kill 843115 00:09:02.850 11:19:46 blockdev_general -- common/autotest_common.sh@972 -- # wait 843115 00:09:03.412 11:19:46 blockdev_general -- bdev/blockdev.sh@758 -- # trap cleanup SIGINT SIGTERM EXIT 00:09:03.412 11:19:46 blockdev_general -- bdev/blockdev.sh@760 -- # run_test bdev_hello_world /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/hello_bdev --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -b Malloc0 '' 00:09:03.412 11:19:46 blockdev_general -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:09:03.412 11:19:46 blockdev_general -- common/autotest_common.sh@1105 -- # xtrace_disable 00:09:03.412 11:19:46 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:09:03.412 ************************************ 00:09:03.412 START TEST bdev_hello_world 00:09:03.412 ************************************ 00:09:03.412 11:19:46 blockdev_general.bdev_hello_world -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/hello_bdev --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -b Malloc0 '' 00:09:03.412 [2024-07-15 11:19:46.959787] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:09:03.412 [2024-07-15 11:19:46.959845] [ DPDK EAL parameters: hello_bdev --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid843491 ] 00:09:03.668 [2024-07-15 11:19:47.086579] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:03.668 [2024-07-15 11:19:47.187797] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:09:03.924 [2024-07-15 11:19:47.346350] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc3 00:09:03.924 [2024-07-15 11:19:47.346412] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc3 00:09:03.924 [2024-07-15 11:19:47.346427] vbdev_passthru.c: 735:bdev_passthru_create_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:09:03.924 [2024-07-15 11:19:47.354352] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:09:03.924 [2024-07-15 11:19:47.354379] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:09:03.924 [2024-07-15 11:19:47.362364] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:09:03.924 [2024-07-15 11:19:47.362390] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:09:03.924 [2024-07-15 11:19:47.438902] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc3 00:09:03.924 [2024-07-15 11:19:47.438962] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:03.924 [2024-07-15 11:19:47.438979] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x13563c0 00:09:03.924 [2024-07-15 11:19:47.438992] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:03.924 [2024-07-15 11:19:47.440488] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:03.924 [2024-07-15 11:19:47.440519] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: TestPT 00:09:04.180 [2024-07-15 11:19:47.610302] hello_bdev.c: 222:hello_start: *NOTICE*: Successfully started the application 00:09:04.180 [2024-07-15 11:19:47.610370] hello_bdev.c: 231:hello_start: *NOTICE*: Opening the bdev Malloc0 00:09:04.180 [2024-07-15 11:19:47.610425] hello_bdev.c: 244:hello_start: *NOTICE*: Opening io channel 00:09:04.180 [2024-07-15 11:19:47.610501] hello_bdev.c: 138:hello_write: *NOTICE*: Writing to the bdev 00:09:04.180 [2024-07-15 11:19:47.610577] hello_bdev.c: 117:write_complete: *NOTICE*: bdev io write completed successfully 00:09:04.180 [2024-07-15 11:19:47.610607] hello_bdev.c: 84:hello_read: *NOTICE*: Reading io 00:09:04.180 [2024-07-15 11:19:47.610670] hello_bdev.c: 65:read_complete: *NOTICE*: Read string from bdev : Hello World! 00:09:04.180 00:09:04.180 [2024-07-15 11:19:47.610709] hello_bdev.c: 74:read_complete: *NOTICE*: Stopping app 00:09:04.437 00:09:04.437 real 0m1.028s 00:09:04.438 user 0m0.652s 00:09:04.438 sys 0m0.318s 00:09:04.438 11:19:47 blockdev_general.bdev_hello_world -- common/autotest_common.sh@1124 -- # xtrace_disable 00:09:04.438 11:19:47 blockdev_general.bdev_hello_world -- common/autotest_common.sh@10 -- # set +x 00:09:04.438 ************************************ 00:09:04.438 END TEST bdev_hello_world 00:09:04.438 ************************************ 00:09:04.438 11:19:47 blockdev_general -- common/autotest_common.sh@1142 -- # return 0 00:09:04.438 11:19:47 blockdev_general -- bdev/blockdev.sh@761 -- # run_test bdev_bounds bdev_bounds '' 00:09:04.438 11:19:47 blockdev_general -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:09:04.438 11:19:47 blockdev_general -- common/autotest_common.sh@1105 -- # xtrace_disable 00:09:04.438 11:19:47 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:09:04.438 ************************************ 00:09:04.438 START TEST bdev_bounds 00:09:04.438 ************************************ 00:09:04.438 11:19:48 blockdev_general.bdev_bounds -- common/autotest_common.sh@1123 -- # bdev_bounds '' 00:09:04.438 11:19:48 blockdev_general.bdev_bounds -- bdev/blockdev.sh@290 -- # bdevio_pid=843683 00:09:04.438 11:19:48 blockdev_general.bdev_bounds -- bdev/blockdev.sh@291 -- # trap 'cleanup; killprocess $bdevio_pid; exit 1' SIGINT SIGTERM EXIT 00:09:04.438 11:19:48 blockdev_general.bdev_bounds -- bdev/blockdev.sh@292 -- # echo 'Process bdevio pid: 843683' 00:09:04.438 Process bdevio pid: 843683 00:09:04.438 11:19:48 blockdev_general.bdev_bounds -- bdev/blockdev.sh@293 -- # waitforlisten 843683 00:09:04.438 11:19:48 blockdev_general.bdev_bounds -- common/autotest_common.sh@829 -- # '[' -z 843683 ']' 00:09:04.438 11:19:48 blockdev_general.bdev_bounds -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:04.438 11:19:48 blockdev_general.bdev_bounds -- common/autotest_common.sh@834 -- # local max_retries=100 00:09:04.438 11:19:48 blockdev_general.bdev_bounds -- bdev/blockdev.sh@289 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevio/bdevio -w -s 0 --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json '' 00:09:04.438 11:19:48 blockdev_general.bdev_bounds -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:04.438 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:04.438 11:19:48 blockdev_general.bdev_bounds -- common/autotest_common.sh@838 -- # xtrace_disable 00:09:04.438 11:19:48 blockdev_general.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:09:04.695 [2024-07-15 11:19:48.071228] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:09:04.695 [2024-07-15 11:19:48.071290] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 0 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid843683 ] 00:09:04.695 [2024-07-15 11:19:48.199957] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:09:04.952 [2024-07-15 11:19:48.310962] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:09:04.952 [2024-07-15 11:19:48.311047] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:09:04.952 [2024-07-15 11:19:48.311052] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:09:04.952 [2024-07-15 11:19:48.465170] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc3 00:09:04.952 [2024-07-15 11:19:48.465231] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc3 00:09:04.952 [2024-07-15 11:19:48.465247] vbdev_passthru.c: 735:bdev_passthru_create_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:09:04.952 [2024-07-15 11:19:48.473178] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:09:04.952 [2024-07-15 11:19:48.473205] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:09:04.952 [2024-07-15 11:19:48.481195] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:09:04.952 [2024-07-15 11:19:48.481219] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:09:05.209 [2024-07-15 11:19:48.557692] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc3 00:09:05.209 [2024-07-15 11:19:48.557744] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:05.209 [2024-07-15 11:19:48.557763] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x21210c0 00:09:05.209 [2024-07-15 11:19:48.557776] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:05.209 [2024-07-15 11:19:48.559247] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:05.209 [2024-07-15 11:19:48.559275] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: TestPT 00:09:05.209 11:19:48 blockdev_general.bdev_bounds -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:09:05.209 11:19:48 blockdev_general.bdev_bounds -- common/autotest_common.sh@862 -- # return 0 00:09:05.209 11:19:48 blockdev_general.bdev_bounds -- bdev/blockdev.sh@294 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevio/tests.py perform_tests 00:09:05.468 I/O targets: 00:09:05.468 Malloc0: 65536 blocks of 512 bytes (32 MiB) 00:09:05.468 Malloc1p0: 32768 blocks of 512 bytes (16 MiB) 00:09:05.468 Malloc1p1: 32768 blocks of 512 bytes (16 MiB) 00:09:05.468 Malloc2p0: 8192 blocks of 512 bytes (4 MiB) 00:09:05.468 Malloc2p1: 8192 blocks of 512 bytes (4 MiB) 00:09:05.468 Malloc2p2: 8192 blocks of 512 bytes (4 MiB) 00:09:05.468 Malloc2p3: 8192 blocks of 512 bytes (4 MiB) 00:09:05.468 Malloc2p4: 8192 blocks of 512 bytes (4 MiB) 00:09:05.468 Malloc2p5: 8192 blocks of 512 bytes (4 MiB) 00:09:05.468 Malloc2p6: 8192 blocks of 512 bytes (4 MiB) 00:09:05.468 Malloc2p7: 8192 blocks of 512 bytes (4 MiB) 00:09:05.468 TestPT: 65536 blocks of 512 bytes (32 MiB) 00:09:05.468 raid0: 131072 blocks of 512 bytes (64 MiB) 00:09:05.468 concat0: 131072 blocks of 512 bytes (64 MiB) 00:09:05.468 raid1: 65536 blocks of 512 bytes (32 MiB) 00:09:05.468 AIO0: 5000 blocks of 2048 bytes (10 MiB) 00:09:05.468 00:09:05.468 00:09:05.468 CUnit - A unit testing framework for C - Version 2.1-3 00:09:05.468 http://cunit.sourceforge.net/ 00:09:05.468 00:09:05.468 00:09:05.468 Suite: bdevio tests on: AIO0 00:09:05.468 Test: blockdev write read block ...passed 00:09:05.468 Test: blockdev write zeroes read block ...passed 00:09:05.468 Test: blockdev write zeroes read no split ...passed 00:09:05.468 Test: blockdev write zeroes read split ...passed 00:09:05.468 Test: blockdev write zeroes read split partial ...passed 00:09:05.468 Test: blockdev reset ...passed 00:09:05.468 Test: blockdev write read 8 blocks ...passed 00:09:05.468 Test: blockdev write read size > 128k ...passed 00:09:05.468 Test: blockdev write read invalid size ...passed 00:09:05.468 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:09:05.468 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:09:05.468 Test: blockdev write read max offset ...passed 00:09:05.468 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:09:05.468 Test: blockdev writev readv 8 blocks ...passed 00:09:05.468 Test: blockdev writev readv 30 x 1block ...passed 00:09:05.468 Test: blockdev writev readv block ...passed 00:09:05.468 Test: blockdev writev readv size > 128k ...passed 00:09:05.468 Test: blockdev writev readv size > 128k in two iovs ...passed 00:09:05.468 Test: blockdev comparev and writev ...passed 00:09:05.468 Test: blockdev nvme passthru rw ...passed 00:09:05.468 Test: blockdev nvme passthru vendor specific ...passed 00:09:05.468 Test: blockdev nvme admin passthru ...passed 00:09:05.468 Test: blockdev copy ...passed 00:09:05.468 Suite: bdevio tests on: raid1 00:09:05.468 Test: blockdev write read block ...passed 00:09:05.468 Test: blockdev write zeroes read block ...passed 00:09:05.468 Test: blockdev write zeroes read no split ...passed 00:09:05.468 Test: blockdev write zeroes read split ...passed 00:09:05.468 Test: blockdev write zeroes read split partial ...passed 00:09:05.468 Test: blockdev reset ...passed 00:09:05.468 Test: blockdev write read 8 blocks ...passed 00:09:05.468 Test: blockdev write read size > 128k ...passed 00:09:05.468 Test: blockdev write read invalid size ...passed 00:09:05.468 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:09:05.468 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:09:05.468 Test: blockdev write read max offset ...passed 00:09:05.468 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:09:05.468 Test: blockdev writev readv 8 blocks ...passed 00:09:05.468 Test: blockdev writev readv 30 x 1block ...passed 00:09:05.468 Test: blockdev writev readv block ...passed 00:09:05.468 Test: blockdev writev readv size > 128k ...passed 00:09:05.468 Test: blockdev writev readv size > 128k in two iovs ...passed 00:09:05.468 Test: blockdev comparev and writev ...passed 00:09:05.468 Test: blockdev nvme passthru rw ...passed 00:09:05.468 Test: blockdev nvme passthru vendor specific ...passed 00:09:05.468 Test: blockdev nvme admin passthru ...passed 00:09:05.468 Test: blockdev copy ...passed 00:09:05.468 Suite: bdevio tests on: concat0 00:09:05.468 Test: blockdev write read block ...passed 00:09:05.468 Test: blockdev write zeroes read block ...passed 00:09:05.468 Test: blockdev write zeroes read no split ...passed 00:09:05.468 Test: blockdev write zeroes read split ...passed 00:09:05.468 Test: blockdev write zeroes read split partial ...passed 00:09:05.468 Test: blockdev reset ...passed 00:09:05.468 Test: blockdev write read 8 blocks ...passed 00:09:05.468 Test: blockdev write read size > 128k ...passed 00:09:05.468 Test: blockdev write read invalid size ...passed 00:09:05.468 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:09:05.468 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:09:05.468 Test: blockdev write read max offset ...passed 00:09:05.468 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:09:05.468 Test: blockdev writev readv 8 blocks ...passed 00:09:05.468 Test: blockdev writev readv 30 x 1block ...passed 00:09:05.468 Test: blockdev writev readv block ...passed 00:09:05.468 Test: blockdev writev readv size > 128k ...passed 00:09:05.468 Test: blockdev writev readv size > 128k in two iovs ...passed 00:09:05.468 Test: blockdev comparev and writev ...passed 00:09:05.468 Test: blockdev nvme passthru rw ...passed 00:09:05.468 Test: blockdev nvme passthru vendor specific ...passed 00:09:05.468 Test: blockdev nvme admin passthru ...passed 00:09:05.468 Test: blockdev copy ...passed 00:09:05.468 Suite: bdevio tests on: raid0 00:09:05.468 Test: blockdev write read block ...passed 00:09:05.468 Test: blockdev write zeroes read block ...passed 00:09:05.468 Test: blockdev write zeroes read no split ...passed 00:09:05.468 Test: blockdev write zeroes read split ...passed 00:09:05.468 Test: blockdev write zeroes read split partial ...passed 00:09:05.468 Test: blockdev reset ...passed 00:09:05.468 Test: blockdev write read 8 blocks ...passed 00:09:05.468 Test: blockdev write read size > 128k ...passed 00:09:05.468 Test: blockdev write read invalid size ...passed 00:09:05.468 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:09:05.468 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:09:05.468 Test: blockdev write read max offset ...passed 00:09:05.468 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:09:05.468 Test: blockdev writev readv 8 blocks ...passed 00:09:05.468 Test: blockdev writev readv 30 x 1block ...passed 00:09:05.468 Test: blockdev writev readv block ...passed 00:09:05.468 Test: blockdev writev readv size > 128k ...passed 00:09:05.468 Test: blockdev writev readv size > 128k in two iovs ...passed 00:09:05.468 Test: blockdev comparev and writev ...passed 00:09:05.468 Test: blockdev nvme passthru rw ...passed 00:09:05.468 Test: blockdev nvme passthru vendor specific ...passed 00:09:05.468 Test: blockdev nvme admin passthru ...passed 00:09:05.468 Test: blockdev copy ...passed 00:09:05.468 Suite: bdevio tests on: TestPT 00:09:05.468 Test: blockdev write read block ...passed 00:09:05.468 Test: blockdev write zeroes read block ...passed 00:09:05.468 Test: blockdev write zeroes read no split ...passed 00:09:05.468 Test: blockdev write zeroes read split ...passed 00:09:05.468 Test: blockdev write zeroes read split partial ...passed 00:09:05.468 Test: blockdev reset ...passed 00:09:05.468 Test: blockdev write read 8 blocks ...passed 00:09:05.468 Test: blockdev write read size > 128k ...passed 00:09:05.468 Test: blockdev write read invalid size ...passed 00:09:05.468 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:09:05.468 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:09:05.468 Test: blockdev write read max offset ...passed 00:09:05.468 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:09:05.468 Test: blockdev writev readv 8 blocks ...passed 00:09:05.468 Test: blockdev writev readv 30 x 1block ...passed 00:09:05.468 Test: blockdev writev readv block ...passed 00:09:05.468 Test: blockdev writev readv size > 128k ...passed 00:09:05.468 Test: blockdev writev readv size > 128k in two iovs ...passed 00:09:05.468 Test: blockdev comparev and writev ...passed 00:09:05.468 Test: blockdev nvme passthru rw ...passed 00:09:05.468 Test: blockdev nvme passthru vendor specific ...passed 00:09:05.468 Test: blockdev nvme admin passthru ...passed 00:09:05.468 Test: blockdev copy ...passed 00:09:05.468 Suite: bdevio tests on: Malloc2p7 00:09:05.468 Test: blockdev write read block ...passed 00:09:05.468 Test: blockdev write zeroes read block ...passed 00:09:05.468 Test: blockdev write zeroes read no split ...passed 00:09:05.468 Test: blockdev write zeroes read split ...passed 00:09:05.468 Test: blockdev write zeroes read split partial ...passed 00:09:05.468 Test: blockdev reset ...passed 00:09:05.468 Test: blockdev write read 8 blocks ...passed 00:09:05.468 Test: blockdev write read size > 128k ...passed 00:09:05.468 Test: blockdev write read invalid size ...passed 00:09:05.468 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:09:05.468 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:09:05.468 Test: blockdev write read max offset ...passed 00:09:05.468 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:09:05.468 Test: blockdev writev readv 8 blocks ...passed 00:09:05.468 Test: blockdev writev readv 30 x 1block ...passed 00:09:05.468 Test: blockdev writev readv block ...passed 00:09:05.468 Test: blockdev writev readv size > 128k ...passed 00:09:05.468 Test: blockdev writev readv size > 128k in two iovs ...passed 00:09:05.468 Test: blockdev comparev and writev ...passed 00:09:05.469 Test: blockdev nvme passthru rw ...passed 00:09:05.469 Test: blockdev nvme passthru vendor specific ...passed 00:09:05.469 Test: blockdev nvme admin passthru ...passed 00:09:05.469 Test: blockdev copy ...passed 00:09:05.469 Suite: bdevio tests on: Malloc2p6 00:09:05.469 Test: blockdev write read block ...passed 00:09:05.469 Test: blockdev write zeroes read block ...passed 00:09:05.469 Test: blockdev write zeroes read no split ...passed 00:09:05.469 Test: blockdev write zeroes read split ...passed 00:09:05.469 Test: blockdev write zeroes read split partial ...passed 00:09:05.469 Test: blockdev reset ...passed 00:09:05.469 Test: blockdev write read 8 blocks ...passed 00:09:05.469 Test: blockdev write read size > 128k ...passed 00:09:05.469 Test: blockdev write read invalid size ...passed 00:09:05.469 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:09:05.469 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:09:05.469 Test: blockdev write read max offset ...passed 00:09:05.469 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:09:05.469 Test: blockdev writev readv 8 blocks ...passed 00:09:05.469 Test: blockdev writev readv 30 x 1block ...passed 00:09:05.469 Test: blockdev writev readv block ...passed 00:09:05.469 Test: blockdev writev readv size > 128k ...passed 00:09:05.469 Test: blockdev writev readv size > 128k in two iovs ...passed 00:09:05.469 Test: blockdev comparev and writev ...passed 00:09:05.469 Test: blockdev nvme passthru rw ...passed 00:09:05.469 Test: blockdev nvme passthru vendor specific ...passed 00:09:05.469 Test: blockdev nvme admin passthru ...passed 00:09:05.469 Test: blockdev copy ...passed 00:09:05.469 Suite: bdevio tests on: Malloc2p5 00:09:05.469 Test: blockdev write read block ...passed 00:09:05.469 Test: blockdev write zeroes read block ...passed 00:09:05.469 Test: blockdev write zeroes read no split ...passed 00:09:05.469 Test: blockdev write zeroes read split ...passed 00:09:05.469 Test: blockdev write zeroes read split partial ...passed 00:09:05.469 Test: blockdev reset ...passed 00:09:05.469 Test: blockdev write read 8 blocks ...passed 00:09:05.469 Test: blockdev write read size > 128k ...passed 00:09:05.469 Test: blockdev write read invalid size ...passed 00:09:05.469 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:09:05.469 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:09:05.469 Test: blockdev write read max offset ...passed 00:09:05.469 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:09:05.469 Test: blockdev writev readv 8 blocks ...passed 00:09:05.469 Test: blockdev writev readv 30 x 1block ...passed 00:09:05.469 Test: blockdev writev readv block ...passed 00:09:05.469 Test: blockdev writev readv size > 128k ...passed 00:09:05.469 Test: blockdev writev readv size > 128k in two iovs ...passed 00:09:05.469 Test: blockdev comparev and writev ...passed 00:09:05.469 Test: blockdev nvme passthru rw ...passed 00:09:05.469 Test: blockdev nvme passthru vendor specific ...passed 00:09:05.469 Test: blockdev nvme admin passthru ...passed 00:09:05.469 Test: blockdev copy ...passed 00:09:05.469 Suite: bdevio tests on: Malloc2p4 00:09:05.469 Test: blockdev write read block ...passed 00:09:05.469 Test: blockdev write zeroes read block ...passed 00:09:05.469 Test: blockdev write zeroes read no split ...passed 00:09:05.469 Test: blockdev write zeroes read split ...passed 00:09:05.469 Test: blockdev write zeroes read split partial ...passed 00:09:05.469 Test: blockdev reset ...passed 00:09:05.469 Test: blockdev write read 8 blocks ...passed 00:09:05.469 Test: blockdev write read size > 128k ...passed 00:09:05.469 Test: blockdev write read invalid size ...passed 00:09:05.469 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:09:05.469 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:09:05.469 Test: blockdev write read max offset ...passed 00:09:05.469 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:09:05.469 Test: blockdev writev readv 8 blocks ...passed 00:09:05.469 Test: blockdev writev readv 30 x 1block ...passed 00:09:05.469 Test: blockdev writev readv block ...passed 00:09:05.469 Test: blockdev writev readv size > 128k ...passed 00:09:05.469 Test: blockdev writev readv size > 128k in two iovs ...passed 00:09:05.469 Test: blockdev comparev and writev ...passed 00:09:05.469 Test: blockdev nvme passthru rw ...passed 00:09:05.469 Test: blockdev nvme passthru vendor specific ...passed 00:09:05.469 Test: blockdev nvme admin passthru ...passed 00:09:05.469 Test: blockdev copy ...passed 00:09:05.469 Suite: bdevio tests on: Malloc2p3 00:09:05.469 Test: blockdev write read block ...passed 00:09:05.469 Test: blockdev write zeroes read block ...passed 00:09:05.469 Test: blockdev write zeroes read no split ...passed 00:09:05.469 Test: blockdev write zeroes read split ...passed 00:09:05.469 Test: blockdev write zeroes read split partial ...passed 00:09:05.469 Test: blockdev reset ...passed 00:09:05.469 Test: blockdev write read 8 blocks ...passed 00:09:05.469 Test: blockdev write read size > 128k ...passed 00:09:05.469 Test: blockdev write read invalid size ...passed 00:09:05.469 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:09:05.469 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:09:05.469 Test: blockdev write read max offset ...passed 00:09:05.469 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:09:05.469 Test: blockdev writev readv 8 blocks ...passed 00:09:05.469 Test: blockdev writev readv 30 x 1block ...passed 00:09:05.469 Test: blockdev writev readv block ...passed 00:09:05.469 Test: blockdev writev readv size > 128k ...passed 00:09:05.469 Test: blockdev writev readv size > 128k in two iovs ...passed 00:09:05.469 Test: blockdev comparev and writev ...passed 00:09:05.469 Test: blockdev nvme passthru rw ...passed 00:09:05.469 Test: blockdev nvme passthru vendor specific ...passed 00:09:05.469 Test: blockdev nvme admin passthru ...passed 00:09:05.469 Test: blockdev copy ...passed 00:09:05.469 Suite: bdevio tests on: Malloc2p2 00:09:05.469 Test: blockdev write read block ...passed 00:09:05.469 Test: blockdev write zeroes read block ...passed 00:09:05.469 Test: blockdev write zeroes read no split ...passed 00:09:05.469 Test: blockdev write zeroes read split ...passed 00:09:05.469 Test: blockdev write zeroes read split partial ...passed 00:09:05.469 Test: blockdev reset ...passed 00:09:05.469 Test: blockdev write read 8 blocks ...passed 00:09:05.469 Test: blockdev write read size > 128k ...passed 00:09:05.469 Test: blockdev write read invalid size ...passed 00:09:05.469 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:09:05.469 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:09:05.469 Test: blockdev write read max offset ...passed 00:09:05.469 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:09:05.469 Test: blockdev writev readv 8 blocks ...passed 00:09:05.469 Test: blockdev writev readv 30 x 1block ...passed 00:09:05.469 Test: blockdev writev readv block ...passed 00:09:05.469 Test: blockdev writev readv size > 128k ...passed 00:09:05.469 Test: blockdev writev readv size > 128k in two iovs ...passed 00:09:05.469 Test: blockdev comparev and writev ...passed 00:09:05.469 Test: blockdev nvme passthru rw ...passed 00:09:05.469 Test: blockdev nvme passthru vendor specific ...passed 00:09:05.469 Test: blockdev nvme admin passthru ...passed 00:09:05.469 Test: blockdev copy ...passed 00:09:05.469 Suite: bdevio tests on: Malloc2p1 00:09:05.469 Test: blockdev write read block ...passed 00:09:05.469 Test: blockdev write zeroes read block ...passed 00:09:05.469 Test: blockdev write zeroes read no split ...passed 00:09:05.469 Test: blockdev write zeroes read split ...passed 00:09:05.469 Test: blockdev write zeroes read split partial ...passed 00:09:05.469 Test: blockdev reset ...passed 00:09:05.469 Test: blockdev write read 8 blocks ...passed 00:09:05.469 Test: blockdev write read size > 128k ...passed 00:09:05.469 Test: blockdev write read invalid size ...passed 00:09:05.469 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:09:05.469 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:09:05.469 Test: blockdev write read max offset ...passed 00:09:05.469 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:09:05.469 Test: blockdev writev readv 8 blocks ...passed 00:09:05.469 Test: blockdev writev readv 30 x 1block ...passed 00:09:05.469 Test: blockdev writev readv block ...passed 00:09:05.469 Test: blockdev writev readv size > 128k ...passed 00:09:05.469 Test: blockdev writev readv size > 128k in two iovs ...passed 00:09:05.469 Test: blockdev comparev and writev ...passed 00:09:05.469 Test: blockdev nvme passthru rw ...passed 00:09:05.469 Test: blockdev nvme passthru vendor specific ...passed 00:09:05.469 Test: blockdev nvme admin passthru ...passed 00:09:05.469 Test: blockdev copy ...passed 00:09:05.469 Suite: bdevio tests on: Malloc2p0 00:09:05.469 Test: blockdev write read block ...passed 00:09:05.469 Test: blockdev write zeroes read block ...passed 00:09:05.469 Test: blockdev write zeroes read no split ...passed 00:09:05.469 Test: blockdev write zeroes read split ...passed 00:09:05.469 Test: blockdev write zeroes read split partial ...passed 00:09:05.469 Test: blockdev reset ...passed 00:09:05.469 Test: blockdev write read 8 blocks ...passed 00:09:05.469 Test: blockdev write read size > 128k ...passed 00:09:05.469 Test: blockdev write read invalid size ...passed 00:09:05.469 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:09:05.469 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:09:05.469 Test: blockdev write read max offset ...passed 00:09:05.469 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:09:05.469 Test: blockdev writev readv 8 blocks ...passed 00:09:05.469 Test: blockdev writev readv 30 x 1block ...passed 00:09:05.469 Test: blockdev writev readv block ...passed 00:09:05.469 Test: blockdev writev readv size > 128k ...passed 00:09:05.469 Test: blockdev writev readv size > 128k in two iovs ...passed 00:09:05.469 Test: blockdev comparev and writev ...passed 00:09:05.469 Test: blockdev nvme passthru rw ...passed 00:09:05.469 Test: blockdev nvme passthru vendor specific ...passed 00:09:05.469 Test: blockdev nvme admin passthru ...passed 00:09:05.469 Test: blockdev copy ...passed 00:09:05.469 Suite: bdevio tests on: Malloc1p1 00:09:05.469 Test: blockdev write read block ...passed 00:09:05.469 Test: blockdev write zeroes read block ...passed 00:09:05.469 Test: blockdev write zeroes read no split ...passed 00:09:05.469 Test: blockdev write zeroes read split ...passed 00:09:05.469 Test: blockdev write zeroes read split partial ...passed 00:09:05.469 Test: blockdev reset ...passed 00:09:05.469 Test: blockdev write read 8 blocks ...passed 00:09:05.469 Test: blockdev write read size > 128k ...passed 00:09:05.469 Test: blockdev write read invalid size ...passed 00:09:05.469 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:09:05.469 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:09:05.469 Test: blockdev write read max offset ...passed 00:09:05.469 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:09:05.469 Test: blockdev writev readv 8 blocks ...passed 00:09:05.469 Test: blockdev writev readv 30 x 1block ...passed 00:09:05.469 Test: blockdev writev readv block ...passed 00:09:05.469 Test: blockdev writev readv size > 128k ...passed 00:09:05.469 Test: blockdev writev readv size > 128k in two iovs ...passed 00:09:05.469 Test: blockdev comparev and writev ...passed 00:09:05.469 Test: blockdev nvme passthru rw ...passed 00:09:05.469 Test: blockdev nvme passthru vendor specific ...passed 00:09:05.469 Test: blockdev nvme admin passthru ...passed 00:09:05.469 Test: blockdev copy ...passed 00:09:05.469 Suite: bdevio tests on: Malloc1p0 00:09:05.469 Test: blockdev write read block ...passed 00:09:05.469 Test: blockdev write zeroes read block ...passed 00:09:05.469 Test: blockdev write zeroes read no split ...passed 00:09:05.469 Test: blockdev write zeroes read split ...passed 00:09:05.469 Test: blockdev write zeroes read split partial ...passed 00:09:05.469 Test: blockdev reset ...passed 00:09:05.469 Test: blockdev write read 8 blocks ...passed 00:09:05.470 Test: blockdev write read size > 128k ...passed 00:09:05.470 Test: blockdev write read invalid size ...passed 00:09:05.470 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:09:05.470 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:09:05.470 Test: blockdev write read max offset ...passed 00:09:05.470 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:09:05.470 Test: blockdev writev readv 8 blocks ...passed 00:09:05.470 Test: blockdev writev readv 30 x 1block ...passed 00:09:05.470 Test: blockdev writev readv block ...passed 00:09:05.470 Test: blockdev writev readv size > 128k ...passed 00:09:05.470 Test: blockdev writev readv size > 128k in two iovs ...passed 00:09:05.470 Test: blockdev comparev and writev ...passed 00:09:05.470 Test: blockdev nvme passthru rw ...passed 00:09:05.470 Test: blockdev nvme passthru vendor specific ...passed 00:09:05.470 Test: blockdev nvme admin passthru ...passed 00:09:05.470 Test: blockdev copy ...passed 00:09:05.470 Suite: bdevio tests on: Malloc0 00:09:05.470 Test: blockdev write read block ...passed 00:09:05.470 Test: blockdev write zeroes read block ...passed 00:09:05.470 Test: blockdev write zeroes read no split ...passed 00:09:05.727 Test: blockdev write zeroes read split ...passed 00:09:05.727 Test: blockdev write zeroes read split partial ...passed 00:09:05.727 Test: blockdev reset ...passed 00:09:05.727 Test: blockdev write read 8 blocks ...passed 00:09:05.727 Test: blockdev write read size > 128k ...passed 00:09:05.727 Test: blockdev write read invalid size ...passed 00:09:05.727 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:09:05.727 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:09:05.727 Test: blockdev write read max offset ...passed 00:09:05.727 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:09:05.727 Test: blockdev writev readv 8 blocks ...passed 00:09:05.727 Test: blockdev writev readv 30 x 1block ...passed 00:09:05.727 Test: blockdev writev readv block ...passed 00:09:05.727 Test: blockdev writev readv size > 128k ...passed 00:09:05.727 Test: blockdev writev readv size > 128k in two iovs ...passed 00:09:05.727 Test: blockdev comparev and writev ...passed 00:09:05.727 Test: blockdev nvme passthru rw ...passed 00:09:05.727 Test: blockdev nvme passthru vendor specific ...passed 00:09:05.727 Test: blockdev nvme admin passthru ...passed 00:09:05.727 Test: blockdev copy ...passed 00:09:05.727 00:09:05.727 Run Summary: Type Total Ran Passed Failed Inactive 00:09:05.727 suites 16 16 n/a 0 0 00:09:05.727 tests 368 368 368 0 0 00:09:05.727 asserts 2224 2224 2224 0 n/a 00:09:05.727 00:09:05.727 Elapsed time = 0.498 seconds 00:09:05.727 0 00:09:05.727 11:19:49 blockdev_general.bdev_bounds -- bdev/blockdev.sh@295 -- # killprocess 843683 00:09:05.727 11:19:49 blockdev_general.bdev_bounds -- common/autotest_common.sh@948 -- # '[' -z 843683 ']' 00:09:05.727 11:19:49 blockdev_general.bdev_bounds -- common/autotest_common.sh@952 -- # kill -0 843683 00:09:05.727 11:19:49 blockdev_general.bdev_bounds -- common/autotest_common.sh@953 -- # uname 00:09:05.727 11:19:49 blockdev_general.bdev_bounds -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:09:05.727 11:19:49 blockdev_general.bdev_bounds -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 843683 00:09:05.727 11:19:49 blockdev_general.bdev_bounds -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:09:05.727 11:19:49 blockdev_general.bdev_bounds -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:09:05.727 11:19:49 blockdev_general.bdev_bounds -- common/autotest_common.sh@966 -- # echo 'killing process with pid 843683' 00:09:05.727 killing process with pid 843683 00:09:05.727 11:19:49 blockdev_general.bdev_bounds -- common/autotest_common.sh@967 -- # kill 843683 00:09:05.727 11:19:49 blockdev_general.bdev_bounds -- common/autotest_common.sh@972 -- # wait 843683 00:09:05.985 11:19:49 blockdev_general.bdev_bounds -- bdev/blockdev.sh@296 -- # trap - SIGINT SIGTERM EXIT 00:09:05.985 00:09:05.985 real 0m1.433s 00:09:05.985 user 0m3.452s 00:09:05.985 sys 0m0.493s 00:09:05.985 11:19:49 blockdev_general.bdev_bounds -- common/autotest_common.sh@1124 -- # xtrace_disable 00:09:05.985 11:19:49 blockdev_general.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:09:05.985 ************************************ 00:09:05.985 END TEST bdev_bounds 00:09:05.985 ************************************ 00:09:05.985 11:19:49 blockdev_general -- common/autotest_common.sh@1142 -- # return 0 00:09:05.985 11:19:49 blockdev_general -- bdev/blockdev.sh@762 -- # run_test bdev_nbd nbd_function_test /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json 'Malloc0 Malloc1p0 Malloc1p1 Malloc2p0 Malloc2p1 Malloc2p2 Malloc2p3 Malloc2p4 Malloc2p5 Malloc2p6 Malloc2p7 TestPT raid0 concat0 raid1 AIO0' '' 00:09:05.985 11:19:49 blockdev_general -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:09:05.985 11:19:49 blockdev_general -- common/autotest_common.sh@1105 -- # xtrace_disable 00:09:05.985 11:19:49 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:09:05.985 ************************************ 00:09:05.985 START TEST bdev_nbd 00:09:05.985 ************************************ 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- common/autotest_common.sh@1123 -- # nbd_function_test /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json 'Malloc0 Malloc1p0 Malloc1p1 Malloc2p0 Malloc2p1 Malloc2p2 Malloc2p3 Malloc2p4 Malloc2p5 Malloc2p6 Malloc2p7 TestPT raid0 concat0 raid1 AIO0' '' 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- bdev/blockdev.sh@300 -- # uname -s 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- bdev/blockdev.sh@300 -- # [[ Linux == Linux ]] 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- bdev/blockdev.sh@302 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- bdev/blockdev.sh@303 -- # local conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- bdev/blockdev.sh@304 -- # bdev_all=('Malloc0' 'Malloc1p0' 'Malloc1p1' 'Malloc2p0' 'Malloc2p1' 'Malloc2p2' 'Malloc2p3' 'Malloc2p4' 'Malloc2p5' 'Malloc2p6' 'Malloc2p7' 'TestPT' 'raid0' 'concat0' 'raid1' 'AIO0') 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- bdev/blockdev.sh@304 -- # local bdev_all 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- bdev/blockdev.sh@305 -- # local bdev_num=16 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- bdev/blockdev.sh@309 -- # [[ -e /sys/module/nbd ]] 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- bdev/blockdev.sh@311 -- # nbd_all=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- bdev/blockdev.sh@311 -- # local nbd_all 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- bdev/blockdev.sh@312 -- # bdev_num=16 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- bdev/blockdev.sh@314 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- bdev/blockdev.sh@314 -- # local nbd_list 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- bdev/blockdev.sh@315 -- # bdev_list=('Malloc0' 'Malloc1p0' 'Malloc1p1' 'Malloc2p0' 'Malloc2p1' 'Malloc2p2' 'Malloc2p3' 'Malloc2p4' 'Malloc2p5' 'Malloc2p6' 'Malloc2p7' 'TestPT' 'raid0' 'concat0' 'raid1' 'AIO0') 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- bdev/blockdev.sh@315 -- # local bdev_list 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- bdev/blockdev.sh@318 -- # nbd_pid=843889 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- bdev/blockdev.sh@319 -- # trap 'cleanup; killprocess $nbd_pid' SIGINT SIGTERM EXIT 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- bdev/blockdev.sh@317 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-nbd.sock -i 0 --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json '' 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- bdev/blockdev.sh@320 -- # waitforlisten 843889 /var/tmp/spdk-nbd.sock 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- common/autotest_common.sh@829 -- # '[' -z 843889 ']' 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- common/autotest_common.sh@834 -- # local max_retries=100 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:09:05.985 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- common/autotest_common.sh@838 -- # xtrace_disable 00:09:05.985 11:19:49 blockdev_general.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:09:06.243 [2024-07-15 11:19:49.604489] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:09:06.243 [2024-07-15 11:19:49.604559] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:06.243 [2024-07-15 11:19:49.728909] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:06.243 [2024-07-15 11:19:49.835035] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:09:06.500 [2024-07-15 11:19:49.999732] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc3 00:09:06.500 [2024-07-15 11:19:49.999790] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc3 00:09:06.500 [2024-07-15 11:19:49.999809] vbdev_passthru.c: 735:bdev_passthru_create_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:09:06.500 [2024-07-15 11:19:50.007738] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:09:06.500 [2024-07-15 11:19:50.007765] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:09:06.500 [2024-07-15 11:19:50.015750] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:09:06.500 [2024-07-15 11:19:50.015774] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:09:06.500 [2024-07-15 11:19:50.093451] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc3 00:09:06.500 [2024-07-15 11:19:50.093506] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:06.500 [2024-07-15 11:19:50.093522] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x13bea40 00:09:06.500 [2024-07-15 11:19:50.093535] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:06.500 [2024-07-15 11:19:50.094990] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:06.500 [2024-07-15 11:19:50.095021] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: TestPT 00:09:07.065 11:19:50 blockdev_general.bdev_nbd -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:09:07.065 11:19:50 blockdev_general.bdev_nbd -- common/autotest_common.sh@862 -- # return 0 00:09:07.065 11:19:50 blockdev_general.bdev_nbd -- bdev/blockdev.sh@322 -- # nbd_rpc_start_stop_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1p0 Malloc1p1 Malloc2p0 Malloc2p1 Malloc2p2 Malloc2p3 Malloc2p4 Malloc2p5 Malloc2p6 Malloc2p7 TestPT raid0 concat0 raid1 AIO0' 00:09:07.065 11:19:50 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@113 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:07.065 11:19:50 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@114 -- # bdev_list=('Malloc0' 'Malloc1p0' 'Malloc1p1' 'Malloc2p0' 'Malloc2p1' 'Malloc2p2' 'Malloc2p3' 'Malloc2p4' 'Malloc2p5' 'Malloc2p6' 'Malloc2p7' 'TestPT' 'raid0' 'concat0' 'raid1' 'AIO0') 00:09:07.065 11:19:50 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@114 -- # local bdev_list 00:09:07.065 11:19:50 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@116 -- # nbd_start_disks_without_nbd_idx /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1p0 Malloc1p1 Malloc2p0 Malloc2p1 Malloc2p2 Malloc2p3 Malloc2p4 Malloc2p5 Malloc2p6 Malloc2p7 TestPT raid0 concat0 raid1 AIO0' 00:09:07.065 11:19:50 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@22 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:07.065 11:19:50 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@23 -- # bdev_list=('Malloc0' 'Malloc1p0' 'Malloc1p1' 'Malloc2p0' 'Malloc2p1' 'Malloc2p2' 'Malloc2p3' 'Malloc2p4' 'Malloc2p5' 'Malloc2p6' 'Malloc2p7' 'TestPT' 'raid0' 'concat0' 'raid1' 'AIO0') 00:09:07.065 11:19:50 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@23 -- # local bdev_list 00:09:07.065 11:19:50 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@24 -- # local i 00:09:07.065 11:19:50 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@25 -- # local nbd_device 00:09:07.065 11:19:50 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i = 0 )) 00:09:07.065 11:19:50 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 16 )) 00:09:07.065 11:19:50 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 00:09:07.322 11:19:50 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd0 00:09:07.322 11:19:50 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd0 00:09:07.322 11:19:50 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd0 00:09:07.322 11:19:50 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:09:07.322 11:19:50 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:07.322 11:19:50 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:07.322 11:19:50 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:07.322 11:19:50 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:09:07.322 11:19:50 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:07.322 11:19:50 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:07.322 11:19:50 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:07.322 11:19:50 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:07.322 1+0 records in 00:09:07.322 1+0 records out 00:09:07.322 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00020959 s, 19.5 MB/s 00:09:07.322 11:19:50 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:07.322 11:19:50 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:07.322 11:19:50 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:07.322 11:19:50 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:07.322 11:19:50 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:07.322 11:19:50 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:09:07.322 11:19:50 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 16 )) 00:09:07.322 11:19:50 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1p0 00:09:07.580 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd1 00:09:07.580 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd1 00:09:07.580 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd1 00:09:07.580 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:09:07.580 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:07.580 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:07.580 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:07.580 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:09:07.580 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:07.580 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:07.580 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:07.580 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:07.580 1+0 records in 00:09:07.580 1+0 records out 00:09:07.580 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000318979 s, 12.8 MB/s 00:09:07.580 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:07.580 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:07.580 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:07.580 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:07.580 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:07.580 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:09:07.580 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 16 )) 00:09:07.580 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1p1 00:09:07.837 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd2 00:09:07.837 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd2 00:09:07.837 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd2 00:09:07.837 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd2 00:09:07.837 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:07.837 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:07.837 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:07.837 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd2 /proc/partitions 00:09:07.837 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:07.837 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:07.837 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:07.837 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd2 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:07.837 1+0 records in 00:09:07.837 1+0 records out 00:09:07.837 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000262959 s, 15.6 MB/s 00:09:07.837 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:07.837 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:07.837 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:07.837 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:07.837 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:07.837 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:09:07.837 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 16 )) 00:09:07.837 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc2p0 00:09:08.094 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd3 00:09:08.094 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd3 00:09:08.094 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd3 00:09:08.094 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd3 00:09:08.094 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:08.094 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:08.094 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:08.094 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd3 /proc/partitions 00:09:08.094 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:08.094 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:08.094 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:08.094 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd3 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:08.094 1+0 records in 00:09:08.094 1+0 records out 00:09:08.094 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00033592 s, 12.2 MB/s 00:09:08.094 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:08.094 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:08.094 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:08.094 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:08.094 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:08.094 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:09:08.094 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 16 )) 00:09:08.094 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc2p1 00:09:08.351 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd4 00:09:08.351 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd4 00:09:08.351 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd4 00:09:08.351 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd4 00:09:08.351 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:08.351 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:08.351 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:08.351 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd4 /proc/partitions 00:09:08.351 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:08.351 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:08.351 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:08.351 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd4 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:08.351 1+0 records in 00:09:08.351 1+0 records out 00:09:08.351 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000306094 s, 13.4 MB/s 00:09:08.351 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:08.351 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:08.351 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:08.351 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:08.351 11:19:51 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:08.351 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:09:08.351 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 16 )) 00:09:08.351 11:19:51 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc2p2 00:09:08.609 11:19:52 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd5 00:09:08.609 11:19:52 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd5 00:09:08.609 11:19:52 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd5 00:09:08.609 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd5 00:09:08.609 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:08.609 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:08.609 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:08.609 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd5 /proc/partitions 00:09:08.609 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:08.609 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:08.609 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:08.609 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd5 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:08.609 1+0 records in 00:09:08.609 1+0 records out 00:09:08.609 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000370162 s, 11.1 MB/s 00:09:08.609 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:08.609 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:08.609 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:08.609 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:08.609 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:08.609 11:19:52 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:09:08.609 11:19:52 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 16 )) 00:09:08.609 11:19:52 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc2p3 00:09:08.865 11:19:52 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd6 00:09:08.865 11:19:52 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd6 00:09:08.865 11:19:52 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd6 00:09:08.865 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd6 00:09:08.865 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:08.865 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:08.865 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:08.865 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd6 /proc/partitions 00:09:08.865 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:08.865 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:08.865 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:08.865 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd6 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:08.865 1+0 records in 00:09:08.865 1+0 records out 00:09:08.865 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000343994 s, 11.9 MB/s 00:09:08.865 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:09.122 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:09.122 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:09.122 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:09.122 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:09.122 11:19:52 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:09:09.122 11:19:52 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 16 )) 00:09:09.122 11:19:52 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc2p4 00:09:09.379 11:19:52 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd7 00:09:09.379 11:19:52 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd7 00:09:09.379 11:19:52 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd7 00:09:09.379 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd7 00:09:09.379 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:09.379 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:09.379 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:09.379 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd7 /proc/partitions 00:09:09.379 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:09.379 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:09.379 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:09.379 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd7 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:09.379 1+0 records in 00:09:09.379 1+0 records out 00:09:09.379 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000385214 s, 10.6 MB/s 00:09:09.379 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:09.379 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:09.379 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:09.379 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:09.379 11:19:52 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:09.379 11:19:52 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:09:09.379 11:19:52 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 16 )) 00:09:09.379 11:19:52 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc2p5 00:09:09.636 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd8 00:09:09.636 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd8 00:09:09.636 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd8 00:09:09.636 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd8 00:09:09.636 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:09.636 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:09.636 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:09.636 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd8 /proc/partitions 00:09:09.636 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:09.636 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:09.636 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:09.636 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd8 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:09.636 1+0 records in 00:09:09.636 1+0 records out 00:09:09.637 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000526454 s, 7.8 MB/s 00:09:09.637 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:09.637 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:09.637 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:09.637 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:09.637 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:09.637 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:09:09.637 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 16 )) 00:09:09.637 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc2p6 00:09:09.893 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd9 00:09:09.893 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd9 00:09:09.893 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd9 00:09:09.893 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd9 00:09:09.893 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:09.893 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:09.893 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:09.893 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd9 /proc/partitions 00:09:09.893 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:09.893 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:09.893 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:09.893 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd9 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:09.893 1+0 records in 00:09:09.893 1+0 records out 00:09:09.893 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000602738 s, 6.8 MB/s 00:09:09.893 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:09.893 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:09.893 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:09.893 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:09.893 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:09.893 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:09:09.893 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 16 )) 00:09:09.893 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc2p7 00:09:10.150 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd10 00:09:10.150 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd10 00:09:10.150 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd10 00:09:10.150 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd10 00:09:10.150 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:10.150 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:10.150 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:10.150 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd10 /proc/partitions 00:09:10.150 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:10.150 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:10.150 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:10.150 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd10 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:10.150 1+0 records in 00:09:10.150 1+0 records out 00:09:10.150 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000489762 s, 8.4 MB/s 00:09:10.150 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:10.150 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:10.150 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:10.150 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:10.150 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:10.150 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:09:10.150 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 16 )) 00:09:10.150 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk TestPT 00:09:10.407 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd11 00:09:10.407 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd11 00:09:10.407 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd11 00:09:10.407 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd11 00:09:10.407 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:10.407 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:10.407 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:10.407 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd11 /proc/partitions 00:09:10.407 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:10.407 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:10.407 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:10.407 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd11 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:10.407 1+0 records in 00:09:10.407 1+0 records out 00:09:10.407 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000656694 s, 6.2 MB/s 00:09:10.407 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:10.407 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:10.407 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:10.407 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:10.407 11:19:53 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:10.407 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:09:10.407 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 16 )) 00:09:10.407 11:19:53 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid0 00:09:10.663 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd12 00:09:10.663 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd12 00:09:10.663 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd12 00:09:10.663 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd12 00:09:10.663 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:10.663 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:10.663 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:10.663 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd12 /proc/partitions 00:09:10.663 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:10.663 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:10.663 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:10.663 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd12 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:10.663 1+0 records in 00:09:10.663 1+0 records out 00:09:10.663 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000749622 s, 5.5 MB/s 00:09:10.663 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:10.663 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:10.663 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:10.663 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:10.663 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:10.663 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:09:10.663 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 16 )) 00:09:10.664 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk concat0 00:09:10.921 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd13 00:09:10.921 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd13 00:09:10.921 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd13 00:09:10.921 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd13 00:09:10.921 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:10.921 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:10.921 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:10.921 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd13 /proc/partitions 00:09:10.921 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:10.921 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:10.921 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:10.921 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd13 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:10.921 1+0 records in 00:09:10.921 1+0 records out 00:09:10.921 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00067267 s, 6.1 MB/s 00:09:10.921 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:10.921 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:10.921 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:10.921 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:10.921 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:10.921 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:09:10.921 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 16 )) 00:09:10.921 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid1 00:09:11.179 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd14 00:09:11.179 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd14 00:09:11.179 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd14 00:09:11.179 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd14 00:09:11.179 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:11.179 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:11.179 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:11.179 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd14 /proc/partitions 00:09:11.179 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:11.179 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:11.179 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:11.179 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd14 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:11.179 1+0 records in 00:09:11.179 1+0 records out 00:09:11.179 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00075891 s, 5.4 MB/s 00:09:11.179 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:11.179 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:11.179 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:11.179 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:11.179 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:11.179 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:09:11.179 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 16 )) 00:09:11.179 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk AIO0 00:09:11.436 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd15 00:09:11.436 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd15 00:09:11.436 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd15 00:09:11.436 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd15 00:09:11.436 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:11.436 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:11.436 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:11.436 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd15 /proc/partitions 00:09:11.436 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:11.436 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:11.436 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:11.436 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd15 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:11.436 1+0 records in 00:09:11.436 1+0 records out 00:09:11.436 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00064403 s, 6.4 MB/s 00:09:11.436 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:11.437 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:11.437 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:11.437 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:11.437 11:19:54 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:11.437 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:09:11.437 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 16 )) 00:09:11.437 11:19:54 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@118 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:09:11.693 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@118 -- # nbd_disks_json='[ 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd0", 00:09:11.693 "bdev_name": "Malloc0" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd1", 00:09:11.693 "bdev_name": "Malloc1p0" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd2", 00:09:11.693 "bdev_name": "Malloc1p1" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd3", 00:09:11.693 "bdev_name": "Malloc2p0" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd4", 00:09:11.693 "bdev_name": "Malloc2p1" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd5", 00:09:11.693 "bdev_name": "Malloc2p2" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd6", 00:09:11.693 "bdev_name": "Malloc2p3" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd7", 00:09:11.693 "bdev_name": "Malloc2p4" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd8", 00:09:11.693 "bdev_name": "Malloc2p5" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd9", 00:09:11.693 "bdev_name": "Malloc2p6" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd10", 00:09:11.693 "bdev_name": "Malloc2p7" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd11", 00:09:11.693 "bdev_name": "TestPT" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd12", 00:09:11.693 "bdev_name": "raid0" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd13", 00:09:11.693 "bdev_name": "concat0" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd14", 00:09:11.693 "bdev_name": "raid1" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd15", 00:09:11.693 "bdev_name": "AIO0" 00:09:11.693 } 00:09:11.693 ]' 00:09:11.693 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@119 -- # nbd_disks_name=($(echo "${nbd_disks_json}" | jq -r '.[] | .nbd_device')) 00:09:11.693 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@119 -- # echo '[ 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd0", 00:09:11.693 "bdev_name": "Malloc0" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd1", 00:09:11.693 "bdev_name": "Malloc1p0" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd2", 00:09:11.693 "bdev_name": "Malloc1p1" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd3", 00:09:11.693 "bdev_name": "Malloc2p0" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd4", 00:09:11.693 "bdev_name": "Malloc2p1" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd5", 00:09:11.693 "bdev_name": "Malloc2p2" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd6", 00:09:11.693 "bdev_name": "Malloc2p3" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd7", 00:09:11.693 "bdev_name": "Malloc2p4" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd8", 00:09:11.693 "bdev_name": "Malloc2p5" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd9", 00:09:11.693 "bdev_name": "Malloc2p6" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd10", 00:09:11.693 "bdev_name": "Malloc2p7" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd11", 00:09:11.693 "bdev_name": "TestPT" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd12", 00:09:11.693 "bdev_name": "raid0" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd13", 00:09:11.693 "bdev_name": "concat0" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd14", 00:09:11.693 "bdev_name": "raid1" 00:09:11.693 }, 00:09:11.693 { 00:09:11.693 "nbd_device": "/dev/nbd15", 00:09:11.693 "bdev_name": "AIO0" 00:09:11.693 } 00:09:11.693 ]' 00:09:11.693 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@119 -- # jq -r '.[] | .nbd_device' 00:09:11.693 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@120 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1 /dev/nbd2 /dev/nbd3 /dev/nbd4 /dev/nbd5 /dev/nbd6 /dev/nbd7 /dev/nbd8 /dev/nbd9 /dev/nbd10 /dev/nbd11 /dev/nbd12 /dev/nbd13 /dev/nbd14 /dev/nbd15' 00:09:11.693 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:11.693 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15') 00:09:11.693 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:09:11.693 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:09:11.693 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:11.694 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:09:11.950 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:09:11.950 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:09:11.950 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:09:11.950 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:11.950 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:11.950 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:09:11.950 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:11.950 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:11.950 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:11.950 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:09:12.206 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:09:12.206 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:09:12.206 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:09:12.206 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:12.206 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:12.206 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:09:12.206 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:12.206 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:12.206 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:12.206 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd2 00:09:12.463 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd2 00:09:12.463 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd2 00:09:12.463 11:19:55 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd2 00:09:12.463 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:12.463 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:12.463 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd2 /proc/partitions 00:09:12.463 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:12.463 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:12.463 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:12.463 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd3 00:09:12.720 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd3 00:09:12.720 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd3 00:09:12.720 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd3 00:09:12.720 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:12.720 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:12.720 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd3 /proc/partitions 00:09:12.720 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:12.720 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:12.720 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:12.720 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd4 00:09:12.977 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd4 00:09:12.977 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd4 00:09:12.977 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd4 00:09:12.977 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:12.977 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:12.977 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd4 /proc/partitions 00:09:12.977 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:12.977 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:12.977 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:12.977 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd5 00:09:13.233 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd5 00:09:13.233 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd5 00:09:13.233 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd5 00:09:13.233 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:13.233 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:13.233 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd5 /proc/partitions 00:09:13.233 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:13.233 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:13.233 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:13.233 11:19:56 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd6 00:09:13.795 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd6 00:09:13.796 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd6 00:09:13.796 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd6 00:09:13.796 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:13.796 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:13.796 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd6 /proc/partitions 00:09:13.796 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:13.796 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:13.796 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:13.796 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd7 00:09:13.796 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd7 00:09:13.796 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd7 00:09:13.796 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd7 00:09:13.796 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:13.796 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:13.796 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd7 /proc/partitions 00:09:13.796 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:13.796 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:13.796 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:13.796 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd8 00:09:14.052 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd8 00:09:14.052 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd8 00:09:14.309 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd8 00:09:14.309 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:14.309 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:14.309 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd8 /proc/partitions 00:09:14.309 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:14.309 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:14.309 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:14.309 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd9 00:09:14.566 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd9 00:09:14.566 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd9 00:09:14.566 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd9 00:09:14.566 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:14.566 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:14.566 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd9 /proc/partitions 00:09:14.566 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:14.566 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:14.566 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:14.566 11:19:57 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd10 00:09:14.823 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd10 00:09:14.823 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd10 00:09:14.823 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd10 00:09:14.823 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:14.823 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:14.823 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd10 /proc/partitions 00:09:14.823 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:14.823 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:14.823 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:14.823 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd11 00:09:15.081 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd11 00:09:15.081 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd11 00:09:15.082 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd11 00:09:15.082 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:15.082 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:15.082 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd11 /proc/partitions 00:09:15.082 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:15.082 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:15.082 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:15.082 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd12 00:09:15.082 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd12 00:09:15.082 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd12 00:09:15.082 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd12 00:09:15.082 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:15.082 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:15.082 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd12 /proc/partitions 00:09:15.340 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:15.340 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:15.340 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:15.340 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd13 00:09:15.598 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd13 00:09:15.598 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd13 00:09:15.598 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd13 00:09:15.598 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:15.598 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:15.598 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd13 /proc/partitions 00:09:15.598 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:15.598 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:15.598 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:15.598 11:19:58 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd14 00:09:15.856 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd14 00:09:15.856 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd14 00:09:15.856 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd14 00:09:15.856 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:15.856 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:15.856 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd14 /proc/partitions 00:09:15.856 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:15.856 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:15.856 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:15.856 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd15 00:09:16.114 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd15 00:09:16.114 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd15 00:09:16.114 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd15 00:09:16.114 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:16.114 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:16.114 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd15 /proc/partitions 00:09:16.114 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@39 -- # sleep 0.1 00:09:16.114 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i++ )) 00:09:16.114 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:16.114 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd15 /proc/partitions 00:09:16.114 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:16.114 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:16.114 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@122 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:09:16.114 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:16.114 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:09:16.372 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:09:16.372 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:09:16.372 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:09:16.631 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:09:16.631 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:09:16.631 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:09:16.631 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:09:16.631 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:09:16.631 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:09:16.631 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@122 -- # count=0 00:09:16.631 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@123 -- # '[' 0 -ne 0 ']' 00:09:16.631 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@127 -- # return 0 00:09:16.631 11:19:59 blockdev_general.bdev_nbd -- bdev/blockdev.sh@323 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1p0 Malloc1p1 Malloc2p0 Malloc2p1 Malloc2p2 Malloc2p3 Malloc2p4 Malloc2p5 Malloc2p6 Malloc2p7 TestPT raid0 concat0 raid1 AIO0' '/dev/nbd0 /dev/nbd1 /dev/nbd10 /dev/nbd11 /dev/nbd12 /dev/nbd13 /dev/nbd14 /dev/nbd15 /dev/nbd2 /dev/nbd3 /dev/nbd4 /dev/nbd5 /dev/nbd6 /dev/nbd7 /dev/nbd8 /dev/nbd9' 00:09:16.631 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:16.631 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1p0' 'Malloc1p1' 'Malloc2p0' 'Malloc2p1' 'Malloc2p2' 'Malloc2p3' 'Malloc2p4' 'Malloc2p5' 'Malloc2p6' 'Malloc2p7' 'TestPT' 'raid0' 'concat0' 'raid1' 'AIO0') 00:09:16.631 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@91 -- # local bdev_list 00:09:16.631 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:09:16.631 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@92 -- # local nbd_list 00:09:16.631 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1p0 Malloc1p1 Malloc2p0 Malloc2p1 Malloc2p2 Malloc2p3 Malloc2p4 Malloc2p5 Malloc2p6 Malloc2p7 TestPT raid0 concat0 raid1 AIO0' '/dev/nbd0 /dev/nbd1 /dev/nbd10 /dev/nbd11 /dev/nbd12 /dev/nbd13 /dev/nbd14 /dev/nbd15 /dev/nbd2 /dev/nbd3 /dev/nbd4 /dev/nbd5 /dev/nbd6 /dev/nbd7 /dev/nbd8 /dev/nbd9' 00:09:16.631 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:16.631 11:19:59 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1p0' 'Malloc1p1' 'Malloc2p0' 'Malloc2p1' 'Malloc2p2' 'Malloc2p3' 'Malloc2p4' 'Malloc2p5' 'Malloc2p6' 'Malloc2p7' 'TestPT' 'raid0' 'concat0' 'raid1' 'AIO0') 00:09:16.631 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@10 -- # local bdev_list 00:09:16.631 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:09:16.631 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@11 -- # local nbd_list 00:09:16.631 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@12 -- # local i 00:09:16.631 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:09:16.631 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 16 )) 00:09:16.631 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:09:16.890 /dev/nbd0 00:09:16.890 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:09:16.890 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:09:16.890 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:09:16.890 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:16.890 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:16.890 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:16.890 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:09:16.890 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:16.890 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:16.890 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:16.890 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:16.890 1+0 records in 00:09:16.890 1+0 records out 00:09:16.890 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000265878 s, 15.4 MB/s 00:09:16.890 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:16.890 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:16.890 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:16.890 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:16.890 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:16.890 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:16.890 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 16 )) 00:09:16.890 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1p0 /dev/nbd1 00:09:17.149 /dev/nbd1 00:09:17.149 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:09:17.149 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:09:17.149 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:09:17.149 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:17.149 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:17.149 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:17.149 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:09:17.149 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:17.149 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:17.149 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:17.149 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:17.149 1+0 records in 00:09:17.149 1+0 records out 00:09:17.149 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000259182 s, 15.8 MB/s 00:09:17.149 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:17.149 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:17.149 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:17.149 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:17.149 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:17.149 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:17.149 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 16 )) 00:09:17.149 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1p1 /dev/nbd10 00:09:17.408 /dev/nbd10 00:09:17.408 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd10 00:09:17.408 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd10 00:09:17.408 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd10 00:09:17.408 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:17.408 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:17.408 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:17.408 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd10 /proc/partitions 00:09:17.408 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:17.408 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:17.408 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:17.408 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd10 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:17.408 1+0 records in 00:09:17.408 1+0 records out 00:09:17.408 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000318715 s, 12.9 MB/s 00:09:17.408 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:17.408 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:17.408 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:17.408 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:17.408 11:20:00 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:17.408 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:17.408 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 16 )) 00:09:17.408 11:20:00 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc2p0 /dev/nbd11 00:09:17.667 /dev/nbd11 00:09:17.667 11:20:01 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd11 00:09:17.667 11:20:01 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd11 00:09:17.667 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd11 00:09:17.667 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:17.667 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:17.667 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:17.667 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd11 /proc/partitions 00:09:17.667 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:17.667 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:17.667 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:17.667 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd11 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:17.667 1+0 records in 00:09:17.667 1+0 records out 00:09:17.667 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000332192 s, 12.3 MB/s 00:09:17.667 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:17.667 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:17.667 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:17.667 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:17.667 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:17.667 11:20:01 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:17.667 11:20:01 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 16 )) 00:09:17.667 11:20:01 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc2p1 /dev/nbd12 00:09:17.926 /dev/nbd12 00:09:17.926 11:20:01 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd12 00:09:17.926 11:20:01 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd12 00:09:17.926 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd12 00:09:17.926 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:17.926 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:17.926 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:17.926 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd12 /proc/partitions 00:09:17.926 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:17.926 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:17.926 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:17.926 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd12 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:17.926 1+0 records in 00:09:17.926 1+0 records out 00:09:17.926 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000374534 s, 10.9 MB/s 00:09:17.926 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:17.926 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:17.926 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:17.926 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:17.926 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:17.926 11:20:01 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:17.926 11:20:01 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 16 )) 00:09:17.926 11:20:01 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc2p2 /dev/nbd13 00:09:18.184 /dev/nbd13 00:09:18.184 11:20:01 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd13 00:09:18.184 11:20:01 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd13 00:09:18.184 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd13 00:09:18.184 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:18.184 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:18.184 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:18.184 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd13 /proc/partitions 00:09:18.184 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:18.184 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:18.184 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:18.184 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd13 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:18.184 1+0 records in 00:09:18.184 1+0 records out 00:09:18.184 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000375847 s, 10.9 MB/s 00:09:18.184 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:18.184 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:18.184 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:18.184 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:18.184 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:18.184 11:20:01 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:18.184 11:20:01 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 16 )) 00:09:18.185 11:20:01 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc2p3 /dev/nbd14 00:09:18.442 /dev/nbd14 00:09:18.442 11:20:01 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd14 00:09:18.442 11:20:01 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd14 00:09:18.443 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd14 00:09:18.443 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:18.443 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:18.443 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:18.443 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd14 /proc/partitions 00:09:18.443 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:18.443 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:18.443 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:18.443 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd14 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:18.443 1+0 records in 00:09:18.443 1+0 records out 00:09:18.443 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000359669 s, 11.4 MB/s 00:09:18.443 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:18.443 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:18.443 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:18.443 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:18.443 11:20:01 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:18.443 11:20:01 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:18.443 11:20:01 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 16 )) 00:09:18.443 11:20:01 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc2p4 /dev/nbd15 00:09:18.701 /dev/nbd15 00:09:18.701 11:20:02 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd15 00:09:18.701 11:20:02 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd15 00:09:18.701 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd15 00:09:18.701 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:18.701 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:18.701 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:18.701 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd15 /proc/partitions 00:09:18.701 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:18.701 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:18.701 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:18.701 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd15 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:18.701 1+0 records in 00:09:18.701 1+0 records out 00:09:18.701 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000369334 s, 11.1 MB/s 00:09:18.701 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:18.701 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:18.701 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:18.701 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:18.701 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:18.701 11:20:02 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:18.701 11:20:02 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 16 )) 00:09:18.701 11:20:02 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc2p5 /dev/nbd2 00:09:18.971 /dev/nbd2 00:09:18.972 11:20:02 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd2 00:09:18.972 11:20:02 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd2 00:09:18.972 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd2 00:09:18.972 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:18.972 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:18.972 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:18.972 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd2 /proc/partitions 00:09:18.972 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:18.972 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:18.972 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:18.972 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd2 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:18.972 1+0 records in 00:09:18.972 1+0 records out 00:09:18.972 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000460267 s, 8.9 MB/s 00:09:18.972 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:18.972 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:18.972 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:18.972 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:18.972 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:18.972 11:20:02 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:18.972 11:20:02 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 16 )) 00:09:18.972 11:20:02 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc2p6 /dev/nbd3 00:09:19.232 /dev/nbd3 00:09:19.232 11:20:02 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd3 00:09:19.232 11:20:02 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd3 00:09:19.232 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd3 00:09:19.232 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:19.232 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:19.232 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:19.232 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd3 /proc/partitions 00:09:19.232 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:19.232 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:19.232 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:19.232 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd3 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:19.232 1+0 records in 00:09:19.232 1+0 records out 00:09:19.232 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000540216 s, 7.6 MB/s 00:09:19.232 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:19.232 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:19.232 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:19.232 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:19.232 11:20:02 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:19.232 11:20:02 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:19.232 11:20:02 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 16 )) 00:09:19.232 11:20:02 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc2p7 /dev/nbd4 00:09:19.490 /dev/nbd4 00:09:19.490 11:20:03 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd4 00:09:19.490 11:20:03 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd4 00:09:19.490 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd4 00:09:19.490 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:19.490 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:19.490 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:19.490 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd4 /proc/partitions 00:09:19.490 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:19.490 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:19.490 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:19.490 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd4 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:19.490 1+0 records in 00:09:19.490 1+0 records out 00:09:19.490 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000647783 s, 6.3 MB/s 00:09:19.748 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:19.748 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:19.748 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:19.748 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:19.748 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:19.748 11:20:03 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:19.748 11:20:03 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 16 )) 00:09:19.748 11:20:03 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk TestPT /dev/nbd5 00:09:19.748 /dev/nbd5 00:09:20.006 11:20:03 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd5 00:09:20.007 11:20:03 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd5 00:09:20.007 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd5 00:09:20.007 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:20.007 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:20.007 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:20.007 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd5 /proc/partitions 00:09:20.007 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:20.007 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:20.007 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:20.007 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd5 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:20.007 1+0 records in 00:09:20.007 1+0 records out 00:09:20.007 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000624582 s, 6.6 MB/s 00:09:20.007 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:20.007 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:20.007 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:20.007 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:20.007 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:20.007 11:20:03 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:20.007 11:20:03 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 16 )) 00:09:20.007 11:20:03 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid0 /dev/nbd6 00:09:20.264 /dev/nbd6 00:09:20.264 11:20:03 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd6 00:09:20.264 11:20:03 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd6 00:09:20.264 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd6 00:09:20.264 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:20.264 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:20.264 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:20.264 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd6 /proc/partitions 00:09:20.264 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:20.264 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:20.264 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:20.264 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd6 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:20.264 1+0 records in 00:09:20.264 1+0 records out 00:09:20.264 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000553392 s, 7.4 MB/s 00:09:20.264 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:20.264 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:20.264 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:20.264 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:20.264 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:20.264 11:20:03 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:20.264 11:20:03 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 16 )) 00:09:20.264 11:20:03 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk concat0 /dev/nbd7 00:09:20.523 /dev/nbd7 00:09:20.523 11:20:03 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd7 00:09:20.523 11:20:03 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd7 00:09:20.523 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd7 00:09:20.523 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:20.523 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:20.523 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:20.523 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd7 /proc/partitions 00:09:20.523 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:20.523 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:20.523 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:20.523 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd7 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:20.523 1+0 records in 00:09:20.523 1+0 records out 00:09:20.523 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000732573 s, 5.6 MB/s 00:09:20.523 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:20.523 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:20.523 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:20.523 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:20.523 11:20:03 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:20.523 11:20:03 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:20.523 11:20:03 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 16 )) 00:09:20.523 11:20:03 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid1 /dev/nbd8 00:09:20.783 /dev/nbd8 00:09:20.783 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd8 00:09:20.783 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd8 00:09:20.783 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd8 00:09:20.783 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:20.783 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:20.783 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:20.783 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd8 /proc/partitions 00:09:20.783 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:20.783 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:20.783 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:20.783 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd8 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:20.783 1+0 records in 00:09:20.783 1+0 records out 00:09:20.783 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000511467 s, 8.0 MB/s 00:09:20.783 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:20.783 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:20.783 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:20.783 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:20.783 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:20.783 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:20.783 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 16 )) 00:09:20.783 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk AIO0 /dev/nbd9 00:09:21.042 /dev/nbd9 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd9 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd9 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd9 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd9 /proc/partitions 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd9 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:09:21.042 1+0 records in 00:09:21.042 1+0 records out 00:09:21.042 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000806639 s, 5.1 MB/s 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 16 )) 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:21.042 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:09:21.301 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd0", 00:09:21.301 "bdev_name": "Malloc0" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd1", 00:09:21.301 "bdev_name": "Malloc1p0" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd10", 00:09:21.301 "bdev_name": "Malloc1p1" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd11", 00:09:21.301 "bdev_name": "Malloc2p0" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd12", 00:09:21.301 "bdev_name": "Malloc2p1" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd13", 00:09:21.301 "bdev_name": "Malloc2p2" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd14", 00:09:21.301 "bdev_name": "Malloc2p3" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd15", 00:09:21.301 "bdev_name": "Malloc2p4" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd2", 00:09:21.301 "bdev_name": "Malloc2p5" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd3", 00:09:21.301 "bdev_name": "Malloc2p6" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd4", 00:09:21.301 "bdev_name": "Malloc2p7" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd5", 00:09:21.301 "bdev_name": "TestPT" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd6", 00:09:21.301 "bdev_name": "raid0" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd7", 00:09:21.301 "bdev_name": "concat0" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd8", 00:09:21.301 "bdev_name": "raid1" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd9", 00:09:21.301 "bdev_name": "AIO0" 00:09:21.301 } 00:09:21.301 ]' 00:09:21.301 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[ 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd0", 00:09:21.301 "bdev_name": "Malloc0" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd1", 00:09:21.301 "bdev_name": "Malloc1p0" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd10", 00:09:21.301 "bdev_name": "Malloc1p1" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd11", 00:09:21.301 "bdev_name": "Malloc2p0" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd12", 00:09:21.301 "bdev_name": "Malloc2p1" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd13", 00:09:21.301 "bdev_name": "Malloc2p2" 00:09:21.301 }, 00:09:21.301 { 00:09:21.301 "nbd_device": "/dev/nbd14", 00:09:21.301 "bdev_name": "Malloc2p3" 00:09:21.301 }, 00:09:21.301 { 00:09:21.302 "nbd_device": "/dev/nbd15", 00:09:21.302 "bdev_name": "Malloc2p4" 00:09:21.302 }, 00:09:21.302 { 00:09:21.302 "nbd_device": "/dev/nbd2", 00:09:21.302 "bdev_name": "Malloc2p5" 00:09:21.302 }, 00:09:21.302 { 00:09:21.302 "nbd_device": "/dev/nbd3", 00:09:21.302 "bdev_name": "Malloc2p6" 00:09:21.302 }, 00:09:21.302 { 00:09:21.302 "nbd_device": "/dev/nbd4", 00:09:21.302 "bdev_name": "Malloc2p7" 00:09:21.302 }, 00:09:21.302 { 00:09:21.302 "nbd_device": "/dev/nbd5", 00:09:21.302 "bdev_name": "TestPT" 00:09:21.302 }, 00:09:21.302 { 00:09:21.302 "nbd_device": "/dev/nbd6", 00:09:21.302 "bdev_name": "raid0" 00:09:21.302 }, 00:09:21.302 { 00:09:21.302 "nbd_device": "/dev/nbd7", 00:09:21.302 "bdev_name": "concat0" 00:09:21.302 }, 00:09:21.302 { 00:09:21.302 "nbd_device": "/dev/nbd8", 00:09:21.302 "bdev_name": "raid1" 00:09:21.302 }, 00:09:21.302 { 00:09:21.302 "nbd_device": "/dev/nbd9", 00:09:21.302 "bdev_name": "AIO0" 00:09:21.302 } 00:09:21.302 ]' 00:09:21.302 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:09:21.302 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:09:21.302 /dev/nbd1 00:09:21.302 /dev/nbd10 00:09:21.302 /dev/nbd11 00:09:21.302 /dev/nbd12 00:09:21.302 /dev/nbd13 00:09:21.302 /dev/nbd14 00:09:21.302 /dev/nbd15 00:09:21.302 /dev/nbd2 00:09:21.302 /dev/nbd3 00:09:21.302 /dev/nbd4 00:09:21.302 /dev/nbd5 00:09:21.302 /dev/nbd6 00:09:21.302 /dev/nbd7 00:09:21.302 /dev/nbd8 00:09:21.302 /dev/nbd9' 00:09:21.302 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:09:21.302 /dev/nbd1 00:09:21.302 /dev/nbd10 00:09:21.302 /dev/nbd11 00:09:21.302 /dev/nbd12 00:09:21.302 /dev/nbd13 00:09:21.302 /dev/nbd14 00:09:21.302 /dev/nbd15 00:09:21.302 /dev/nbd2 00:09:21.302 /dev/nbd3 00:09:21.302 /dev/nbd4 00:09:21.302 /dev/nbd5 00:09:21.302 /dev/nbd6 00:09:21.302 /dev/nbd7 00:09:21.302 /dev/nbd8 00:09:21.302 /dev/nbd9' 00:09:21.302 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:09:21.302 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=16 00:09:21.302 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 16 00:09:21.302 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@95 -- # count=16 00:09:21.302 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@96 -- # '[' 16 -ne 16 ']' 00:09:21.302 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1 /dev/nbd10 /dev/nbd11 /dev/nbd12 /dev/nbd13 /dev/nbd14 /dev/nbd15 /dev/nbd2 /dev/nbd3 /dev/nbd4 /dev/nbd5 /dev/nbd6 /dev/nbd7 /dev/nbd8 /dev/nbd9' write 00:09:21.302 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:09:21.302 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:09:21.302 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=write 00:09:21.302 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest 00:09:21.302 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:09:21.302 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest bs=4096 count=256 00:09:21.302 256+0 records in 00:09:21.302 256+0 records out 00:09:21.302 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0109386 s, 95.9 MB/s 00:09:21.302 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:21.302 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:09:21.560 256+0 records in 00:09:21.560 256+0 records out 00:09:21.560 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.182138 s, 5.8 MB/s 00:09:21.560 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:21.560 11:20:04 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:09:21.821 256+0 records in 00:09:21.821 256+0 records out 00:09:21.821 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.183974 s, 5.7 MB/s 00:09:21.821 11:20:05 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:21.821 11:20:05 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd10 bs=4096 count=256 oflag=direct 00:09:21.821 256+0 records in 00:09:21.821 256+0 records out 00:09:21.821 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.183359 s, 5.7 MB/s 00:09:21.821 11:20:05 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:21.821 11:20:05 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd11 bs=4096 count=256 oflag=direct 00:09:22.118 256+0 records in 00:09:22.118 256+0 records out 00:09:22.118 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.183912 s, 5.7 MB/s 00:09:22.118 11:20:05 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:22.118 11:20:05 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd12 bs=4096 count=256 oflag=direct 00:09:22.405 256+0 records in 00:09:22.405 256+0 records out 00:09:22.405 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.183835 s, 5.7 MB/s 00:09:22.405 11:20:05 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:22.405 11:20:05 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd13 bs=4096 count=256 oflag=direct 00:09:22.405 256+0 records in 00:09:22.405 256+0 records out 00:09:22.405 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.18471 s, 5.7 MB/s 00:09:22.405 11:20:05 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:22.405 11:20:05 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd14 bs=4096 count=256 oflag=direct 00:09:22.661 256+0 records in 00:09:22.661 256+0 records out 00:09:22.661 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.184045 s, 5.7 MB/s 00:09:22.661 11:20:06 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:22.661 11:20:06 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd15 bs=4096 count=256 oflag=direct 00:09:22.918 256+0 records in 00:09:22.918 256+0 records out 00:09:22.918 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.183929 s, 5.7 MB/s 00:09:22.918 11:20:06 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:22.918 11:20:06 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd2 bs=4096 count=256 oflag=direct 00:09:22.918 256+0 records in 00:09:22.918 256+0 records out 00:09:22.918 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.182774 s, 5.7 MB/s 00:09:22.918 11:20:06 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:22.918 11:20:06 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd3 bs=4096 count=256 oflag=direct 00:09:23.176 256+0 records in 00:09:23.176 256+0 records out 00:09:23.176 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.18386 s, 5.7 MB/s 00:09:23.176 11:20:06 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:23.176 11:20:06 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd4 bs=4096 count=256 oflag=direct 00:09:23.435 256+0 records in 00:09:23.435 256+0 records out 00:09:23.435 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.184051 s, 5.7 MB/s 00:09:23.435 11:20:06 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:23.435 11:20:06 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd5 bs=4096 count=256 oflag=direct 00:09:23.692 256+0 records in 00:09:23.692 256+0 records out 00:09:23.692 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.183517 s, 5.7 MB/s 00:09:23.692 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:23.692 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd6 bs=4096 count=256 oflag=direct 00:09:23.692 256+0 records in 00:09:23.692 256+0 records out 00:09:23.692 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.184835 s, 5.7 MB/s 00:09:23.692 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:23.692 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd7 bs=4096 count=256 oflag=direct 00:09:23.949 256+0 records in 00:09:23.949 256+0 records out 00:09:23.949 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.18556 s, 5.7 MB/s 00:09:23.949 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:23.949 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd8 bs=4096 count=256 oflag=direct 00:09:24.205 256+0 records in 00:09:24.205 256+0 records out 00:09:24.205 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.187857 s, 5.6 MB/s 00:09:24.205 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:24.205 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd9 bs=4096 count=256 oflag=direct 00:09:24.205 256+0 records in 00:09:24.205 256+0 records out 00:09:24.205 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.182256 s, 5.8 MB/s 00:09:24.205 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1 /dev/nbd10 /dev/nbd11 /dev/nbd12 /dev/nbd13 /dev/nbd14 /dev/nbd15 /dev/nbd2 /dev/nbd3 /dev/nbd4 /dev/nbd5 /dev/nbd6 /dev/nbd7 /dev/nbd8 /dev/nbd9' verify 00:09:24.205 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:09:24.205 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=verify 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd0 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd1 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd10 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd11 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd12 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd13 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd14 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd15 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd2 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd3 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd4 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd5 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd6 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd7 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd8 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd9 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1 /dev/nbd10 /dev/nbd11 /dev/nbd12 /dev/nbd13 /dev/nbd14 /dev/nbd15 /dev/nbd2 /dev/nbd3 /dev/nbd4 /dev/nbd5 /dev/nbd6 /dev/nbd7 /dev/nbd8 /dev/nbd9' 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:24.463 11:20:07 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:09:24.719 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:09:24.719 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:09:24.719 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:09:24.719 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:24.719 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:24.719 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:09:24.719 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:24.719 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:24.719 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:24.719 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:09:24.976 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:09:24.976 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:09:24.976 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:09:24.976 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:24.976 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:24.976 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:09:24.976 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:24.976 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:24.976 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:24.976 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd10 00:09:25.234 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd10 00:09:25.234 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd10 00:09:25.234 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd10 00:09:25.234 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:25.234 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:25.234 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd10 /proc/partitions 00:09:25.234 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:25.234 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:25.234 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:25.234 11:20:08 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd11 00:09:25.491 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd11 00:09:25.491 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd11 00:09:25.491 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd11 00:09:25.491 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:25.491 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:25.491 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd11 /proc/partitions 00:09:25.491 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:25.491 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:25.491 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:25.491 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd12 00:09:25.747 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd12 00:09:25.747 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd12 00:09:25.747 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd12 00:09:25.747 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:25.747 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:25.747 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd12 /proc/partitions 00:09:26.005 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:26.005 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:26.005 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:26.005 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd13 00:09:26.263 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd13 00:09:26.263 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd13 00:09:26.263 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd13 00:09:26.263 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:26.263 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:26.263 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd13 /proc/partitions 00:09:26.263 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@39 -- # sleep 0.1 00:09:26.263 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i++ )) 00:09:26.263 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:26.263 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd13 /proc/partitions 00:09:26.263 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:26.263 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:26.263 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:26.263 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd14 00:09:26.521 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd14 00:09:26.521 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd14 00:09:26.521 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd14 00:09:26.521 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:26.521 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:26.521 11:20:09 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd14 /proc/partitions 00:09:26.521 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:26.521 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:26.521 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:26.521 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd15 00:09:26.779 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd15 00:09:26.779 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd15 00:09:26.779 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd15 00:09:26.779 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:26.779 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:26.779 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd15 /proc/partitions 00:09:26.779 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:26.779 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:26.779 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:26.779 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd2 00:09:27.036 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd2 00:09:27.036 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd2 00:09:27.036 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd2 00:09:27.036 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:27.036 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:27.036 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd2 /proc/partitions 00:09:27.036 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:27.036 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:27.036 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:27.036 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd3 00:09:27.294 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd3 00:09:27.294 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd3 00:09:27.294 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd3 00:09:27.294 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:27.294 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:27.294 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd3 /proc/partitions 00:09:27.294 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:27.294 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:27.294 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:27.294 11:20:10 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd4 00:09:27.552 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd4 00:09:27.552 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd4 00:09:27.552 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd4 00:09:27.552 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:27.552 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:27.552 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd4 /proc/partitions 00:09:27.552 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:27.552 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:27.552 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:27.552 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd5 00:09:27.810 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd5 00:09:27.810 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd5 00:09:27.810 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd5 00:09:27.810 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:27.810 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:27.810 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd5 /proc/partitions 00:09:27.810 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:27.810 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:27.810 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:27.810 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd6 00:09:28.376 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd6 00:09:28.376 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd6 00:09:28.376 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd6 00:09:28.376 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:28.376 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:28.376 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd6 /proc/partitions 00:09:28.376 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:28.376 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:28.376 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:28.376 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd7 00:09:28.376 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd7 00:09:28.376 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd7 00:09:28.376 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd7 00:09:28.376 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:28.376 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:28.376 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd7 /proc/partitions 00:09:28.376 11:20:11 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@39 -- # sleep 0.1 00:09:28.634 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i++ )) 00:09:28.634 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:28.634 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd7 /proc/partitions 00:09:28.634 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:28.634 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:28.634 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:28.634 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd8 00:09:28.909 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd8 00:09:28.909 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd8 00:09:28.909 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd8 00:09:28.909 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:28.909 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:28.909 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd8 /proc/partitions 00:09:28.909 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:28.909 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:28.909 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:28.909 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd9 00:09:29.167 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd9 00:09:29.167 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd9 00:09:29.167 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd9 00:09:29.167 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:29.167 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:29.167 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd9 /proc/partitions 00:09:29.167 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:29.167 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:29.167 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:09:29.167 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:29.167 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:09:29.425 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:09:29.425 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:09:29.425 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:09:29.425 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:09:29.425 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:09:29.425 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:09:29.425 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:09:29.425 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:09:29.425 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:09:29.425 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@104 -- # count=0 00:09:29.425 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:09:29.425 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@109 -- # return 0 00:09:29.425 11:20:12 blockdev_general.bdev_nbd -- bdev/blockdev.sh@324 -- # nbd_with_lvol_verify /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1 /dev/nbd10 /dev/nbd11 /dev/nbd12 /dev/nbd13 /dev/nbd14 /dev/nbd15 /dev/nbd2 /dev/nbd3 /dev/nbd4 /dev/nbd5 /dev/nbd6 /dev/nbd7 /dev/nbd8 /dev/nbd9' 00:09:29.425 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@131 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:29.425 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@132 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:09:29.425 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@132 -- # local nbd_list 00:09:29.425 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@133 -- # local mkfs_ret 00:09:29.425 11:20:12 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@135 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create -b malloc_lvol_verify 16 512 00:09:29.682 malloc_lvol_verify 00:09:29.682 11:20:13 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@136 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create_lvstore malloc_lvol_verify lvs 00:09:29.939 71afcd32-d756-4477-86d1-7a66d662ede7 00:09:29.939 11:20:13 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@137 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create lvol 4 -l lvs 00:09:30.196 dff04ee8-bf3a-453e-9f95-b5278688926b 00:09:30.196 11:20:13 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@138 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk lvs/lvol /dev/nbd0 00:09:30.454 /dev/nbd0 00:09:30.454 11:20:13 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@140 -- # mkfs.ext4 /dev/nbd0 00:09:30.454 mke2fs 1.46.5 (30-Dec-2021) 00:09:30.454 Discarding device blocks: 0/4096 done 00:09:30.454 Creating filesystem with 4096 1k blocks and 1024 inodes 00:09:30.454 00:09:30.454 Allocating group tables: 0/1 done 00:09:30.454 Writing inode tables: 0/1 done 00:09:30.454 Creating journal (1024 blocks): done 00:09:30.454 Writing superblocks and filesystem accounting information: 0/1 done 00:09:30.454 00:09:30.454 11:20:13 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@141 -- # mkfs_ret=0 00:09:30.454 11:20:13 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@142 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:09:30.454 11:20:13 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:30.454 11:20:13 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:09:30.454 11:20:13 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:09:30.454 11:20:13 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:09:30.454 11:20:13 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:30.454 11:20:13 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@143 -- # '[' 0 -ne 0 ']' 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- bdev/nbd_common.sh@147 -- # return 0 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- bdev/blockdev.sh@326 -- # killprocess 843889 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- common/autotest_common.sh@948 -- # '[' -z 843889 ']' 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- common/autotest_common.sh@952 -- # kill -0 843889 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- common/autotest_common.sh@953 -- # uname 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 843889 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- common/autotest_common.sh@966 -- # echo 'killing process with pid 843889' 00:09:30.711 killing process with pid 843889 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- common/autotest_common.sh@967 -- # kill 843889 00:09:30.711 11:20:14 blockdev_general.bdev_nbd -- common/autotest_common.sh@972 -- # wait 843889 00:09:31.328 11:20:14 blockdev_general.bdev_nbd -- bdev/blockdev.sh@327 -- # trap - SIGINT SIGTERM EXIT 00:09:31.328 00:09:31.328 real 0m25.341s 00:09:31.328 user 0m30.838s 00:09:31.328 sys 0m14.533s 00:09:31.328 11:20:14 blockdev_general.bdev_nbd -- common/autotest_common.sh@1124 -- # xtrace_disable 00:09:31.328 11:20:14 blockdev_general.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:09:31.328 ************************************ 00:09:31.328 END TEST bdev_nbd 00:09:31.328 ************************************ 00:09:31.328 11:20:14 blockdev_general -- common/autotest_common.sh@1142 -- # return 0 00:09:31.328 11:20:14 blockdev_general -- bdev/blockdev.sh@763 -- # [[ y == y ]] 00:09:31.328 11:20:14 blockdev_general -- bdev/blockdev.sh@764 -- # '[' bdev = nvme ']' 00:09:31.328 11:20:14 blockdev_general -- bdev/blockdev.sh@764 -- # '[' bdev = gpt ']' 00:09:31.328 11:20:14 blockdev_general -- bdev/blockdev.sh@768 -- # run_test bdev_fio fio_test_suite '' 00:09:31.328 11:20:14 blockdev_general -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:09:31.328 11:20:14 blockdev_general -- common/autotest_common.sh@1105 -- # xtrace_disable 00:09:31.328 11:20:14 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:09:31.585 ************************************ 00:09:31.585 START TEST bdev_fio 00:09:31.585 ************************************ 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- common/autotest_common.sh@1123 -- # fio_test_suite '' 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- bdev/blockdev.sh@331 -- # local env_context 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- bdev/blockdev.sh@335 -- # pushd /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev 00:09:31.585 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev /var/jenkins/workspace/crypto-phy-autotest/spdk 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- bdev/blockdev.sh@336 -- # trap 'rm -f ./*.state; popd; exit 1' SIGINT SIGTERM EXIT 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- bdev/blockdev.sh@339 -- # echo '' 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- bdev/blockdev.sh@339 -- # sed s/--env-context=// 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- bdev/blockdev.sh@339 -- # env_context= 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- bdev/blockdev.sh@340 -- # fio_config_gen /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio verify AIO '' 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- common/autotest_common.sh@1280 -- # local config_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- common/autotest_common.sh@1281 -- # local workload=verify 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- common/autotest_common.sh@1282 -- # local bdev_type=AIO 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- common/autotest_common.sh@1283 -- # local env_context= 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- common/autotest_common.sh@1284 -- # local fio_dir=/usr/src/fio 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- common/autotest_common.sh@1286 -- # '[' -e /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio ']' 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- common/autotest_common.sh@1291 -- # '[' -z verify ']' 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -n '' ']' 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- common/autotest_common.sh@1299 -- # touch /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- common/autotest_common.sh@1301 -- # cat 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- common/autotest_common.sh@1313 -- # '[' verify == verify ']' 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- common/autotest_common.sh@1314 -- # cat 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- common/autotest_common.sh@1323 -- # '[' AIO == AIO ']' 00:09:31.585 11:20:14 blockdev_general.bdev_fio -- common/autotest_common.sh@1324 -- # /usr/src/fio/fio --version 00:09:31.585 11:20:15 blockdev_general.bdev_fio -- common/autotest_common.sh@1324 -- # [[ fio-3.35 == *\f\i\o\-\3* ]] 00:09:31.585 11:20:15 blockdev_general.bdev_fio -- common/autotest_common.sh@1325 -- # echo serialize_overlap=1 00:09:31.585 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:09:31.585 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_Malloc0]' 00:09:31.585 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=Malloc0 00:09:31.585 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:09:31.585 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_Malloc1p0]' 00:09:31.585 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=Malloc1p0 00:09:31.585 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:09:31.585 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_Malloc1p1]' 00:09:31.585 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=Malloc1p1 00:09:31.585 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:09:31.585 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_Malloc2p0]' 00:09:31.585 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=Malloc2p0 00:09:31.585 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_Malloc2p1]' 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=Malloc2p1 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_Malloc2p2]' 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=Malloc2p2 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_Malloc2p3]' 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=Malloc2p3 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_Malloc2p4]' 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=Malloc2p4 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_Malloc2p5]' 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=Malloc2p5 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_Malloc2p6]' 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=Malloc2p6 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_Malloc2p7]' 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=Malloc2p7 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_TestPT]' 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=TestPT 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_raid0]' 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=raid0 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_concat0]' 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=concat0 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_raid1]' 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=raid1 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_AIO0]' 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=AIO0 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@347 -- # local 'fio_params=--ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json' 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- bdev/blockdev.sh@349 -- # run_test bdev_fio_rw_verify fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- common/autotest_common.sh@1099 -- # '[' 11 -le 1 ']' 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- common/autotest_common.sh@1105 -- # xtrace_disable 00:09:31.586 11:20:15 blockdev_general.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:09:31.586 ************************************ 00:09:31.586 START TEST bdev_fio_rw_verify 00:09:31.586 ************************************ 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1123 -- # fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1339 -- # local sanitizers 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1341 -- # shift 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # local asan_lib= 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # grep libasan 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # asan_lib= 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # asan_lib= 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev' 00:09:31.586 11:20:15 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:09:31.844 job_Malloc0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:31.844 job_Malloc1p0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:31.844 job_Malloc1p1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:31.844 job_Malloc2p0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:31.844 job_Malloc2p1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:31.844 job_Malloc2p2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:31.844 job_Malloc2p3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:31.844 job_Malloc2p4: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:31.844 job_Malloc2p5: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:31.844 job_Malloc2p6: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:31.844 job_Malloc2p7: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:31.844 job_TestPT: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:31.844 job_raid0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:31.844 job_concat0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:31.844 job_raid1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:31.844 job_AIO0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:31.844 fio-3.35 00:09:31.844 Starting 16 threads 00:09:44.047 00:09:44.047 job_Malloc0: (groupid=0, jobs=16): err= 0: pid=848092: Mon Jul 15 11:20:26 2024 00:09:44.047 read: IOPS=96.4k, BW=377MiB/s (395MB/s)(3767MiB/10001msec) 00:09:44.047 slat (usec): min=3, max=451, avg=32.75, stdev=15.03 00:09:44.047 clat (usec): min=10, max=1229, avg=268.52, stdev=140.13 00:09:44.047 lat (usec): min=24, max=1339, avg=301.27, stdev=150.47 00:09:44.047 clat percentiles (usec): 00:09:44.047 | 50.000th=[ 251], 99.000th=[ 725], 99.900th=[ 898], 99.990th=[ 996], 00:09:44.047 | 99.999th=[ 1074] 00:09:44.047 write: IOPS=150k, BW=586MiB/s (614MB/s)(5779MiB/9864msec); 0 zone resets 00:09:44.047 slat (usec): min=6, max=2391, avg=46.10, stdev=16.46 00:09:44.047 clat (usec): min=12, max=2986, avg=323.31, stdev=165.77 00:09:44.047 lat (usec): min=34, max=3041, avg=369.41, stdev=176.43 00:09:44.047 clat percentiles (usec): 00:09:44.047 | 50.000th=[ 297], 99.000th=[ 840], 99.900th=[ 1074], 99.990th=[ 1172], 00:09:44.047 | 99.999th=[ 1582] 00:09:44.047 bw ( KiB/s): min=431606, max=777161, per=99.50%, avg=596945.00, stdev=6994.71, samples=304 00:09:44.047 iops : min=107901, max=194289, avg=149235.95, stdev=1748.67, samples=304 00:09:44.047 lat (usec) : 20=0.01%, 50=0.57%, 100=4.60%, 250=37.57%, 500=47.19% 00:09:44.047 lat (usec) : 750=8.34%, 1000=1.57% 00:09:44.047 lat (msec) : 2=0.15%, 4=0.01% 00:09:44.047 cpu : usr=99.15%, sys=0.45%, ctx=530, majf=0, minf=1112 00:09:44.047 IO depths : 1=12.5%, 2=25.0%, 4=50.0%, 8=12.5%, 16=0.0%, 32=0.0%, >=64=0.0% 00:09:44.047 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:44.047 complete : 0=0.0%, 4=89.0%, 8=11.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:44.047 issued rwts: total=964351,1479445,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:44.047 latency : target=0, window=0, percentile=100.00%, depth=8 00:09:44.047 00:09:44.047 Run status group 0 (all jobs): 00:09:44.047 READ: bw=377MiB/s (395MB/s), 377MiB/s-377MiB/s (395MB/s-395MB/s), io=3767MiB (3950MB), run=10001-10001msec 00:09:44.047 WRITE: bw=586MiB/s (614MB/s), 586MiB/s-586MiB/s (614MB/s-614MB/s), io=5779MiB (6060MB), run=9864-9864msec 00:09:44.047 00:09:44.047 real 0m11.552s 00:09:44.047 user 2m44.749s 00:09:44.047 sys 0m1.783s 00:09:44.047 11:20:26 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1124 -- # xtrace_disable 00:09:44.047 11:20:26 blockdev_general.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@10 -- # set +x 00:09:44.047 ************************************ 00:09:44.047 END TEST bdev_fio_rw_verify 00:09:44.047 ************************************ 00:09:44.047 11:20:26 blockdev_general.bdev_fio -- common/autotest_common.sh@1142 -- # return 0 00:09:44.047 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@350 -- # rm -f 00:09:44.047 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@351 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:09:44.047 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@354 -- # fio_config_gen /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio trim '' '' 00:09:44.047 11:20:26 blockdev_general.bdev_fio -- common/autotest_common.sh@1280 -- # local config_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:09:44.047 11:20:26 blockdev_general.bdev_fio -- common/autotest_common.sh@1281 -- # local workload=trim 00:09:44.047 11:20:26 blockdev_general.bdev_fio -- common/autotest_common.sh@1282 -- # local bdev_type= 00:09:44.047 11:20:26 blockdev_general.bdev_fio -- common/autotest_common.sh@1283 -- # local env_context= 00:09:44.047 11:20:26 blockdev_general.bdev_fio -- common/autotest_common.sh@1284 -- # local fio_dir=/usr/src/fio 00:09:44.047 11:20:26 blockdev_general.bdev_fio -- common/autotest_common.sh@1286 -- # '[' -e /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio ']' 00:09:44.047 11:20:26 blockdev_general.bdev_fio -- common/autotest_common.sh@1291 -- # '[' -z trim ']' 00:09:44.047 11:20:26 blockdev_general.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -n '' ']' 00:09:44.047 11:20:26 blockdev_general.bdev_fio -- common/autotest_common.sh@1299 -- # touch /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:09:44.047 11:20:26 blockdev_general.bdev_fio -- common/autotest_common.sh@1301 -- # cat 00:09:44.047 11:20:26 blockdev_general.bdev_fio -- common/autotest_common.sh@1313 -- # '[' trim == verify ']' 00:09:44.047 11:20:26 blockdev_general.bdev_fio -- common/autotest_common.sh@1328 -- # '[' trim == trim ']' 00:09:44.047 11:20:26 blockdev_general.bdev_fio -- common/autotest_common.sh@1329 -- # echo rw=trimwrite 00:09:44.047 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@355 -- # jq -r 'select(.supported_io_types.unmap == true) | .name' 00:09:44.049 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@355 -- # printf '%s\n' '{' ' "name": "Malloc0",' ' "aliases": [' ' "5ce527dc-eb58-413a-827d-f3c0a4fd090b"' ' ],' ' "product_name": "Malloc disk",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "5ce527dc-eb58-413a-827d-f3c0a4fd090b",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 20000,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {}' '}' '{' ' "name": "Malloc1p0",' ' "aliases": [' ' "5fadbd15-2a90-55bd-ae35-f5123bafe8da"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 32768,' ' "uuid": "5fadbd15-2a90-55bd-ae35-f5123bafe8da",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc1",' ' "offset_blocks": 0' ' }' ' }' '}' '{' ' "name": "Malloc1p1",' ' "aliases": [' ' "2d0d2f9e-1b12-5e40-8b76-3029d8b089cc"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 32768,' ' "uuid": "2d0d2f9e-1b12-5e40-8b76-3029d8b089cc",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc1",' ' "offset_blocks": 32768' ' }' ' }' '}' '{' ' "name": "Malloc2p0",' ' "aliases": [' ' "d022053a-fcc7-56c1-af68-8f04876a4109"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "d022053a-fcc7-56c1-af68-8f04876a4109",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 0' ' }' ' }' '}' '{' ' "name": "Malloc2p1",' ' "aliases": [' ' "b6ff7cf9-30d0-53bb-856f-e4fd483be15c"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "b6ff7cf9-30d0-53bb-856f-e4fd483be15c",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 8192' ' }' ' }' '}' '{' ' "name": "Malloc2p2",' ' "aliases": [' ' "6df4933e-5f3c-5662-8ad9-b1cb90107c3c"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "6df4933e-5f3c-5662-8ad9-b1cb90107c3c",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 16384' ' }' ' }' '}' '{' ' "name": "Malloc2p3",' ' "aliases": [' ' "f27852e2-a19f-5771-a71c-11be9d704d5d"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "f27852e2-a19f-5771-a71c-11be9d704d5d",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 24576' ' }' ' }' '}' '{' ' "name": "Malloc2p4",' ' "aliases": [' ' "86f93d9d-194b-5784-8a78-9f5e07868a7e"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "86f93d9d-194b-5784-8a78-9f5e07868a7e",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 32768' ' }' ' }' '}' '{' ' "name": "Malloc2p5",' ' "aliases": [' ' "da83a269-dd19-5e99-a12c-13edcb7db0cd"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "da83a269-dd19-5e99-a12c-13edcb7db0cd",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 40960' ' }' ' }' '}' '{' ' "name": "Malloc2p6",' ' "aliases": [' ' "9d07fa39-831f-5784-b785-279997ba3af0"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "9d07fa39-831f-5784-b785-279997ba3af0",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 49152' ' }' ' }' '}' '{' ' "name": "Malloc2p7",' ' "aliases": [' ' "6c40d456-1f09-5c69-9c20-4523d2ef39e8"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "6c40d456-1f09-5c69-9c20-4523d2ef39e8",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 57344' ' }' ' }' '}' '{' ' "name": "TestPT",' ' "aliases": [' ' "34c56ada-8252-54db-aeb7-e85474d6df8d"' ' ],' ' "product_name": "passthru",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "34c56ada-8252-54db-aeb7-e85474d6df8d",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "passthru": {' ' "name": "TestPT",' ' "base_bdev_name": "Malloc3"' ' }' ' }' '}' '{' ' "name": "raid0",' ' "aliases": [' ' "23b1ca62-1017-4838-909e-76ea1f6cad62"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "23b1ca62-1017-4838-909e-76ea1f6cad62",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "raid": {' ' "uuid": "23b1ca62-1017-4838-909e-76ea1f6cad62",' ' "strip_size_kb": 64,' ' "state": "online",' ' "raid_level": "raid0",' ' "superblock": false,' ' "num_base_bdevs": 2,' ' "num_base_bdevs_discovered": 2,' ' "num_base_bdevs_operational": 2,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc4",' ' "uuid": "6de1596c-08db-4b80-86ea-96d56c0345c7",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc5",' ' "uuid": "05f2ee9c-ccde-4341-ac99-c1ef2c8f5ce7",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' '{' ' "name": "concat0",' ' "aliases": [' ' "e2eb45e4-69ab-4edf-95dd-c81cd873e951"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "e2eb45e4-69ab-4edf-95dd-c81cd873e951",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "raid": {' ' "uuid": "e2eb45e4-69ab-4edf-95dd-c81cd873e951",' ' "strip_size_kb": 64,' ' "state": "online",' ' "raid_level": "concat",' ' "superblock": false,' ' "num_base_bdevs": 2,' ' "num_base_bdevs_discovered": 2,' ' "num_base_bdevs_operational": 2,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc6",' ' "uuid": "e5067475-6ef9-43ae-bcb9-d5ba9d228e5e",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc7",' ' "uuid": "64e1c577-11ea-4bed-84d9-9fee33627e6c",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' '{' ' "name": "raid1",' ' "aliases": [' ' "80122521-8fa1-4582-a8d0-a569befdbe0f"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "80122521-8fa1-4582-a8d0-a569befdbe0f",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "raid": {' ' "uuid": "80122521-8fa1-4582-a8d0-a569befdbe0f",' ' "strip_size_kb": 0,' ' "state": "online",' ' "raid_level": "raid1",' ' "superblock": false,' ' "num_base_bdevs": 2,' ' "num_base_bdevs_discovered": 2,' ' "num_base_bdevs_operational": 2,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc8",' ' "uuid": "91863e0b-b6e3-4ae1-85b3-dbd0ab80231f",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc9",' ' "uuid": "e8e513fa-3b18-46dd-ad2f-1e2ffbb5ae8e",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' '{' ' "name": "AIO0",' ' "aliases": [' ' "0afa1591-1717-40c4-a4dd-e5b16747b803"' ' ],' ' "product_name": "AIO disk",' ' "block_size": 2048,' ' "num_blocks": 5000,' ' "uuid": "0afa1591-1717-40c4-a4dd-e5b16747b803",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "aio": {' ' "filename": "/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/aiofile",' ' "block_size_override": true,' ' "readonly": false,' ' "fallocate": false' ' }' ' }' '}' 00:09:44.049 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@355 -- # [[ -n Malloc0 00:09:44.049 Malloc1p0 00:09:44.049 Malloc1p1 00:09:44.049 Malloc2p0 00:09:44.049 Malloc2p1 00:09:44.049 Malloc2p2 00:09:44.049 Malloc2p3 00:09:44.049 Malloc2p4 00:09:44.049 Malloc2p5 00:09:44.049 Malloc2p6 00:09:44.049 Malloc2p7 00:09:44.049 TestPT 00:09:44.049 raid0 00:09:44.049 concat0 ]] 00:09:44.049 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@356 -- # jq -r 'select(.supported_io_types.unmap == true) | .name' 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@356 -- # printf '%s\n' '{' ' "name": "Malloc0",' ' "aliases": [' ' "5ce527dc-eb58-413a-827d-f3c0a4fd090b"' ' ],' ' "product_name": "Malloc disk",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "5ce527dc-eb58-413a-827d-f3c0a4fd090b",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 20000,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {}' '}' '{' ' "name": "Malloc1p0",' ' "aliases": [' ' "5fadbd15-2a90-55bd-ae35-f5123bafe8da"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 32768,' ' "uuid": "5fadbd15-2a90-55bd-ae35-f5123bafe8da",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc1",' ' "offset_blocks": 0' ' }' ' }' '}' '{' ' "name": "Malloc1p1",' ' "aliases": [' ' "2d0d2f9e-1b12-5e40-8b76-3029d8b089cc"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 32768,' ' "uuid": "2d0d2f9e-1b12-5e40-8b76-3029d8b089cc",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc1",' ' "offset_blocks": 32768' ' }' ' }' '}' '{' ' "name": "Malloc2p0",' ' "aliases": [' ' "d022053a-fcc7-56c1-af68-8f04876a4109"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "d022053a-fcc7-56c1-af68-8f04876a4109",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 0' ' }' ' }' '}' '{' ' "name": "Malloc2p1",' ' "aliases": [' ' "b6ff7cf9-30d0-53bb-856f-e4fd483be15c"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "b6ff7cf9-30d0-53bb-856f-e4fd483be15c",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 8192' ' }' ' }' '}' '{' ' "name": "Malloc2p2",' ' "aliases": [' ' "6df4933e-5f3c-5662-8ad9-b1cb90107c3c"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "6df4933e-5f3c-5662-8ad9-b1cb90107c3c",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 16384' ' }' ' }' '}' '{' ' "name": "Malloc2p3",' ' "aliases": [' ' "f27852e2-a19f-5771-a71c-11be9d704d5d"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "f27852e2-a19f-5771-a71c-11be9d704d5d",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 24576' ' }' ' }' '}' '{' ' "name": "Malloc2p4",' ' "aliases": [' ' "86f93d9d-194b-5784-8a78-9f5e07868a7e"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "86f93d9d-194b-5784-8a78-9f5e07868a7e",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 32768' ' }' ' }' '}' '{' ' "name": "Malloc2p5",' ' "aliases": [' ' "da83a269-dd19-5e99-a12c-13edcb7db0cd"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "da83a269-dd19-5e99-a12c-13edcb7db0cd",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 40960' ' }' ' }' '}' '{' ' "name": "Malloc2p6",' ' "aliases": [' ' "9d07fa39-831f-5784-b785-279997ba3af0"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "9d07fa39-831f-5784-b785-279997ba3af0",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 49152' ' }' ' }' '}' '{' ' "name": "Malloc2p7",' ' "aliases": [' ' "6c40d456-1f09-5c69-9c20-4523d2ef39e8"' ' ],' ' "product_name": "Split Disk",' ' "block_size": 512,' ' "num_blocks": 8192,' ' "uuid": "6c40d456-1f09-5c69-9c20-4523d2ef39e8",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "split": {' ' "base_bdev": "Malloc2",' ' "offset_blocks": 57344' ' }' ' }' '}' '{' ' "name": "TestPT",' ' "aliases": [' ' "34c56ada-8252-54db-aeb7-e85474d6df8d"' ' ],' ' "product_name": "passthru",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "34c56ada-8252-54db-aeb7-e85474d6df8d",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": true,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": true,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": true,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "passthru": {' ' "name": "TestPT",' ' "base_bdev_name": "Malloc3"' ' }' ' }' '}' '{' ' "name": "raid0",' ' "aliases": [' ' "23b1ca62-1017-4838-909e-76ea1f6cad62"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "23b1ca62-1017-4838-909e-76ea1f6cad62",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "raid": {' ' "uuid": "23b1ca62-1017-4838-909e-76ea1f6cad62",' ' "strip_size_kb": 64,' ' "state": "online",' ' "raid_level": "raid0",' ' "superblock": false,' ' "num_base_bdevs": 2,' ' "num_base_bdevs_discovered": 2,' ' "num_base_bdevs_operational": 2,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc4",' ' "uuid": "6de1596c-08db-4b80-86ea-96d56c0345c7",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc5",' ' "uuid": "05f2ee9c-ccde-4341-ac99-c1ef2c8f5ce7",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' '{' ' "name": "concat0",' ' "aliases": [' ' "e2eb45e4-69ab-4edf-95dd-c81cd873e951"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "e2eb45e4-69ab-4edf-95dd-c81cd873e951",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "raid": {' ' "uuid": "e2eb45e4-69ab-4edf-95dd-c81cd873e951",' ' "strip_size_kb": 64,' ' "state": "online",' ' "raid_level": "concat",' ' "superblock": false,' ' "num_base_bdevs": 2,' ' "num_base_bdevs_discovered": 2,' ' "num_base_bdevs_operational": 2,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc6",' ' "uuid": "e5067475-6ef9-43ae-bcb9-d5ba9d228e5e",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc7",' ' "uuid": "64e1c577-11ea-4bed-84d9-9fee33627e6c",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' '{' ' "name": "raid1",' ' "aliases": [' ' "80122521-8fa1-4582-a8d0-a569befdbe0f"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "80122521-8fa1-4582-a8d0-a569befdbe0f",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "raid": {' ' "uuid": "80122521-8fa1-4582-a8d0-a569befdbe0f",' ' "strip_size_kb": 0,' ' "state": "online",' ' "raid_level": "raid1",' ' "superblock": false,' ' "num_base_bdevs": 2,' ' "num_base_bdevs_discovered": 2,' ' "num_base_bdevs_operational": 2,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc8",' ' "uuid": "91863e0b-b6e3-4ae1-85b3-dbd0ab80231f",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc9",' ' "uuid": "e8e513fa-3b18-46dd-ad2f-1e2ffbb5ae8e",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' '{' ' "name": "AIO0",' ' "aliases": [' ' "0afa1591-1717-40c4-a4dd-e5b16747b803"' ' ],' ' "product_name": "AIO disk",' ' "block_size": 2048,' ' "num_blocks": 5000,' ' "uuid": "0afa1591-1717-40c4-a4dd-e5b16747b803",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "aio": {' ' "filename": "/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/aiofile",' ' "block_size_override": true,' ' "readonly": false,' ' "fallocate": false' ' }' ' }' '}' 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_Malloc0]' 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=Malloc0 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_Malloc1p0]' 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=Malloc1p0 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_Malloc1p1]' 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=Malloc1p1 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_Malloc2p0]' 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=Malloc2p0 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_Malloc2p1]' 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=Malloc2p1 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_Malloc2p2]' 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=Malloc2p2 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_Malloc2p3]' 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=Malloc2p3 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_Malloc2p4]' 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=Malloc2p4 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_Malloc2p5]' 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=Malloc2p5 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_Malloc2p6]' 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=Malloc2p6 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_Malloc2p7]' 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=Malloc2p7 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_TestPT]' 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=TestPT 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_raid0]' 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=raid0 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_concat0]' 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=concat0 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- bdev/blockdev.sh@367 -- # run_test bdev_fio_trim fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --verify_state_save=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- common/autotest_common.sh@1099 -- # '[' 11 -le 1 ']' 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- common/autotest_common.sh@1105 -- # xtrace_disable 00:09:44.050 11:20:26 blockdev_general.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:09:44.050 ************************************ 00:09:44.050 START TEST bdev_fio_trim 00:09:44.050 ************************************ 00:09:44.050 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1123 -- # fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --verify_state_save=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:09:44.050 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --verify_state_save=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:09:44.050 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:09:44.050 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:09:44.050 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1339 -- # local sanitizers 00:09:44.050 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:09:44.050 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1341 -- # shift 00:09:44.050 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1343 -- # local asan_lib= 00:09:44.050 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:09:44.050 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:09:44.050 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # grep libasan 00:09:44.050 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:09:44.050 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # asan_lib= 00:09:44.050 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:09:44.050 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:09:44.050 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:09:44.050 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:09:44.051 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:09:44.051 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # asan_lib= 00:09:44.051 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:09:44.051 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev' 00:09:44.051 11:20:26 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --verify_state_save=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:09:44.051 job_Malloc0: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:44.051 job_Malloc1p0: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:44.051 job_Malloc1p1: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:44.051 job_Malloc2p0: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:44.051 job_Malloc2p1: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:44.051 job_Malloc2p2: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:44.051 job_Malloc2p3: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:44.051 job_Malloc2p4: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:44.051 job_Malloc2p5: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:44.051 job_Malloc2p6: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:44.051 job_Malloc2p7: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:44.051 job_TestPT: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:44.051 job_raid0: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:44.051 job_concat0: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:09:44.051 fio-3.35 00:09:44.051 Starting 14 threads 00:09:56.327 00:09:56.327 job_Malloc0: (groupid=0, jobs=14): err= 0: pid=849794: Mon Jul 15 11:20:38 2024 00:09:56.327 write: IOPS=124k, BW=486MiB/s (510MB/s)(4860MiB/10002msec); 0 zone resets 00:09:56.327 slat (usec): min=8, max=419, avg=39.95, stdev=11.02 00:09:56.327 clat (usec): min=30, max=3754, avg=281.36, stdev=99.43 00:09:56.327 lat (usec): min=48, max=3795, avg=321.31, stdev=103.73 00:09:56.327 clat percentiles (usec): 00:09:56.327 | 50.000th=[ 269], 99.000th=[ 515], 99.900th=[ 562], 99.990th=[ 603], 00:09:56.327 | 99.999th=[ 832] 00:09:56.327 bw ( KiB/s): min=435104, max=655897, per=100.00%, avg=499081.32, stdev=4305.77, samples=266 00:09:56.327 iops : min=108776, max=163969, avg=124770.26, stdev=1076.41, samples=266 00:09:56.327 trim: IOPS=124k, BW=486MiB/s (510MB/s)(4860MiB/10002msec); 0 zone resets 00:09:56.327 slat (usec): min=5, max=3401, avg=26.57, stdev= 7.72 00:09:56.327 clat (usec): min=49, max=3795, avg=321.48, stdev=103.74 00:09:56.327 lat (usec): min=56, max=3813, avg=348.05, stdev=106.75 00:09:56.327 clat percentiles (usec): 00:09:56.327 | 50.000th=[ 310], 99.000th=[ 562], 99.900th=[ 611], 99.990th=[ 652], 00:09:56.327 | 99.999th=[ 873] 00:09:56.327 bw ( KiB/s): min=435104, max=655889, per=100.00%, avg=499081.32, stdev=4305.76, samples=266 00:09:56.327 iops : min=108776, max=163967, avg=124770.16, stdev=1076.40, samples=266 00:09:56.327 lat (usec) : 50=0.01%, 100=0.59%, 250=34.56%, 500=61.24%, 750=3.60% 00:09:56.327 lat (usec) : 1000=0.01% 00:09:56.327 lat (msec) : 2=0.01%, 4=0.01% 00:09:56.327 cpu : usr=99.62%, sys=0.00%, ctx=607, majf=0, minf=1013 00:09:56.327 IO depths : 1=12.5%, 2=25.0%, 4=50.0%, 8=12.5%, 16=0.0%, 32=0.0%, >=64=0.0% 00:09:56.327 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:56.327 complete : 0=0.0%, 4=88.9%, 8=11.1%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:56.327 issued rwts: total=0,1244280,1244282,0 short=0,0,0,0 dropped=0,0,0,0 00:09:56.327 latency : target=0, window=0, percentile=100.00%, depth=8 00:09:56.327 00:09:56.327 Run status group 0 (all jobs): 00:09:56.327 WRITE: bw=486MiB/s (510MB/s), 486MiB/s-486MiB/s (510MB/s-510MB/s), io=4860MiB (5097MB), run=10002-10002msec 00:09:56.327 TRIM: bw=486MiB/s (510MB/s), 486MiB/s-486MiB/s (510MB/s-510MB/s), io=4860MiB (5097MB), run=10002-10002msec 00:09:56.327 00:09:56.327 real 0m11.847s 00:09:56.327 user 2m25.896s 00:09:56.327 sys 0m0.683s 00:09:56.327 11:20:38 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1124 -- # xtrace_disable 00:09:56.327 11:20:38 blockdev_general.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@10 -- # set +x 00:09:56.327 ************************************ 00:09:56.327 END TEST bdev_fio_trim 00:09:56.327 ************************************ 00:09:56.327 11:20:38 blockdev_general.bdev_fio -- common/autotest_common.sh@1142 -- # return 0 00:09:56.327 11:20:38 blockdev_general.bdev_fio -- bdev/blockdev.sh@368 -- # rm -f 00:09:56.327 11:20:38 blockdev_general.bdev_fio -- bdev/blockdev.sh@369 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:09:56.327 11:20:38 blockdev_general.bdev_fio -- bdev/blockdev.sh@370 -- # popd 00:09:56.327 /var/jenkins/workspace/crypto-phy-autotest/spdk 00:09:56.327 11:20:38 blockdev_general.bdev_fio -- bdev/blockdev.sh@371 -- # trap - SIGINT SIGTERM EXIT 00:09:56.327 00:09:56.327 real 0m23.754s 00:09:56.327 user 5m10.837s 00:09:56.327 sys 0m2.662s 00:09:56.327 11:20:38 blockdev_general.bdev_fio -- common/autotest_common.sh@1124 -- # xtrace_disable 00:09:56.327 11:20:38 blockdev_general.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:09:56.327 ************************************ 00:09:56.327 END TEST bdev_fio 00:09:56.327 ************************************ 00:09:56.327 11:20:38 blockdev_general -- common/autotest_common.sh@1142 -- # return 0 00:09:56.327 11:20:38 blockdev_general -- bdev/blockdev.sh@775 -- # trap cleanup SIGINT SIGTERM EXIT 00:09:56.327 11:20:38 blockdev_general -- bdev/blockdev.sh@777 -- # run_test bdev_verify /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:09:56.327 11:20:38 blockdev_general -- common/autotest_common.sh@1099 -- # '[' 16 -le 1 ']' 00:09:56.327 11:20:38 blockdev_general -- common/autotest_common.sh@1105 -- # xtrace_disable 00:09:56.327 11:20:38 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:09:56.327 ************************************ 00:09:56.327 START TEST bdev_verify 00:09:56.327 ************************************ 00:09:56.327 11:20:38 blockdev_general.bdev_verify -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:09:56.327 [2024-07-15 11:20:38.857756] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:09:56.327 [2024-07-15 11:20:38.857822] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid851246 ] 00:09:56.327 [2024-07-15 11:20:38.990120] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:09:56.327 [2024-07-15 11:20:39.096801] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:09:56.327 [2024-07-15 11:20:39.096806] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:09:56.328 [2024-07-15 11:20:39.252486] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc3 00:09:56.328 [2024-07-15 11:20:39.252545] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc3 00:09:56.328 [2024-07-15 11:20:39.252559] vbdev_passthru.c: 735:bdev_passthru_create_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:09:56.328 [2024-07-15 11:20:39.260487] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:09:56.328 [2024-07-15 11:20:39.260514] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:09:56.328 [2024-07-15 11:20:39.268504] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:09:56.328 [2024-07-15 11:20:39.268537] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:09:56.328 [2024-07-15 11:20:39.345724] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc3 00:09:56.328 [2024-07-15 11:20:39.345781] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:56.328 [2024-07-15 11:20:39.345801] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x289f4d0 00:09:56.328 [2024-07-15 11:20:39.345814] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:56.328 [2024-07-15 11:20:39.347455] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:56.328 [2024-07-15 11:20:39.347487] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: TestPT 00:09:56.328 Running I/O for 5 seconds... 00:10:01.590 00:10:01.590 Latency(us) 00:10:01.590 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:10:01.590 Job: Malloc0 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:10:01.590 Verification LBA range: start 0x0 length 0x1000 00:10:01.590 Malloc0 : 5.20 1083.97 4.23 0.00 0.00 117821.19 609.06 260776.29 00:10:01.590 Job: Malloc0 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:10:01.590 Verification LBA range: start 0x1000 length 0x1000 00:10:01.590 Malloc0 : 5.22 1078.38 4.21 0.00 0.00 118428.54 616.18 412135.96 00:10:01.590 Job: Malloc1p0 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:10:01.590 Verification LBA range: start 0x0 length 0x800 00:10:01.590 Malloc1p0 : 5.25 561.29 2.19 0.00 0.00 226672.66 3675.71 246187.41 00:10:01.590 Job: Malloc1p0 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x800 length 0x800 00:10:01.591 Malloc1p0 : 5.22 563.45 2.20 0.00 0.00 225844.66 3647.22 229774.91 00:10:01.591 Job: Malloc1p1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x0 length 0x800 00:10:01.591 Malloc1p1 : 5.25 560.90 2.19 0.00 0.00 226125.26 3519.00 240716.58 00:10:01.591 Job: Malloc1p1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x800 length 0x800 00:10:01.591 Malloc1p1 : 5.23 563.19 2.20 0.00 0.00 225237.03 3519.00 227951.30 00:10:01.591 Job: Malloc2p0 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x0 length 0x200 00:10:01.591 Malloc2p0 : 5.25 560.64 2.19 0.00 0.00 225541.66 3490.50 237069.36 00:10:01.591 Job: Malloc2p0 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x200 length 0x200 00:10:01.591 Malloc2p0 : 5.23 562.97 2.20 0.00 0.00 224615.50 3462.01 223392.28 00:10:01.591 Job: Malloc2p1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x0 length 0x200 00:10:01.591 Malloc2p1 : 5.25 560.27 2.19 0.00 0.00 224988.58 3447.76 233422.14 00:10:01.591 Job: Malloc2p1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x200 length 0x200 00:10:01.591 Malloc2p1 : 5.23 562.74 2.20 0.00 0.00 224007.63 3447.76 217921.45 00:10:01.591 Job: Malloc2p2 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x0 length 0x200 00:10:01.591 Malloc2p2 : 5.26 560.04 2.19 0.00 0.00 224414.61 3405.02 227039.50 00:10:01.591 Job: Malloc2p2 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x200 length 0x200 00:10:01.591 Malloc2p2 : 5.23 562.51 2.20 0.00 0.00 223432.68 3376.53 215186.03 00:10:01.591 Job: Malloc2p3 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x0 length 0x200 00:10:01.591 Malloc2p3 : 5.26 559.81 2.19 0.00 0.00 223840.91 3390.78 223392.28 00:10:01.591 Job: Malloc2p3 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x200 length 0x200 00:10:01.591 Malloc2p3 : 5.24 562.26 2.20 0.00 0.00 222873.95 3362.28 209715.20 00:10:01.591 Job: Malloc2p4 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x0 length 0x200 00:10:01.591 Malloc2p4 : 5.26 559.59 2.19 0.00 0.00 223270.83 3447.76 218833.25 00:10:01.591 Job: Malloc2p4 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x200 length 0x200 00:10:01.591 Malloc2p4 : 5.24 562.01 2.20 0.00 0.00 222333.32 3433.52 205156.17 00:10:01.591 Job: Malloc2p5 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x0 length 0x200 00:10:01.591 Malloc2p5 : 5.26 559.36 2.19 0.00 0.00 222701.35 3390.78 215186.03 00:10:01.591 Job: Malloc2p5 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x200 length 0x200 00:10:01.591 Malloc2p5 : 5.24 561.76 2.19 0.00 0.00 221782.24 3333.79 200597.15 00:10:01.591 Job: Malloc2p6 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x0 length 0x200 00:10:01.591 Malloc2p6 : 5.27 559.14 2.18 0.00 0.00 222121.49 3333.79 210627.01 00:10:01.591 Job: Malloc2p6 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x200 length 0x200 00:10:01.591 Malloc2p6 : 5.24 561.52 2.19 0.00 0.00 221209.23 3333.79 198773.54 00:10:01.591 Job: Malloc2p7 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x0 length 0x200 00:10:01.591 Malloc2p7 : 5.27 558.91 2.18 0.00 0.00 221563.55 3476.26 208803.39 00:10:01.591 Job: Malloc2p7 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x200 length 0x200 00:10:01.591 Malloc2p7 : 5.25 561.25 2.19 0.00 0.00 220666.22 3447.76 194214.51 00:10:01.591 Job: TestPT (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x0 length 0x1000 00:10:01.591 TestPT : 5.28 557.18 2.18 0.00 0.00 221481.10 13050.21 209715.20 00:10:01.591 Job: TestPT (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x1000 length 0x1000 00:10:01.591 TestPT : 5.27 538.51 2.10 0.00 0.00 228496.97 19147.91 280836.01 00:10:01.591 Job: raid0 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x0 length 0x2000 00:10:01.591 raid0 : 5.27 558.43 2.18 0.00 0.00 220120.31 3462.01 181449.24 00:10:01.591 Job: raid0 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x2000 length 0x2000 00:10:01.591 raid0 : 5.25 560.67 2.19 0.00 0.00 219256.66 3462.01 167772.16 00:10:01.591 Job: concat0 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x0 length 0x2000 00:10:01.591 concat0 : 5.27 558.18 2.18 0.00 0.00 219566.75 3490.50 178713.82 00:10:01.591 Job: concat0 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x2000 length 0x2000 00:10:01.591 concat0 : 5.28 581.88 2.27 0.00 0.00 210693.74 3476.26 168683.97 00:10:01.591 Job: raid1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x0 length 0x1000 00:10:01.591 raid1 : 5.28 557.92 2.18 0.00 0.00 219074.02 4017.64 175978.41 00:10:01.591 Job: raid1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x1000 length 0x1000 00:10:01.591 raid1 : 5.28 581.64 2.27 0.00 0.00 210202.25 4131.62 174154.80 00:10:01.591 Job: AIO0 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x0 length 0x4e2 00:10:01.591 AIO0 : 5.29 580.93 2.27 0.00 0.00 209737.00 819.20 178713.82 00:10:01.591 Job: AIO0 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:10:01.591 Verification LBA range: start 0x4e2 length 0x4e2 00:10:01.591 AIO0 : 5.28 581.44 2.27 0.00 0.00 209611.48 1524.42 182361.04 00:10:01.591 =================================================================================================================== 00:10:01.591 Total : 19042.73 74.39 0.00 0.00 209685.14 609.06 412135.96 00:10:01.849 00:10:01.849 real 0m6.574s 00:10:01.849 user 0m12.141s 00:10:01.849 sys 0m0.427s 00:10:01.849 11:20:45 blockdev_general.bdev_verify -- common/autotest_common.sh@1124 -- # xtrace_disable 00:10:01.850 11:20:45 blockdev_general.bdev_verify -- common/autotest_common.sh@10 -- # set +x 00:10:01.850 ************************************ 00:10:01.850 END TEST bdev_verify 00:10:01.850 ************************************ 00:10:01.850 11:20:45 blockdev_general -- common/autotest_common.sh@1142 -- # return 0 00:10:01.850 11:20:45 blockdev_general -- bdev/blockdev.sh@778 -- # run_test bdev_verify_big_io /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:10:01.850 11:20:45 blockdev_general -- common/autotest_common.sh@1099 -- # '[' 16 -le 1 ']' 00:10:01.850 11:20:45 blockdev_general -- common/autotest_common.sh@1105 -- # xtrace_disable 00:10:01.850 11:20:45 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:10:02.108 ************************************ 00:10:02.108 START TEST bdev_verify_big_io 00:10:02.108 ************************************ 00:10:02.108 11:20:45 blockdev_general.bdev_verify_big_io -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:10:02.108 [2024-07-15 11:20:45.510872] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:10:02.108 [2024-07-15 11:20:45.510939] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid852136 ] 00:10:02.108 [2024-07-15 11:20:45.637847] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:10:02.366 [2024-07-15 11:20:45.736438] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:10:02.366 [2024-07-15 11:20:45.736443] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:10:02.366 [2024-07-15 11:20:45.901364] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc3 00:10:02.366 [2024-07-15 11:20:45.901424] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc3 00:10:02.366 [2024-07-15 11:20:45.901439] vbdev_passthru.c: 735:bdev_passthru_create_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:10:02.366 [2024-07-15 11:20:45.909373] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:10:02.366 [2024-07-15 11:20:45.909400] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:10:02.366 [2024-07-15 11:20:45.917390] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:10:02.366 [2024-07-15 11:20:45.917416] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:10:02.625 [2024-07-15 11:20:45.994608] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc3 00:10:02.625 [2024-07-15 11:20:45.994660] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:02.625 [2024-07-15 11:20:45.994679] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x13ef4d0 00:10:02.625 [2024-07-15 11:20:45.994692] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:02.625 [2024-07-15 11:20:45.996323] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:02.625 [2024-07-15 11:20:45.996354] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: TestPT 00:10:02.625 [2024-07-15 11:20:46.185835] bdevperf.c:1818:bdevperf_construct_job: *WARNING*: Due to constraints of verify job, queue depth (-q, 128) can't exceed the number of IO requests which can be submitted to the bdev Malloc2p0 simultaneously (32). Queue depth is limited to 32 00:10:02.625 [2024-07-15 11:20:46.187362] bdevperf.c:1818:bdevperf_construct_job: *WARNING*: Due to constraints of verify job, queue depth (-q, 128) can't exceed the number of IO requests which can be submitted to the bdev Malloc2p0 simultaneously (32). Queue depth is limited to 32 00:10:02.626 [2024-07-15 11:20:46.189510] bdevperf.c:1818:bdevperf_construct_job: *WARNING*: Due to constraints of verify job, queue depth (-q, 128) can't exceed the number of IO requests which can be submitted to the bdev Malloc2p1 simultaneously (32). Queue depth is limited to 32 00:10:02.626 [2024-07-15 11:20:46.191017] bdevperf.c:1818:bdevperf_construct_job: *WARNING*: Due to constraints of verify job, queue depth (-q, 128) can't exceed the number of IO requests which can be submitted to the bdev Malloc2p1 simultaneously (32). Queue depth is limited to 32 00:10:02.626 [2024-07-15 11:20:46.193151] bdevperf.c:1818:bdevperf_construct_job: *WARNING*: Due to constraints of verify job, queue depth (-q, 128) can't exceed the number of IO requests which can be submitted to the bdev Malloc2p2 simultaneously (32). Queue depth is limited to 32 00:10:02.626 [2024-07-15 11:20:46.194597] bdevperf.c:1818:bdevperf_construct_job: *WARNING*: Due to constraints of verify job, queue depth (-q, 128) can't exceed the number of IO requests which can be submitted to the bdev Malloc2p2 simultaneously (32). Queue depth is limited to 32 00:10:02.626 [2024-07-15 11:20:46.196343] bdevperf.c:1818:bdevperf_construct_job: *WARNING*: Due to constraints of verify job, queue depth (-q, 128) can't exceed the number of IO requests which can be submitted to the bdev Malloc2p3 simultaneously (32). Queue depth is limited to 32 00:10:02.626 [2024-07-15 11:20:46.198005] bdevperf.c:1818:bdevperf_construct_job: *WARNING*: Due to constraints of verify job, queue depth (-q, 128) can't exceed the number of IO requests which can be submitted to the bdev Malloc2p3 simultaneously (32). Queue depth is limited to 32 00:10:02.626 [2024-07-15 11:20:46.199073] bdevperf.c:1818:bdevperf_construct_job: *WARNING*: Due to constraints of verify job, queue depth (-q, 128) can't exceed the number of IO requests which can be submitted to the bdev Malloc2p4 simultaneously (32). Queue depth is limited to 32 00:10:02.626 [2024-07-15 11:20:46.200708] bdevperf.c:1818:bdevperf_construct_job: *WARNING*: Due to constraints of verify job, queue depth (-q, 128) can't exceed the number of IO requests which can be submitted to the bdev Malloc2p4 simultaneously (32). Queue depth is limited to 32 00:10:02.626 [2024-07-15 11:20:46.201792] bdevperf.c:1818:bdevperf_construct_job: *WARNING*: Due to constraints of verify job, queue depth (-q, 128) can't exceed the number of IO requests which can be submitted to the bdev Malloc2p5 simultaneously (32). Queue depth is limited to 32 00:10:02.626 [2024-07-15 11:20:46.203465] bdevperf.c:1818:bdevperf_construct_job: *WARNING*: Due to constraints of verify job, queue depth (-q, 128) can't exceed the number of IO requests which can be submitted to the bdev Malloc2p5 simultaneously (32). Queue depth is limited to 32 00:10:02.626 [2024-07-15 11:20:46.204539] bdevperf.c:1818:bdevperf_construct_job: *WARNING*: Due to constraints of verify job, queue depth (-q, 128) can't exceed the number of IO requests which can be submitted to the bdev Malloc2p6 simultaneously (32). Queue depth is limited to 32 00:10:02.626 [2024-07-15 11:20:46.206020] bdevperf.c:1818:bdevperf_construct_job: *WARNING*: Due to constraints of verify job, queue depth (-q, 128) can't exceed the number of IO requests which can be submitted to the bdev Malloc2p6 simultaneously (32). Queue depth is limited to 32 00:10:02.626 [2024-07-15 11:20:46.206901] bdevperf.c:1818:bdevperf_construct_job: *WARNING*: Due to constraints of verify job, queue depth (-q, 128) can't exceed the number of IO requests which can be submitted to the bdev Malloc2p7 simultaneously (32). Queue depth is limited to 32 00:10:02.626 [2024-07-15 11:20:46.208328] bdevperf.c:1818:bdevperf_construct_job: *WARNING*: Due to constraints of verify job, queue depth (-q, 128) can't exceed the number of IO requests which can be submitted to the bdev Malloc2p7 simultaneously (32). Queue depth is limited to 32 00:10:02.883 [2024-07-15 11:20:46.232296] bdevperf.c:1818:bdevperf_construct_job: *WARNING*: Due to constraints of verify job, queue depth (-q, 128) can't exceed the number of IO requests which can be submitted to the bdev AIO0 simultaneously (78). Queue depth is limited to 78 00:10:02.883 [2024-07-15 11:20:46.234297] bdevperf.c:1818:bdevperf_construct_job: *WARNING*: Due to constraints of verify job, queue depth (-q, 128) can't exceed the number of IO requests which can be submitted to the bdev AIO0 simultaneously (78). Queue depth is limited to 78 00:10:02.883 Running I/O for 5 seconds... 00:10:10.991 00:10:10.991 Latency(us) 00:10:10.991 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:10:10.991 Job: Malloc0 (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x0 length 0x100 00:10:10.991 Malloc0 : 5.79 154.78 9.67 0.00 0.00 811016.84 886.87 2261276.94 00:10:10.991 Job: Malloc0 (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x100 length 0x100 00:10:10.991 Malloc0 : 5.98 149.91 9.37 0.00 0.00 837690.83 890.43 2596821.26 00:10:10.991 Job: Malloc1p0 (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x0 length 0x80 00:10:10.991 Malloc1p0 : 6.19 74.99 4.69 0.00 0.00 1570231.49 3177.07 2684354.56 00:10:10.991 Job: Malloc1p0 (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x80 length 0x80 00:10:10.991 Malloc1p0 : 6.50 54.18 3.39 0.00 0.00 2158105.64 2706.92 3545098.69 00:10:10.991 Job: Malloc1p1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x0 length 0x80 00:10:10.991 Malloc1p1 : 6.55 36.62 2.29 0.00 0.00 3073143.78 1695.39 5193642.52 00:10:10.991 Job: Malloc1p1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x80 length 0x80 00:10:10.991 Malloc1p1 : 6.78 37.77 2.36 0.00 0.00 2952046.79 1723.88 5076931.45 00:10:10.991 Job: Malloc2p0 (Core Mask 0x1, workload: verify, depth: 32, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x0 length 0x20 00:10:10.991 Malloc2p0 : 6.19 25.85 1.62 0.00 0.00 1100448.78 598.37 1911143.74 00:10:10.991 Job: Malloc2p0 (Core Mask 0x2, workload: verify, depth: 32, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x20 length 0x20 00:10:10.991 Malloc2p0 : 6.18 25.88 1.62 0.00 0.00 1089360.15 619.74 1677721.60 00:10:10.991 Job: Malloc2p1 (Core Mask 0x1, workload: verify, depth: 32, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x0 length 0x20 00:10:10.991 Malloc2p1 : 6.19 25.84 1.62 0.00 0.00 1090987.20 612.62 1881965.97 00:10:10.991 Job: Malloc2p1 (Core Mask 0x2, workload: verify, depth: 32, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x20 length 0x20 00:10:10.991 Malloc2p1 : 6.18 25.88 1.62 0.00 0.00 1079873.81 619.74 1648543.83 00:10:10.991 Job: Malloc2p2 (Core Mask 0x1, workload: verify, depth: 32, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x0 length 0x20 00:10:10.991 Malloc2p2 : 6.19 25.84 1.61 0.00 0.00 1081436.46 601.93 1860082.64 00:10:10.991 Job: Malloc2p2 (Core Mask 0x2, workload: verify, depth: 32, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x20 length 0x20 00:10:10.991 Malloc2p2 : 6.18 25.87 1.62 0.00 0.00 1070466.01 619.74 1626660.51 00:10:10.991 Job: Malloc2p3 (Core Mask 0x1, workload: verify, depth: 32, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x0 length 0x20 00:10:10.991 Malloc2p3 : 6.19 25.83 1.61 0.00 0.00 1071398.66 605.50 1823610.43 00:10:10.991 Job: Malloc2p3 (Core Mask 0x2, workload: verify, depth: 32, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x20 length 0x20 00:10:10.991 Malloc2p3 : 6.19 25.87 1.62 0.00 0.00 1060821.54 619.74 1604777.18 00:10:10.991 Job: Malloc2p4 (Core Mask 0x1, workload: verify, depth: 32, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x0 length 0x20 00:10:10.991 Malloc2p4 : 6.20 25.83 1.61 0.00 0.00 1061903.73 612.62 1801727.11 00:10:10.991 Job: Malloc2p4 (Core Mask 0x2, workload: verify, depth: 32, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x20 length 0x20 00:10:10.991 Malloc2p4 : 6.26 28.09 1.76 0.00 0.00 981343.19 623.30 1582893.86 00:10:10.991 Job: Malloc2p5 (Core Mask 0x1, workload: verify, depth: 32, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x0 length 0x20 00:10:10.991 Malloc2p5 : 6.20 25.82 1.61 0.00 0.00 1052528.52 612.62 1779843.78 00:10:10.991 Job: Malloc2p5 (Core Mask 0x2, workload: verify, depth: 32, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x20 length 0x20 00:10:10.991 Malloc2p5 : 6.27 28.09 1.76 0.00 0.00 972269.27 626.87 1553716.09 00:10:10.991 Job: Malloc2p6 (Core Mask 0x1, workload: verify, depth: 32, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x0 length 0x20 00:10:10.991 Malloc2p6 : 6.20 25.82 1.61 0.00 0.00 1043191.22 623.30 1757960.46 00:10:10.991 Job: Malloc2p6 (Core Mask 0x2, workload: verify, depth: 32, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x20 length 0x20 00:10:10.991 Malloc2p6 : 6.27 28.08 1.76 0.00 0.00 963592.08 644.67 1531832.77 00:10:10.991 Job: Malloc2p7 (Core Mask 0x1, workload: verify, depth: 32, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x0 length 0x20 00:10:10.991 Malloc2p7 : 6.20 25.81 1.61 0.00 0.00 1034072.16 648.24 1736077.13 00:10:10.991 Job: Malloc2p7 (Core Mask 0x2, workload: verify, depth: 32, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x20 length 0x20 00:10:10.991 Malloc2p7 : 6.27 28.08 1.75 0.00 0.00 954669.58 641.11 1509949.44 00:10:10.991 Job: TestPT (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x0 length 0x100 00:10:10.991 TestPT : 6.78 38.07 2.38 0.00 0.00 2668055.16 91636.42 3938998.54 00:10:10.991 Job: TestPT (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x100 length 0x100 00:10:10.991 TestPT : 6.55 36.64 2.29 0.00 0.00 2820103.36 79327.05 3690987.52 00:10:10.991 Job: raid0 (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x0 length 0x200 00:10:10.991 raid0 : 6.84 42.08 2.63 0.00 0.00 2335755.17 1595.66 4639264.95 00:10:10.991 Job: raid0 (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x200 length 0x200 00:10:10.991 raid0 : 6.55 48.84 3.05 0.00 0.00 2070347.94 1609.91 4522553.88 00:10:10.991 Job: concat0 (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x0 length 0x200 00:10:10.991 concat0 : 6.85 49.08 3.07 0.00 0.00 1973663.86 1581.41 4464198.34 00:10:10.991 Job: concat0 (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x200 length 0x200 00:10:10.991 concat0 : 6.64 52.99 3.31 0.00 0.00 1841126.37 1595.66 4347487.28 00:10:10.991 Job: raid1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x0 length 0x100 00:10:10.991 raid1 : 6.85 63.09 3.94 0.00 0.00 1519558.19 2037.31 4289131.74 00:10:10.991 Job: raid1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x100 length 0x100 00:10:10.991 raid1 : 6.83 60.93 3.81 0.00 0.00 1570380.58 2037.31 4172420.67 00:10:10.991 Job: AIO0 (Core Mask 0x1, workload: verify, depth: 78, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x0 length 0x4e 00:10:10.991 AIO0 : 6.85 66.27 4.14 0.00 0.00 861300.56 790.71 2728121.21 00:10:10.991 Job: AIO0 (Core Mask 0x2, workload: verify, depth: 78, IO size: 65536) 00:10:10.991 Verification LBA range: start 0x4e length 0x4e 00:10:10.991 AIO0 : 6.86 64.43 4.03 0.00 0.00 883958.29 819.20 2392576.89 00:10:10.991 =================================================================================================================== 00:10:10.991 Total : 1453.12 90.82 0.00 0.00 1435045.98 598.37 5193642.52 00:10:10.991 00:10:10.991 real 0m8.196s 00:10:10.991 user 0m15.402s 00:10:10.991 sys 0m0.403s 00:10:10.991 11:20:53 blockdev_general.bdev_verify_big_io -- common/autotest_common.sh@1124 -- # xtrace_disable 00:10:10.991 11:20:53 blockdev_general.bdev_verify_big_io -- common/autotest_common.sh@10 -- # set +x 00:10:10.991 ************************************ 00:10:10.991 END TEST bdev_verify_big_io 00:10:10.991 ************************************ 00:10:10.992 11:20:53 blockdev_general -- common/autotest_common.sh@1142 -- # return 0 00:10:10.992 11:20:53 blockdev_general -- bdev/blockdev.sh@779 -- # run_test bdev_write_zeroes /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:10:10.992 11:20:53 blockdev_general -- common/autotest_common.sh@1099 -- # '[' 13 -le 1 ']' 00:10:10.992 11:20:53 blockdev_general -- common/autotest_common.sh@1105 -- # xtrace_disable 00:10:10.992 11:20:53 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:10:10.992 ************************************ 00:10:10.992 START TEST bdev_write_zeroes 00:10:10.992 ************************************ 00:10:10.992 11:20:53 blockdev_general.bdev_write_zeroes -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:10:10.992 [2024-07-15 11:20:53.777557] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:10:10.992 [2024-07-15 11:20:53.777618] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid853204 ] 00:10:10.992 [2024-07-15 11:20:53.904343] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:10.992 [2024-07-15 11:20:54.004158] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:10:10.992 [2024-07-15 11:20:54.147393] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc3 00:10:10.992 [2024-07-15 11:20:54.147447] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc3 00:10:10.992 [2024-07-15 11:20:54.147462] vbdev_passthru.c: 735:bdev_passthru_create_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:10:10.992 [2024-07-15 11:20:54.155403] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:10:10.992 [2024-07-15 11:20:54.155431] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:10:10.992 [2024-07-15 11:20:54.163412] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:10:10.992 [2024-07-15 11:20:54.163437] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:10:10.992 [2024-07-15 11:20:54.235877] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc3 00:10:10.992 [2024-07-15 11:20:54.235935] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:10.992 [2024-07-15 11:20:54.235954] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x16f4c10 00:10:10.992 [2024-07-15 11:20:54.235967] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:10.992 [2024-07-15 11:20:54.237436] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:10.992 [2024-07-15 11:20:54.237466] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: TestPT 00:10:10.992 Running I/O for 1 seconds... 00:10:12.367 00:10:12.367 Latency(us) 00:10:12.367 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:10:12.367 Job: Malloc0 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:10:12.367 Malloc0 : 1.03 4974.68 19.43 0.00 0.00 25721.02 662.48 43310.75 00:10:12.367 Job: Malloc1p0 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:10:12.367 Malloc1p0 : 1.03 4967.47 19.40 0.00 0.00 25710.26 926.05 42398.94 00:10:12.367 Job: Malloc1p1 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:10:12.367 Malloc1p1 : 1.03 4959.92 19.37 0.00 0.00 25688.83 918.93 41487.14 00:10:12.367 Job: Malloc2p0 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:10:12.367 Malloc2p0 : 1.03 4952.32 19.34 0.00 0.00 25672.12 918.93 40575.33 00:10:12.367 Job: Malloc2p1 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:10:12.367 Malloc2p1 : 1.06 4968.56 19.41 0.00 0.00 25537.27 911.81 39891.48 00:10:12.367 Job: Malloc2p2 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:10:12.367 Malloc2p2 : 1.06 4961.18 19.38 0.00 0.00 25520.43 918.93 38979.67 00:10:12.367 Job: Malloc2p3 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:10:12.367 Malloc2p3 : 1.06 4953.79 19.35 0.00 0.00 25500.90 918.93 38067.87 00:10:12.367 Job: Malloc2p4 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:10:12.367 Malloc2p4 : 1.06 4946.42 19.32 0.00 0.00 25481.04 918.93 37156.06 00:10:12.367 Job: Malloc2p5 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:10:12.367 Malloc2p5 : 1.06 4939.14 19.29 0.00 0.00 25460.75 918.93 36244.26 00:10:12.368 Job: Malloc2p6 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:10:12.368 Malloc2p6 : 1.06 4931.83 19.26 0.00 0.00 25441.71 918.93 35332.45 00:10:12.368 Job: Malloc2p7 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:10:12.368 Malloc2p7 : 1.07 4924.59 19.24 0.00 0.00 25422.28 918.93 34192.70 00:10:12.368 Job: TestPT (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:10:12.368 TestPT : 1.07 4917.38 19.21 0.00 0.00 25402.67 940.30 33280.89 00:10:12.368 Job: raid0 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:10:12.368 raid0 : 1.07 4909.48 19.18 0.00 0.00 25374.98 1624.15 31685.23 00:10:12.368 Job: concat0 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:10:12.368 concat0 : 1.07 4901.75 19.15 0.00 0.00 25323.49 1624.15 30089.57 00:10:12.368 Job: raid1 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:10:12.368 raid1 : 1.07 4892.10 19.11 0.00 0.00 25266.26 2564.45 27468.13 00:10:12.368 Job: AIO0 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:10:12.368 AIO0 : 1.07 4886.21 19.09 0.00 0.00 25173.12 1075.65 26898.25 00:10:12.368 =================================================================================================================== 00:10:12.368 Total : 78986.82 308.54 0.00 0.00 25479.74 662.48 43310.75 00:10:12.368 00:10:12.368 real 0m2.179s 00:10:12.368 user 0m1.813s 00:10:12.368 sys 0m0.320s 00:10:12.368 11:20:55 blockdev_general.bdev_write_zeroes -- common/autotest_common.sh@1124 -- # xtrace_disable 00:10:12.368 11:20:55 blockdev_general.bdev_write_zeroes -- common/autotest_common.sh@10 -- # set +x 00:10:12.368 ************************************ 00:10:12.368 END TEST bdev_write_zeroes 00:10:12.368 ************************************ 00:10:12.368 11:20:55 blockdev_general -- common/autotest_common.sh@1142 -- # return 0 00:10:12.368 11:20:55 blockdev_general -- bdev/blockdev.sh@782 -- # run_test bdev_json_nonenclosed /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:10:12.368 11:20:55 blockdev_general -- common/autotest_common.sh@1099 -- # '[' 13 -le 1 ']' 00:10:12.368 11:20:55 blockdev_general -- common/autotest_common.sh@1105 -- # xtrace_disable 00:10:12.368 11:20:55 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:10:12.627 ************************************ 00:10:12.627 START TEST bdev_json_nonenclosed 00:10:12.627 ************************************ 00:10:12.627 11:20:55 blockdev_general.bdev_json_nonenclosed -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:10:12.627 [2024-07-15 11:20:56.078974] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:10:12.627 [2024-07-15 11:20:56.079104] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid853498 ] 00:10:12.886 [2024-07-15 11:20:56.276018] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:12.886 [2024-07-15 11:20:56.381088] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:10:12.886 [2024-07-15 11:20:56.381162] json_config.c: 608:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: not enclosed in {}. 00:10:12.886 [2024-07-15 11:20:56.381182] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:10:12.886 [2024-07-15 11:20:56.381196] app.c:1053:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:10:13.144 00:10:13.144 real 0m0.513s 00:10:13.144 user 0m0.296s 00:10:13.144 sys 0m0.212s 00:10:13.144 11:20:56 blockdev_general.bdev_json_nonenclosed -- common/autotest_common.sh@1123 -- # es=234 00:10:13.144 11:20:56 blockdev_general.bdev_json_nonenclosed -- common/autotest_common.sh@1124 -- # xtrace_disable 00:10:13.144 11:20:56 blockdev_general.bdev_json_nonenclosed -- common/autotest_common.sh@10 -- # set +x 00:10:13.144 ************************************ 00:10:13.144 END TEST bdev_json_nonenclosed 00:10:13.144 ************************************ 00:10:13.144 11:20:56 blockdev_general -- common/autotest_common.sh@1142 -- # return 234 00:10:13.144 11:20:56 blockdev_general -- bdev/blockdev.sh@782 -- # true 00:10:13.144 11:20:56 blockdev_general -- bdev/blockdev.sh@785 -- # run_test bdev_json_nonarray /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:10:13.144 11:20:56 blockdev_general -- common/autotest_common.sh@1099 -- # '[' 13 -le 1 ']' 00:10:13.144 11:20:56 blockdev_general -- common/autotest_common.sh@1105 -- # xtrace_disable 00:10:13.144 11:20:56 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:10:13.144 ************************************ 00:10:13.144 START TEST bdev_json_nonarray 00:10:13.144 ************************************ 00:10:13.144 11:20:56 blockdev_general.bdev_json_nonarray -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:10:13.144 [2024-07-15 11:20:56.667937] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:10:13.144 [2024-07-15 11:20:56.668071] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid853597 ] 00:10:13.403 [2024-07-15 11:20:56.866269] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:13.403 [2024-07-15 11:20:56.967859] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:10:13.403 [2024-07-15 11:20:56.967956] json_config.c: 614:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: 'subsystems' should be an array. 00:10:13.403 [2024-07-15 11:20:56.967978] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:10:13.403 [2024-07-15 11:20:56.967991] app.c:1053:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:10:13.661 00:10:13.661 real 0m0.508s 00:10:13.661 user 0m0.297s 00:10:13.661 sys 0m0.207s 00:10:13.661 11:20:57 blockdev_general.bdev_json_nonarray -- common/autotest_common.sh@1123 -- # es=234 00:10:13.661 11:20:57 blockdev_general.bdev_json_nonarray -- common/autotest_common.sh@1124 -- # xtrace_disable 00:10:13.661 11:20:57 blockdev_general.bdev_json_nonarray -- common/autotest_common.sh@10 -- # set +x 00:10:13.661 ************************************ 00:10:13.661 END TEST bdev_json_nonarray 00:10:13.661 ************************************ 00:10:13.661 11:20:57 blockdev_general -- common/autotest_common.sh@1142 -- # return 234 00:10:13.661 11:20:57 blockdev_general -- bdev/blockdev.sh@785 -- # true 00:10:13.661 11:20:57 blockdev_general -- bdev/blockdev.sh@787 -- # [[ bdev == bdev ]] 00:10:13.661 11:20:57 blockdev_general -- bdev/blockdev.sh@788 -- # run_test bdev_qos qos_test_suite '' 00:10:13.661 11:20:57 blockdev_general -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:10:13.661 11:20:57 blockdev_general -- common/autotest_common.sh@1105 -- # xtrace_disable 00:10:13.661 11:20:57 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:10:13.661 ************************************ 00:10:13.661 START TEST bdev_qos 00:10:13.661 ************************************ 00:10:13.661 11:20:57 blockdev_general.bdev_qos -- common/autotest_common.sh@1123 -- # qos_test_suite '' 00:10:13.661 11:20:57 blockdev_general.bdev_qos -- bdev/blockdev.sh@446 -- # QOS_PID=853638 00:10:13.662 11:20:57 blockdev_general.bdev_qos -- bdev/blockdev.sh@447 -- # echo 'Process qos testing pid: 853638' 00:10:13.662 Process qos testing pid: 853638 00:10:13.662 11:20:57 blockdev_general.bdev_qos -- bdev/blockdev.sh@445 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -z -m 0x2 -q 256 -o 4096 -w randread -t 60 '' 00:10:13.662 11:20:57 blockdev_general.bdev_qos -- bdev/blockdev.sh@448 -- # trap 'cleanup; killprocess $QOS_PID; exit 1' SIGINT SIGTERM EXIT 00:10:13.662 11:20:57 blockdev_general.bdev_qos -- bdev/blockdev.sh@449 -- # waitforlisten 853638 00:10:13.662 11:20:57 blockdev_general.bdev_qos -- common/autotest_common.sh@829 -- # '[' -z 853638 ']' 00:10:13.662 11:20:57 blockdev_general.bdev_qos -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:13.662 11:20:57 blockdev_general.bdev_qos -- common/autotest_common.sh@834 -- # local max_retries=100 00:10:13.662 11:20:57 blockdev_general.bdev_qos -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:13.662 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:13.662 11:20:57 blockdev_general.bdev_qos -- common/autotest_common.sh@838 -- # xtrace_disable 00:10:13.662 11:20:57 blockdev_general.bdev_qos -- common/autotest_common.sh@10 -- # set +x 00:10:13.662 [2024-07-15 11:20:57.219676] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:10:13.662 [2024-07-15 11:20:57.219743] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid853638 ] 00:10:13.920 [2024-07-15 11:20:57.337550] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:13.920 [2024-07-15 11:20:57.433884] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:10:14.855 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:10:14.855 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@862 -- # return 0 00:10:14.855 11:20:58 blockdev_general.bdev_qos -- bdev/blockdev.sh@451 -- # rpc_cmd bdev_malloc_create -b Malloc_0 128 512 00:10:14.855 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:14.855 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@10 -- # set +x 00:10:14.855 Malloc_0 00:10:14.855 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:14.855 11:20:58 blockdev_general.bdev_qos -- bdev/blockdev.sh@452 -- # waitforbdev Malloc_0 00:10:14.855 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@897 -- # local bdev_name=Malloc_0 00:10:14.855 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:10:14.855 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@899 -- # local i 00:10:14.855 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:10:14.855 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:10:14.855 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@902 -- # rpc_cmd bdev_wait_for_examine 00:10:14.855 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:14.855 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@10 -- # set +x 00:10:14.855 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:14.855 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@904 -- # rpc_cmd bdev_get_bdevs -b Malloc_0 -t 2000 00:10:14.855 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:14.855 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@10 -- # set +x 00:10:14.855 [ 00:10:14.855 { 00:10:14.855 "name": "Malloc_0", 00:10:14.855 "aliases": [ 00:10:14.855 "3ac3bb13-3608-45c1-a077-2f4e28241305" 00:10:14.855 ], 00:10:14.855 "product_name": "Malloc disk", 00:10:14.855 "block_size": 512, 00:10:14.855 "num_blocks": 262144, 00:10:14.855 "uuid": "3ac3bb13-3608-45c1-a077-2f4e28241305", 00:10:14.855 "assigned_rate_limits": { 00:10:14.855 "rw_ios_per_sec": 0, 00:10:14.855 "rw_mbytes_per_sec": 0, 00:10:14.855 "r_mbytes_per_sec": 0, 00:10:14.855 "w_mbytes_per_sec": 0 00:10:14.855 }, 00:10:14.855 "claimed": false, 00:10:14.855 "zoned": false, 00:10:14.855 "supported_io_types": { 00:10:14.855 "read": true, 00:10:14.855 "write": true, 00:10:14.855 "unmap": true, 00:10:14.855 "flush": true, 00:10:14.855 "reset": true, 00:10:14.855 "nvme_admin": false, 00:10:14.855 "nvme_io": false, 00:10:14.855 "nvme_io_md": false, 00:10:14.855 "write_zeroes": true, 00:10:14.855 "zcopy": true, 00:10:14.855 "get_zone_info": false, 00:10:14.855 "zone_management": false, 00:10:14.855 "zone_append": false, 00:10:14.855 "compare": false, 00:10:14.855 "compare_and_write": false, 00:10:14.855 "abort": true, 00:10:14.855 "seek_hole": false, 00:10:14.855 "seek_data": false, 00:10:14.856 "copy": true, 00:10:14.856 "nvme_iov_md": false 00:10:14.856 }, 00:10:14.856 "memory_domains": [ 00:10:14.856 { 00:10:14.856 "dma_device_id": "system", 00:10:14.856 "dma_device_type": 1 00:10:14.856 }, 00:10:14.856 { 00:10:14.856 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:14.856 "dma_device_type": 2 00:10:14.856 } 00:10:14.856 ], 00:10:14.856 "driver_specific": {} 00:10:14.856 } 00:10:14.856 ] 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@905 -- # return 0 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- bdev/blockdev.sh@453 -- # rpc_cmd bdev_null_create Null_1 128 512 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@10 -- # set +x 00:10:14.856 Null_1 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- bdev/blockdev.sh@454 -- # waitforbdev Null_1 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@897 -- # local bdev_name=Null_1 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@899 -- # local i 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@902 -- # rpc_cmd bdev_wait_for_examine 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@10 -- # set +x 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@904 -- # rpc_cmd bdev_get_bdevs -b Null_1 -t 2000 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@10 -- # set +x 00:10:14.856 [ 00:10:14.856 { 00:10:14.856 "name": "Null_1", 00:10:14.856 "aliases": [ 00:10:14.856 "4bccdfa9-a97d-4107-8c2f-0908ccebc8a6" 00:10:14.856 ], 00:10:14.856 "product_name": "Null disk", 00:10:14.856 "block_size": 512, 00:10:14.856 "num_blocks": 262144, 00:10:14.856 "uuid": "4bccdfa9-a97d-4107-8c2f-0908ccebc8a6", 00:10:14.856 "assigned_rate_limits": { 00:10:14.856 "rw_ios_per_sec": 0, 00:10:14.856 "rw_mbytes_per_sec": 0, 00:10:14.856 "r_mbytes_per_sec": 0, 00:10:14.856 "w_mbytes_per_sec": 0 00:10:14.856 }, 00:10:14.856 "claimed": false, 00:10:14.856 "zoned": false, 00:10:14.856 "supported_io_types": { 00:10:14.856 "read": true, 00:10:14.856 "write": true, 00:10:14.856 "unmap": false, 00:10:14.856 "flush": false, 00:10:14.856 "reset": true, 00:10:14.856 "nvme_admin": false, 00:10:14.856 "nvme_io": false, 00:10:14.856 "nvme_io_md": false, 00:10:14.856 "write_zeroes": true, 00:10:14.856 "zcopy": false, 00:10:14.856 "get_zone_info": false, 00:10:14.856 "zone_management": false, 00:10:14.856 "zone_append": false, 00:10:14.856 "compare": false, 00:10:14.856 "compare_and_write": false, 00:10:14.856 "abort": true, 00:10:14.856 "seek_hole": false, 00:10:14.856 "seek_data": false, 00:10:14.856 "copy": false, 00:10:14.856 "nvme_iov_md": false 00:10:14.856 }, 00:10:14.856 "driver_specific": {} 00:10:14.856 } 00:10:14.856 ] 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- common/autotest_common.sh@905 -- # return 0 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- bdev/blockdev.sh@457 -- # qos_function_test 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- bdev/blockdev.sh@410 -- # local qos_lower_iops_limit=1000 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- bdev/blockdev.sh@411 -- # local qos_lower_bw_limit=2 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- bdev/blockdev.sh@412 -- # local io_result=0 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- bdev/blockdev.sh@413 -- # local iops_limit=0 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- bdev/blockdev.sh@414 -- # local bw_limit=0 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- bdev/blockdev.sh@456 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- bdev/blockdev.sh@416 -- # get_io_result IOPS Malloc_0 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- bdev/blockdev.sh@375 -- # local limit_type=IOPS 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- bdev/blockdev.sh@376 -- # local qos_dev=Malloc_0 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- bdev/blockdev.sh@377 -- # local iostat_result 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- bdev/blockdev.sh@378 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/iostat.py -d -i 1 -t 5 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- bdev/blockdev.sh@378 -- # grep Malloc_0 00:10:14.856 11:20:58 blockdev_general.bdev_qos -- bdev/blockdev.sh@378 -- # tail -1 00:10:14.856 Running I/O for 60 seconds... 00:10:20.124 11:21:03 blockdev_general.bdev_qos -- bdev/blockdev.sh@378 -- # iostat_result='Malloc_0 62820.26 251281.03 0.00 0.00 252928.00 0.00 0.00 ' 00:10:20.124 11:21:03 blockdev_general.bdev_qos -- bdev/blockdev.sh@379 -- # '[' IOPS = IOPS ']' 00:10:20.124 11:21:03 blockdev_general.bdev_qos -- bdev/blockdev.sh@380 -- # awk '{print $2}' 00:10:20.124 11:21:03 blockdev_general.bdev_qos -- bdev/blockdev.sh@380 -- # iostat_result=62820.26 00:10:20.124 11:21:03 blockdev_general.bdev_qos -- bdev/blockdev.sh@385 -- # echo 62820 00:10:20.124 11:21:03 blockdev_general.bdev_qos -- bdev/blockdev.sh@416 -- # io_result=62820 00:10:20.124 11:21:03 blockdev_general.bdev_qos -- bdev/blockdev.sh@418 -- # iops_limit=15000 00:10:20.124 11:21:03 blockdev_general.bdev_qos -- bdev/blockdev.sh@419 -- # '[' 15000 -gt 1000 ']' 00:10:20.124 11:21:03 blockdev_general.bdev_qos -- bdev/blockdev.sh@422 -- # rpc_cmd bdev_set_qos_limit --rw_ios_per_sec 15000 Malloc_0 00:10:20.124 11:21:03 blockdev_general.bdev_qos -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:20.124 11:21:03 blockdev_general.bdev_qos -- common/autotest_common.sh@10 -- # set +x 00:10:20.124 11:21:03 blockdev_general.bdev_qos -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:20.124 11:21:03 blockdev_general.bdev_qos -- bdev/blockdev.sh@423 -- # run_test bdev_qos_iops run_qos_test 15000 IOPS Malloc_0 00:10:20.124 11:21:03 blockdev_general.bdev_qos -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:10:20.124 11:21:03 blockdev_general.bdev_qos -- common/autotest_common.sh@1105 -- # xtrace_disable 00:10:20.124 11:21:03 blockdev_general.bdev_qos -- common/autotest_common.sh@10 -- # set +x 00:10:20.124 ************************************ 00:10:20.124 START TEST bdev_qos_iops 00:10:20.124 ************************************ 00:10:20.124 11:21:03 blockdev_general.bdev_qos.bdev_qos_iops -- common/autotest_common.sh@1123 -- # run_qos_test 15000 IOPS Malloc_0 00:10:20.124 11:21:03 blockdev_general.bdev_qos.bdev_qos_iops -- bdev/blockdev.sh@389 -- # local qos_limit=15000 00:10:20.124 11:21:03 blockdev_general.bdev_qos.bdev_qos_iops -- bdev/blockdev.sh@390 -- # local qos_result=0 00:10:20.124 11:21:03 blockdev_general.bdev_qos.bdev_qos_iops -- bdev/blockdev.sh@392 -- # get_io_result IOPS Malloc_0 00:10:20.124 11:21:03 blockdev_general.bdev_qos.bdev_qos_iops -- bdev/blockdev.sh@375 -- # local limit_type=IOPS 00:10:20.124 11:21:03 blockdev_general.bdev_qos.bdev_qos_iops -- bdev/blockdev.sh@376 -- # local qos_dev=Malloc_0 00:10:20.124 11:21:03 blockdev_general.bdev_qos.bdev_qos_iops -- bdev/blockdev.sh@377 -- # local iostat_result 00:10:20.124 11:21:03 blockdev_general.bdev_qos.bdev_qos_iops -- bdev/blockdev.sh@378 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/iostat.py -d -i 1 -t 5 00:10:20.124 11:21:03 blockdev_general.bdev_qos.bdev_qos_iops -- bdev/blockdev.sh@378 -- # grep Malloc_0 00:10:20.124 11:21:03 blockdev_general.bdev_qos.bdev_qos_iops -- bdev/blockdev.sh@378 -- # tail -1 00:10:25.384 11:21:08 blockdev_general.bdev_qos.bdev_qos_iops -- bdev/blockdev.sh@378 -- # iostat_result='Malloc_0 14998.84 59995.37 0.00 0.00 60960.00 0.00 0.00 ' 00:10:25.384 11:21:08 blockdev_general.bdev_qos.bdev_qos_iops -- bdev/blockdev.sh@379 -- # '[' IOPS = IOPS ']' 00:10:25.384 11:21:08 blockdev_general.bdev_qos.bdev_qos_iops -- bdev/blockdev.sh@380 -- # awk '{print $2}' 00:10:25.384 11:21:08 blockdev_general.bdev_qos.bdev_qos_iops -- bdev/blockdev.sh@380 -- # iostat_result=14998.84 00:10:25.384 11:21:08 blockdev_general.bdev_qos.bdev_qos_iops -- bdev/blockdev.sh@385 -- # echo 14998 00:10:25.384 11:21:08 blockdev_general.bdev_qos.bdev_qos_iops -- bdev/blockdev.sh@392 -- # qos_result=14998 00:10:25.384 11:21:08 blockdev_general.bdev_qos.bdev_qos_iops -- bdev/blockdev.sh@393 -- # '[' IOPS = BANDWIDTH ']' 00:10:25.384 11:21:08 blockdev_general.bdev_qos.bdev_qos_iops -- bdev/blockdev.sh@396 -- # lower_limit=13500 00:10:25.384 11:21:08 blockdev_general.bdev_qos.bdev_qos_iops -- bdev/blockdev.sh@397 -- # upper_limit=16500 00:10:25.384 11:21:08 blockdev_general.bdev_qos.bdev_qos_iops -- bdev/blockdev.sh@400 -- # '[' 14998 -lt 13500 ']' 00:10:25.384 11:21:08 blockdev_general.bdev_qos.bdev_qos_iops -- bdev/blockdev.sh@400 -- # '[' 14998 -gt 16500 ']' 00:10:25.384 00:10:25.384 real 0m5.225s 00:10:25.384 user 0m0.099s 00:10:25.384 sys 0m0.044s 00:10:25.384 11:21:08 blockdev_general.bdev_qos.bdev_qos_iops -- common/autotest_common.sh@1124 -- # xtrace_disable 00:10:25.384 11:21:08 blockdev_general.bdev_qos.bdev_qos_iops -- common/autotest_common.sh@10 -- # set +x 00:10:25.384 ************************************ 00:10:25.384 END TEST bdev_qos_iops 00:10:25.384 ************************************ 00:10:25.384 11:21:08 blockdev_general.bdev_qos -- common/autotest_common.sh@1142 -- # return 0 00:10:25.384 11:21:08 blockdev_general.bdev_qos -- bdev/blockdev.sh@427 -- # get_io_result BANDWIDTH Null_1 00:10:25.385 11:21:08 blockdev_general.bdev_qos -- bdev/blockdev.sh@375 -- # local limit_type=BANDWIDTH 00:10:25.385 11:21:08 blockdev_general.bdev_qos -- bdev/blockdev.sh@376 -- # local qos_dev=Null_1 00:10:25.385 11:21:08 blockdev_general.bdev_qos -- bdev/blockdev.sh@377 -- # local iostat_result 00:10:25.385 11:21:08 blockdev_general.bdev_qos -- bdev/blockdev.sh@378 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/iostat.py -d -i 1 -t 5 00:10:25.385 11:21:08 blockdev_general.bdev_qos -- bdev/blockdev.sh@378 -- # grep Null_1 00:10:25.385 11:21:08 blockdev_general.bdev_qos -- bdev/blockdev.sh@378 -- # tail -1 00:10:30.686 11:21:13 blockdev_general.bdev_qos -- bdev/blockdev.sh@378 -- # iostat_result='Null_1 20156.56 80626.25 0.00 0.00 81920.00 0.00 0.00 ' 00:10:30.686 11:21:13 blockdev_general.bdev_qos -- bdev/blockdev.sh@379 -- # '[' BANDWIDTH = IOPS ']' 00:10:30.686 11:21:13 blockdev_general.bdev_qos -- bdev/blockdev.sh@381 -- # '[' BANDWIDTH = BANDWIDTH ']' 00:10:30.686 11:21:13 blockdev_general.bdev_qos -- bdev/blockdev.sh@382 -- # awk '{print $6}' 00:10:30.686 11:21:13 blockdev_general.bdev_qos -- bdev/blockdev.sh@382 -- # iostat_result=81920.00 00:10:30.686 11:21:13 blockdev_general.bdev_qos -- bdev/blockdev.sh@385 -- # echo 81920 00:10:30.686 11:21:13 blockdev_general.bdev_qos -- bdev/blockdev.sh@427 -- # bw_limit=81920 00:10:30.686 11:21:13 blockdev_general.bdev_qos -- bdev/blockdev.sh@428 -- # bw_limit=8 00:10:30.686 11:21:13 blockdev_general.bdev_qos -- bdev/blockdev.sh@429 -- # '[' 8 -lt 2 ']' 00:10:30.686 11:21:13 blockdev_general.bdev_qos -- bdev/blockdev.sh@432 -- # rpc_cmd bdev_set_qos_limit --rw_mbytes_per_sec 8 Null_1 00:10:30.686 11:21:13 blockdev_general.bdev_qos -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:30.686 11:21:13 blockdev_general.bdev_qos -- common/autotest_common.sh@10 -- # set +x 00:10:30.686 11:21:14 blockdev_general.bdev_qos -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:30.686 11:21:14 blockdev_general.bdev_qos -- bdev/blockdev.sh@433 -- # run_test bdev_qos_bw run_qos_test 8 BANDWIDTH Null_1 00:10:30.686 11:21:14 blockdev_general.bdev_qos -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:10:30.686 11:21:14 blockdev_general.bdev_qos -- common/autotest_common.sh@1105 -- # xtrace_disable 00:10:30.686 11:21:14 blockdev_general.bdev_qos -- common/autotest_common.sh@10 -- # set +x 00:10:30.686 ************************************ 00:10:30.686 START TEST bdev_qos_bw 00:10:30.686 ************************************ 00:10:30.686 11:21:14 blockdev_general.bdev_qos.bdev_qos_bw -- common/autotest_common.sh@1123 -- # run_qos_test 8 BANDWIDTH Null_1 00:10:30.686 11:21:14 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@389 -- # local qos_limit=8 00:10:30.686 11:21:14 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@390 -- # local qos_result=0 00:10:30.686 11:21:14 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@392 -- # get_io_result BANDWIDTH Null_1 00:10:30.686 11:21:14 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@375 -- # local limit_type=BANDWIDTH 00:10:30.686 11:21:14 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@376 -- # local qos_dev=Null_1 00:10:30.686 11:21:14 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@377 -- # local iostat_result 00:10:30.686 11:21:14 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@378 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/iostat.py -d -i 1 -t 5 00:10:30.686 11:21:14 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@378 -- # grep Null_1 00:10:30.686 11:21:14 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@378 -- # tail -1 00:10:35.951 11:21:19 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@378 -- # iostat_result='Null_1 2048.36 8193.45 0.00 0.00 8432.00 0.00 0.00 ' 00:10:35.951 11:21:19 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@379 -- # '[' BANDWIDTH = IOPS ']' 00:10:35.951 11:21:19 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@381 -- # '[' BANDWIDTH = BANDWIDTH ']' 00:10:35.951 11:21:19 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@382 -- # awk '{print $6}' 00:10:35.951 11:21:19 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@382 -- # iostat_result=8432.00 00:10:35.951 11:21:19 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@385 -- # echo 8432 00:10:35.951 11:21:19 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@392 -- # qos_result=8432 00:10:35.951 11:21:19 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@393 -- # '[' BANDWIDTH = BANDWIDTH ']' 00:10:35.951 11:21:19 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@394 -- # qos_limit=8192 00:10:35.951 11:21:19 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@396 -- # lower_limit=7372 00:10:35.951 11:21:19 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@397 -- # upper_limit=9011 00:10:35.951 11:21:19 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@400 -- # '[' 8432 -lt 7372 ']' 00:10:35.951 11:21:19 blockdev_general.bdev_qos.bdev_qos_bw -- bdev/blockdev.sh@400 -- # '[' 8432 -gt 9011 ']' 00:10:35.951 00:10:35.951 real 0m5.296s 00:10:35.951 user 0m0.109s 00:10:35.951 sys 0m0.051s 00:10:35.952 11:21:19 blockdev_general.bdev_qos.bdev_qos_bw -- common/autotest_common.sh@1124 -- # xtrace_disable 00:10:35.952 11:21:19 blockdev_general.bdev_qos.bdev_qos_bw -- common/autotest_common.sh@10 -- # set +x 00:10:35.952 ************************************ 00:10:35.952 END TEST bdev_qos_bw 00:10:35.952 ************************************ 00:10:35.952 11:21:19 blockdev_general.bdev_qos -- common/autotest_common.sh@1142 -- # return 0 00:10:35.952 11:21:19 blockdev_general.bdev_qos -- bdev/blockdev.sh@436 -- # rpc_cmd bdev_set_qos_limit --r_mbytes_per_sec 2 Malloc_0 00:10:35.952 11:21:19 blockdev_general.bdev_qos -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:35.952 11:21:19 blockdev_general.bdev_qos -- common/autotest_common.sh@10 -- # set +x 00:10:35.952 11:21:19 blockdev_general.bdev_qos -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:35.952 11:21:19 blockdev_general.bdev_qos -- bdev/blockdev.sh@437 -- # run_test bdev_qos_ro_bw run_qos_test 2 BANDWIDTH Malloc_0 00:10:35.952 11:21:19 blockdev_general.bdev_qos -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:10:35.952 11:21:19 blockdev_general.bdev_qos -- common/autotest_common.sh@1105 -- # xtrace_disable 00:10:35.952 11:21:19 blockdev_general.bdev_qos -- common/autotest_common.sh@10 -- # set +x 00:10:35.952 ************************************ 00:10:35.952 START TEST bdev_qos_ro_bw 00:10:35.952 ************************************ 00:10:35.952 11:21:19 blockdev_general.bdev_qos.bdev_qos_ro_bw -- common/autotest_common.sh@1123 -- # run_qos_test 2 BANDWIDTH Malloc_0 00:10:35.952 11:21:19 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@389 -- # local qos_limit=2 00:10:35.952 11:21:19 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@390 -- # local qos_result=0 00:10:35.952 11:21:19 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@392 -- # get_io_result BANDWIDTH Malloc_0 00:10:35.952 11:21:19 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@375 -- # local limit_type=BANDWIDTH 00:10:35.952 11:21:19 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@376 -- # local qos_dev=Malloc_0 00:10:35.952 11:21:19 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@377 -- # local iostat_result 00:10:35.952 11:21:19 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@378 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/iostat.py -d -i 1 -t 5 00:10:35.952 11:21:19 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@378 -- # grep Malloc_0 00:10:35.952 11:21:19 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@378 -- # tail -1 00:10:41.219 11:21:24 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@378 -- # iostat_result='Malloc_0 512.76 2051.03 0.00 0.00 2064.00 0.00 0.00 ' 00:10:41.219 11:21:24 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@379 -- # '[' BANDWIDTH = IOPS ']' 00:10:41.219 11:21:24 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@381 -- # '[' BANDWIDTH = BANDWIDTH ']' 00:10:41.219 11:21:24 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@382 -- # awk '{print $6}' 00:10:41.219 11:21:24 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@382 -- # iostat_result=2064.00 00:10:41.219 11:21:24 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@385 -- # echo 2064 00:10:41.219 11:21:24 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@392 -- # qos_result=2064 00:10:41.219 11:21:24 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@393 -- # '[' BANDWIDTH = BANDWIDTH ']' 00:10:41.219 11:21:24 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@394 -- # qos_limit=2048 00:10:41.219 11:21:24 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@396 -- # lower_limit=1843 00:10:41.219 11:21:24 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@397 -- # upper_limit=2252 00:10:41.219 11:21:24 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@400 -- # '[' 2064 -lt 1843 ']' 00:10:41.219 11:21:24 blockdev_general.bdev_qos.bdev_qos_ro_bw -- bdev/blockdev.sh@400 -- # '[' 2064 -gt 2252 ']' 00:10:41.219 00:10:41.219 real 0m5.180s 00:10:41.219 user 0m0.106s 00:10:41.219 sys 0m0.052s 00:10:41.219 11:21:24 blockdev_general.bdev_qos.bdev_qos_ro_bw -- common/autotest_common.sh@1124 -- # xtrace_disable 00:10:41.219 11:21:24 blockdev_general.bdev_qos.bdev_qos_ro_bw -- common/autotest_common.sh@10 -- # set +x 00:10:41.219 ************************************ 00:10:41.219 END TEST bdev_qos_ro_bw 00:10:41.219 ************************************ 00:10:41.219 11:21:24 blockdev_general.bdev_qos -- common/autotest_common.sh@1142 -- # return 0 00:10:41.219 11:21:24 blockdev_general.bdev_qos -- bdev/blockdev.sh@459 -- # rpc_cmd bdev_malloc_delete Malloc_0 00:10:41.219 11:21:24 blockdev_general.bdev_qos -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:41.219 11:21:24 blockdev_general.bdev_qos -- common/autotest_common.sh@10 -- # set +x 00:10:41.787 11:21:25 blockdev_general.bdev_qos -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:41.787 11:21:25 blockdev_general.bdev_qos -- bdev/blockdev.sh@460 -- # rpc_cmd bdev_null_delete Null_1 00:10:41.787 11:21:25 blockdev_general.bdev_qos -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:41.787 11:21:25 blockdev_general.bdev_qos -- common/autotest_common.sh@10 -- # set +x 00:10:41.787 00:10:41.787 Latency(us) 00:10:41.787 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:10:41.787 Job: Malloc_0 (Core Mask 0x2, workload: randread, depth: 256, IO size: 4096) 00:10:41.787 Malloc_0 : 26.70 20744.29 81.03 0.00 0.00 12224.22 2008.82 503316.48 00:10:41.787 Job: Null_1 (Core Mask 0x2, workload: randread, depth: 256, IO size: 4096) 00:10:41.787 Null_1 : 26.85 20390.05 79.65 0.00 0.00 12521.64 833.45 150447.86 00:10:41.787 =================================================================================================================== 00:10:41.787 Total : 41134.34 160.68 0.00 0.00 12372.07 833.45 503316.48 00:10:41.787 0 00:10:41.787 11:21:25 blockdev_general.bdev_qos -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:41.787 11:21:25 blockdev_general.bdev_qos -- bdev/blockdev.sh@461 -- # killprocess 853638 00:10:41.787 11:21:25 blockdev_general.bdev_qos -- common/autotest_common.sh@948 -- # '[' -z 853638 ']' 00:10:41.787 11:21:25 blockdev_general.bdev_qos -- common/autotest_common.sh@952 -- # kill -0 853638 00:10:41.787 11:21:25 blockdev_general.bdev_qos -- common/autotest_common.sh@953 -- # uname 00:10:41.787 11:21:25 blockdev_general.bdev_qos -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:10:42.046 11:21:25 blockdev_general.bdev_qos -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 853638 00:10:42.046 11:21:25 blockdev_general.bdev_qos -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:10:42.046 11:21:25 blockdev_general.bdev_qos -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:10:42.046 11:21:25 blockdev_general.bdev_qos -- common/autotest_common.sh@966 -- # echo 'killing process with pid 853638' 00:10:42.046 killing process with pid 853638 00:10:42.046 11:21:25 blockdev_general.bdev_qos -- common/autotest_common.sh@967 -- # kill 853638 00:10:42.046 Received shutdown signal, test time was about 26.911989 seconds 00:10:42.046 00:10:42.046 Latency(us) 00:10:42.046 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:10:42.046 =================================================================================================================== 00:10:42.046 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:10:42.046 11:21:25 blockdev_general.bdev_qos -- common/autotest_common.sh@972 -- # wait 853638 00:10:42.305 11:21:25 blockdev_general.bdev_qos -- bdev/blockdev.sh@462 -- # trap - SIGINT SIGTERM EXIT 00:10:42.305 00:10:42.305 real 0m28.493s 00:10:42.305 user 0m29.310s 00:10:42.305 sys 0m0.854s 00:10:42.305 11:21:25 blockdev_general.bdev_qos -- common/autotest_common.sh@1124 -- # xtrace_disable 00:10:42.305 11:21:25 blockdev_general.bdev_qos -- common/autotest_common.sh@10 -- # set +x 00:10:42.305 ************************************ 00:10:42.305 END TEST bdev_qos 00:10:42.305 ************************************ 00:10:42.305 11:21:25 blockdev_general -- common/autotest_common.sh@1142 -- # return 0 00:10:42.305 11:21:25 blockdev_general -- bdev/blockdev.sh@789 -- # run_test bdev_qd_sampling qd_sampling_test_suite '' 00:10:42.305 11:21:25 blockdev_general -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:10:42.305 11:21:25 blockdev_general -- common/autotest_common.sh@1105 -- # xtrace_disable 00:10:42.305 11:21:25 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:10:42.305 ************************************ 00:10:42.305 START TEST bdev_qd_sampling 00:10:42.305 ************************************ 00:10:42.305 11:21:25 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@1123 -- # qd_sampling_test_suite '' 00:10:42.305 11:21:25 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@538 -- # QD_DEV=Malloc_QD 00:10:42.305 11:21:25 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@541 -- # QD_PID=858034 00:10:42.305 11:21:25 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@542 -- # echo 'Process bdev QD sampling period testing pid: 858034' 00:10:42.305 Process bdev QD sampling period testing pid: 858034 00:10:42.305 11:21:25 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@540 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -z -m 0x3 -q 256 -o 4096 -w randread -t 5 -C '' 00:10:42.305 11:21:25 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@543 -- # trap 'cleanup; killprocess $QD_PID; exit 1' SIGINT SIGTERM EXIT 00:10:42.305 11:21:25 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@544 -- # waitforlisten 858034 00:10:42.305 11:21:25 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@829 -- # '[' -z 858034 ']' 00:10:42.305 11:21:25 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:42.305 11:21:25 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@834 -- # local max_retries=100 00:10:42.305 11:21:25 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:42.305 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:42.305 11:21:25 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@838 -- # xtrace_disable 00:10:42.305 11:21:25 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@10 -- # set +x 00:10:42.305 [2024-07-15 11:21:25.803993] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:10:42.305 [2024-07-15 11:21:25.804074] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid858034 ] 00:10:42.564 [2024-07-15 11:21:25.936651] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:10:42.564 [2024-07-15 11:21:26.042246] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:10:42.564 [2024-07-15 11:21:26.042251] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@862 -- # return 0 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@546 -- # rpc_cmd bdev_malloc_create -b Malloc_QD 128 512 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@10 -- # set +x 00:10:43.130 Malloc_QD 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@547 -- # waitforbdev Malloc_QD 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@897 -- # local bdev_name=Malloc_QD 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@899 -- # local i 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@902 -- # rpc_cmd bdev_wait_for_examine 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@10 -- # set +x 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@904 -- # rpc_cmd bdev_get_bdevs -b Malloc_QD -t 2000 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@10 -- # set +x 00:10:43.130 [ 00:10:43.130 { 00:10:43.130 "name": "Malloc_QD", 00:10:43.130 "aliases": [ 00:10:43.130 "badde97e-fe77-4191-a023-f28194ffc1b1" 00:10:43.130 ], 00:10:43.130 "product_name": "Malloc disk", 00:10:43.130 "block_size": 512, 00:10:43.130 "num_blocks": 262144, 00:10:43.130 "uuid": "badde97e-fe77-4191-a023-f28194ffc1b1", 00:10:43.130 "assigned_rate_limits": { 00:10:43.130 "rw_ios_per_sec": 0, 00:10:43.130 "rw_mbytes_per_sec": 0, 00:10:43.130 "r_mbytes_per_sec": 0, 00:10:43.130 "w_mbytes_per_sec": 0 00:10:43.130 }, 00:10:43.130 "claimed": false, 00:10:43.130 "zoned": false, 00:10:43.130 "supported_io_types": { 00:10:43.130 "read": true, 00:10:43.130 "write": true, 00:10:43.130 "unmap": true, 00:10:43.130 "flush": true, 00:10:43.130 "reset": true, 00:10:43.130 "nvme_admin": false, 00:10:43.130 "nvme_io": false, 00:10:43.130 "nvme_io_md": false, 00:10:43.130 "write_zeroes": true, 00:10:43.130 "zcopy": true, 00:10:43.130 "get_zone_info": false, 00:10:43.130 "zone_management": false, 00:10:43.130 "zone_append": false, 00:10:43.130 "compare": false, 00:10:43.130 "compare_and_write": false, 00:10:43.130 "abort": true, 00:10:43.130 "seek_hole": false, 00:10:43.130 "seek_data": false, 00:10:43.130 "copy": true, 00:10:43.130 "nvme_iov_md": false 00:10:43.130 }, 00:10:43.130 "memory_domains": [ 00:10:43.130 { 00:10:43.130 "dma_device_id": "system", 00:10:43.130 "dma_device_type": 1 00:10:43.130 }, 00:10:43.130 { 00:10:43.130 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:43.130 "dma_device_type": 2 00:10:43.130 } 00:10:43.130 ], 00:10:43.130 "driver_specific": {} 00:10:43.130 } 00:10:43.130 ] 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@905 -- # return 0 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@550 -- # sleep 2 00:10:43.130 11:21:26 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@549 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:43.389 Running I/O for 5 seconds... 00:10:45.286 11:21:28 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@551 -- # qd_sampling_function_test Malloc_QD 00:10:45.286 11:21:28 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@519 -- # local bdev_name=Malloc_QD 00:10:45.286 11:21:28 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@520 -- # local sampling_period=10 00:10:45.286 11:21:28 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@521 -- # local iostats 00:10:45.286 11:21:28 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@523 -- # rpc_cmd bdev_set_qd_sampling_period Malloc_QD 10 00:10:45.286 11:21:28 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:45.286 11:21:28 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@10 -- # set +x 00:10:45.286 11:21:28 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:45.287 11:21:28 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@525 -- # rpc_cmd bdev_get_iostat -b Malloc_QD 00:10:45.287 11:21:28 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:45.287 11:21:28 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@10 -- # set +x 00:10:45.287 11:21:28 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:45.287 11:21:28 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@525 -- # iostats='{ 00:10:45.287 "tick_rate": 2300000000, 00:10:45.287 "ticks": 5342093551083922, 00:10:45.287 "bdevs": [ 00:10:45.287 { 00:10:45.287 "name": "Malloc_QD", 00:10:45.287 "bytes_read": 774943232, 00:10:45.287 "num_read_ops": 189188, 00:10:45.287 "bytes_written": 0, 00:10:45.287 "num_write_ops": 0, 00:10:45.287 "bytes_unmapped": 0, 00:10:45.287 "num_unmap_ops": 0, 00:10:45.287 "bytes_copied": 0, 00:10:45.287 "num_copy_ops": 0, 00:10:45.287 "read_latency_ticks": 2231591866280, 00:10:45.287 "max_read_latency_ticks": 14632110, 00:10:45.287 "min_read_latency_ticks": 286006, 00:10:45.287 "write_latency_ticks": 0, 00:10:45.287 "max_write_latency_ticks": 0, 00:10:45.287 "min_write_latency_ticks": 0, 00:10:45.287 "unmap_latency_ticks": 0, 00:10:45.287 "max_unmap_latency_ticks": 0, 00:10:45.287 "min_unmap_latency_ticks": 0, 00:10:45.287 "copy_latency_ticks": 0, 00:10:45.287 "max_copy_latency_ticks": 0, 00:10:45.287 "min_copy_latency_ticks": 0, 00:10:45.287 "io_error": {}, 00:10:45.287 "queue_depth_polling_period": 10, 00:10:45.287 "queue_depth": 512, 00:10:45.287 "io_time": 20, 00:10:45.287 "weighted_io_time": 10240 00:10:45.287 } 00:10:45.287 ] 00:10:45.287 }' 00:10:45.287 11:21:28 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@527 -- # jq -r '.bdevs[0].queue_depth_polling_period' 00:10:45.287 11:21:28 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@527 -- # qd_sampling_period=10 00:10:45.287 11:21:28 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@529 -- # '[' 10 == null ']' 00:10:45.287 11:21:28 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@529 -- # '[' 10 -ne 10 ']' 00:10:45.287 11:21:28 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@553 -- # rpc_cmd bdev_malloc_delete Malloc_QD 00:10:45.287 11:21:28 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:45.287 11:21:28 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@10 -- # set +x 00:10:45.287 00:10:45.287 Latency(us) 00:10:45.287 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:10:45.287 Job: Malloc_QD (Core Mask 0x1, workload: randread, depth: 256, IO size: 4096) 00:10:45.287 Malloc_QD : 1.98 48994.07 191.38 0.00 0.00 5211.57 1880.60 5613.30 00:10:45.287 Job: Malloc_QD (Core Mask 0x2, workload: randread, depth: 256, IO size: 4096) 00:10:45.287 Malloc_QD : 1.98 50679.31 197.97 0.00 0.00 5038.28 1852.10 6382.64 00:10:45.287 =================================================================================================================== 00:10:45.287 Total : 99673.39 389.35 0.00 0.00 5123.46 1852.10 6382.64 00:10:45.287 0 00:10:45.287 11:21:28 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:45.287 11:21:28 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@554 -- # killprocess 858034 00:10:45.287 11:21:28 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@948 -- # '[' -z 858034 ']' 00:10:45.287 11:21:28 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@952 -- # kill -0 858034 00:10:45.287 11:21:28 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@953 -- # uname 00:10:45.287 11:21:28 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:10:45.287 11:21:28 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 858034 00:10:45.546 11:21:28 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:10:45.546 11:21:28 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:10:45.546 11:21:28 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@966 -- # echo 'killing process with pid 858034' 00:10:45.546 killing process with pid 858034 00:10:45.546 11:21:28 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@967 -- # kill 858034 00:10:45.546 Received shutdown signal, test time was about 2.052341 seconds 00:10:45.546 00:10:45.546 Latency(us) 00:10:45.546 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:10:45.546 =================================================================================================================== 00:10:45.546 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:10:45.546 11:21:28 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@972 -- # wait 858034 00:10:45.546 11:21:29 blockdev_general.bdev_qd_sampling -- bdev/blockdev.sh@555 -- # trap - SIGINT SIGTERM EXIT 00:10:45.546 00:10:45.546 real 0m3.362s 00:10:45.546 user 0m6.489s 00:10:45.546 sys 0m0.436s 00:10:45.546 11:21:29 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@1124 -- # xtrace_disable 00:10:45.546 11:21:29 blockdev_general.bdev_qd_sampling -- common/autotest_common.sh@10 -- # set +x 00:10:45.546 ************************************ 00:10:45.546 END TEST bdev_qd_sampling 00:10:45.546 ************************************ 00:10:45.805 11:21:29 blockdev_general -- common/autotest_common.sh@1142 -- # return 0 00:10:45.805 11:21:29 blockdev_general -- bdev/blockdev.sh@790 -- # run_test bdev_error error_test_suite '' 00:10:45.805 11:21:29 blockdev_general -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:10:45.805 11:21:29 blockdev_general -- common/autotest_common.sh@1105 -- # xtrace_disable 00:10:45.805 11:21:29 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:10:45.805 ************************************ 00:10:45.805 START TEST bdev_error 00:10:45.805 ************************************ 00:10:45.805 11:21:29 blockdev_general.bdev_error -- common/autotest_common.sh@1123 -- # error_test_suite '' 00:10:45.805 11:21:29 blockdev_general.bdev_error -- bdev/blockdev.sh@466 -- # DEV_1=Dev_1 00:10:45.805 11:21:29 blockdev_general.bdev_error -- bdev/blockdev.sh@467 -- # DEV_2=Dev_2 00:10:45.805 11:21:29 blockdev_general.bdev_error -- bdev/blockdev.sh@468 -- # ERR_DEV=EE_Dev_1 00:10:45.805 11:21:29 blockdev_general.bdev_error -- bdev/blockdev.sh@472 -- # ERR_PID=858476 00:10:45.805 11:21:29 blockdev_general.bdev_error -- bdev/blockdev.sh@473 -- # echo 'Process error testing pid: 858476' 00:10:45.805 Process error testing pid: 858476 00:10:45.805 11:21:29 blockdev_general.bdev_error -- bdev/blockdev.sh@474 -- # waitforlisten 858476 00:10:45.805 11:21:29 blockdev_general.bdev_error -- common/autotest_common.sh@829 -- # '[' -z 858476 ']' 00:10:45.805 11:21:29 blockdev_general.bdev_error -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:45.805 11:21:29 blockdev_general.bdev_error -- common/autotest_common.sh@834 -- # local max_retries=100 00:10:45.805 11:21:29 blockdev_general.bdev_error -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:45.805 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:45.805 11:21:29 blockdev_general.bdev_error -- common/autotest_common.sh@838 -- # xtrace_disable 00:10:45.805 11:21:29 blockdev_general.bdev_error -- bdev/blockdev.sh@471 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -z -m 0x2 -q 16 -o 4096 -w randread -t 5 -f '' 00:10:45.805 11:21:29 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:45.805 [2024-07-15 11:21:29.283604] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:10:45.805 [2024-07-15 11:21:29.283739] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid858476 ] 00:10:46.064 [2024-07-15 11:21:29.468940] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:46.064 [2024-07-15 11:21:29.571433] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:10:46.630 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:10:46.630 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@862 -- # return 0 00:10:46.630 11:21:30 blockdev_general.bdev_error -- bdev/blockdev.sh@476 -- # rpc_cmd bdev_malloc_create -b Dev_1 128 512 00:10:46.630 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:46.630 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:46.630 Dev_1 00:10:46.630 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:46.630 11:21:30 blockdev_general.bdev_error -- bdev/blockdev.sh@477 -- # waitforbdev Dev_1 00:10:46.630 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@897 -- # local bdev_name=Dev_1 00:10:46.630 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:10:46.630 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@899 -- # local i 00:10:46.630 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:10:46.630 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:10:46.630 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@902 -- # rpc_cmd bdev_wait_for_examine 00:10:46.630 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:46.630 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:46.630 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:46.630 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@904 -- # rpc_cmd bdev_get_bdevs -b Dev_1 -t 2000 00:10:46.630 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:46.630 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:46.889 [ 00:10:46.889 { 00:10:46.889 "name": "Dev_1", 00:10:46.889 "aliases": [ 00:10:46.889 "ab6d65b0-a316-40ba-a04a-42ad561f8f90" 00:10:46.889 ], 00:10:46.889 "product_name": "Malloc disk", 00:10:46.889 "block_size": 512, 00:10:46.889 "num_blocks": 262144, 00:10:46.889 "uuid": "ab6d65b0-a316-40ba-a04a-42ad561f8f90", 00:10:46.889 "assigned_rate_limits": { 00:10:46.889 "rw_ios_per_sec": 0, 00:10:46.889 "rw_mbytes_per_sec": 0, 00:10:46.889 "r_mbytes_per_sec": 0, 00:10:46.889 "w_mbytes_per_sec": 0 00:10:46.889 }, 00:10:46.889 "claimed": false, 00:10:46.889 "zoned": false, 00:10:46.889 "supported_io_types": { 00:10:46.889 "read": true, 00:10:46.889 "write": true, 00:10:46.889 "unmap": true, 00:10:46.889 "flush": true, 00:10:46.889 "reset": true, 00:10:46.889 "nvme_admin": false, 00:10:46.889 "nvme_io": false, 00:10:46.889 "nvme_io_md": false, 00:10:46.889 "write_zeroes": true, 00:10:46.889 "zcopy": true, 00:10:46.889 "get_zone_info": false, 00:10:46.889 "zone_management": false, 00:10:46.889 "zone_append": false, 00:10:46.889 "compare": false, 00:10:46.889 "compare_and_write": false, 00:10:46.889 "abort": true, 00:10:46.889 "seek_hole": false, 00:10:46.889 "seek_data": false, 00:10:46.889 "copy": true, 00:10:46.889 "nvme_iov_md": false 00:10:46.889 }, 00:10:46.889 "memory_domains": [ 00:10:46.889 { 00:10:46.889 "dma_device_id": "system", 00:10:46.889 "dma_device_type": 1 00:10:46.889 }, 00:10:46.889 { 00:10:46.889 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:46.889 "dma_device_type": 2 00:10:46.889 } 00:10:46.889 ], 00:10:46.889 "driver_specific": {} 00:10:46.889 } 00:10:46.889 ] 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@905 -- # return 0 00:10:46.889 11:21:30 blockdev_general.bdev_error -- bdev/blockdev.sh@478 -- # rpc_cmd bdev_error_create Dev_1 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:46.889 true 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:46.889 11:21:30 blockdev_general.bdev_error -- bdev/blockdev.sh@479 -- # rpc_cmd bdev_malloc_create -b Dev_2 128 512 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:46.889 Dev_2 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:46.889 11:21:30 blockdev_general.bdev_error -- bdev/blockdev.sh@480 -- # waitforbdev Dev_2 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@897 -- # local bdev_name=Dev_2 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@899 -- # local i 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@902 -- # rpc_cmd bdev_wait_for_examine 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@904 -- # rpc_cmd bdev_get_bdevs -b Dev_2 -t 2000 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:46.889 [ 00:10:46.889 { 00:10:46.889 "name": "Dev_2", 00:10:46.889 "aliases": [ 00:10:46.889 "8a1dad9b-f77e-4dc5-a6f6-031626a1135d" 00:10:46.889 ], 00:10:46.889 "product_name": "Malloc disk", 00:10:46.889 "block_size": 512, 00:10:46.889 "num_blocks": 262144, 00:10:46.889 "uuid": "8a1dad9b-f77e-4dc5-a6f6-031626a1135d", 00:10:46.889 "assigned_rate_limits": { 00:10:46.889 "rw_ios_per_sec": 0, 00:10:46.889 "rw_mbytes_per_sec": 0, 00:10:46.889 "r_mbytes_per_sec": 0, 00:10:46.889 "w_mbytes_per_sec": 0 00:10:46.889 }, 00:10:46.889 "claimed": false, 00:10:46.889 "zoned": false, 00:10:46.889 "supported_io_types": { 00:10:46.889 "read": true, 00:10:46.889 "write": true, 00:10:46.889 "unmap": true, 00:10:46.889 "flush": true, 00:10:46.889 "reset": true, 00:10:46.889 "nvme_admin": false, 00:10:46.889 "nvme_io": false, 00:10:46.889 "nvme_io_md": false, 00:10:46.889 "write_zeroes": true, 00:10:46.889 "zcopy": true, 00:10:46.889 "get_zone_info": false, 00:10:46.889 "zone_management": false, 00:10:46.889 "zone_append": false, 00:10:46.889 "compare": false, 00:10:46.889 "compare_and_write": false, 00:10:46.889 "abort": true, 00:10:46.889 "seek_hole": false, 00:10:46.889 "seek_data": false, 00:10:46.889 "copy": true, 00:10:46.889 "nvme_iov_md": false 00:10:46.889 }, 00:10:46.889 "memory_domains": [ 00:10:46.889 { 00:10:46.889 "dma_device_id": "system", 00:10:46.889 "dma_device_type": 1 00:10:46.889 }, 00:10:46.889 { 00:10:46.889 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:46.889 "dma_device_type": 2 00:10:46.889 } 00:10:46.889 ], 00:10:46.889 "driver_specific": {} 00:10:46.889 } 00:10:46.889 ] 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@905 -- # return 0 00:10:46.889 11:21:30 blockdev_general.bdev_error -- bdev/blockdev.sh@481 -- # rpc_cmd bdev_error_inject_error EE_Dev_1 all failure -n 5 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:46.889 11:21:30 blockdev_general.bdev_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:46.889 11:21:30 blockdev_general.bdev_error -- bdev/blockdev.sh@484 -- # sleep 1 00:10:46.889 11:21:30 blockdev_general.bdev_error -- bdev/blockdev.sh@483 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 1 perform_tests 00:10:46.889 Running I/O for 5 seconds... 00:10:47.822 11:21:31 blockdev_general.bdev_error -- bdev/blockdev.sh@487 -- # kill -0 858476 00:10:47.822 11:21:31 blockdev_general.bdev_error -- bdev/blockdev.sh@488 -- # echo 'Process is existed as continue on error is set. Pid: 858476' 00:10:47.822 Process is existed as continue on error is set. Pid: 858476 00:10:47.822 11:21:31 blockdev_general.bdev_error -- bdev/blockdev.sh@495 -- # rpc_cmd bdev_error_delete EE_Dev_1 00:10:47.822 11:21:31 blockdev_general.bdev_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:47.822 11:21:31 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:47.822 11:21:31 blockdev_general.bdev_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:47.822 11:21:31 blockdev_general.bdev_error -- bdev/blockdev.sh@496 -- # rpc_cmd bdev_malloc_delete Dev_1 00:10:47.822 11:21:31 blockdev_general.bdev_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:47.822 11:21:31 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:47.822 11:21:31 blockdev_general.bdev_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:47.822 11:21:31 blockdev_general.bdev_error -- bdev/blockdev.sh@497 -- # sleep 5 00:10:48.080 Timeout while waiting for response: 00:10:48.080 00:10:48.080 00:10:52.323 00:10:52.323 Latency(us) 00:10:52.323 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:10:52.323 Job: EE_Dev_1 (Core Mask 0x2, workload: randread, depth: 16, IO size: 4096) 00:10:52.323 EE_Dev_1 : 0.89 37444.27 146.27 5.60 0.00 423.76 129.11 669.61 00:10:52.323 Job: Dev_2 (Core Mask 0x2, workload: randread, depth: 16, IO size: 4096) 00:10:52.323 Dev_2 : 5.00 81280.54 317.50 0.00 0.00 193.33 66.78 22225.25 00:10:52.323 =================================================================================================================== 00:10:52.323 Total : 118724.81 463.77 5.60 0.00 210.84 66.78 22225.25 00:10:52.890 11:21:36 blockdev_general.bdev_error -- bdev/blockdev.sh@499 -- # killprocess 858476 00:10:52.890 11:21:36 blockdev_general.bdev_error -- common/autotest_common.sh@948 -- # '[' -z 858476 ']' 00:10:52.890 11:21:36 blockdev_general.bdev_error -- common/autotest_common.sh@952 -- # kill -0 858476 00:10:52.890 11:21:36 blockdev_general.bdev_error -- common/autotest_common.sh@953 -- # uname 00:10:52.890 11:21:36 blockdev_general.bdev_error -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:10:52.890 11:21:36 blockdev_general.bdev_error -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 858476 00:10:52.891 11:21:36 blockdev_general.bdev_error -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:10:52.891 11:21:36 blockdev_general.bdev_error -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:10:52.891 11:21:36 blockdev_general.bdev_error -- common/autotest_common.sh@966 -- # echo 'killing process with pid 858476' 00:10:52.891 killing process with pid 858476 00:10:52.891 11:21:36 blockdev_general.bdev_error -- common/autotest_common.sh@967 -- # kill 858476 00:10:52.891 Received shutdown signal, test time was about 5.000000 seconds 00:10:52.891 00:10:52.891 Latency(us) 00:10:52.891 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:10:52.891 =================================================================================================================== 00:10:52.891 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:10:52.891 11:21:36 blockdev_general.bdev_error -- common/autotest_common.sh@972 -- # wait 858476 00:10:53.150 11:21:36 blockdev_general.bdev_error -- bdev/blockdev.sh@503 -- # ERR_PID=859514 00:10:53.150 11:21:36 blockdev_general.bdev_error -- bdev/blockdev.sh@504 -- # echo 'Process error testing pid: 859514' 00:10:53.150 Process error testing pid: 859514 00:10:53.150 11:21:36 blockdev_general.bdev_error -- bdev/blockdev.sh@502 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -z -m 0x2 -q 16 -o 4096 -w randread -t 5 '' 00:10:53.150 11:21:36 blockdev_general.bdev_error -- bdev/blockdev.sh@505 -- # waitforlisten 859514 00:10:53.150 11:21:36 blockdev_general.bdev_error -- common/autotest_common.sh@829 -- # '[' -z 859514 ']' 00:10:53.150 11:21:36 blockdev_general.bdev_error -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:53.150 11:21:36 blockdev_general.bdev_error -- common/autotest_common.sh@834 -- # local max_retries=100 00:10:53.150 11:21:36 blockdev_general.bdev_error -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:53.150 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:53.150 11:21:36 blockdev_general.bdev_error -- common/autotest_common.sh@838 -- # xtrace_disable 00:10:53.150 11:21:36 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:53.409 [2024-07-15 11:21:36.758351] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:10:53.409 [2024-07-15 11:21:36.758431] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid859514 ] 00:10:53.409 [2024-07-15 11:21:36.878898] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:53.409 [2024-07-15 11:21:36.974982] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:10:54.389 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:10:54.389 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@862 -- # return 0 00:10:54.389 11:21:37 blockdev_general.bdev_error -- bdev/blockdev.sh@507 -- # rpc_cmd bdev_malloc_create -b Dev_1 128 512 00:10:54.389 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:54.389 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:54.389 Dev_1 00:10:54.389 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:54.389 11:21:37 blockdev_general.bdev_error -- bdev/blockdev.sh@508 -- # waitforbdev Dev_1 00:10:54.389 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@897 -- # local bdev_name=Dev_1 00:10:54.389 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:10:54.389 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@899 -- # local i 00:10:54.389 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:10:54.389 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:10:54.389 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@902 -- # rpc_cmd bdev_wait_for_examine 00:10:54.389 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:54.389 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:54.389 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:54.389 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@904 -- # rpc_cmd bdev_get_bdevs -b Dev_1 -t 2000 00:10:54.389 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:54.389 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:54.389 [ 00:10:54.389 { 00:10:54.389 "name": "Dev_1", 00:10:54.389 "aliases": [ 00:10:54.389 "5a743ebf-c250-4551-9de3-55d73dae2cc8" 00:10:54.389 ], 00:10:54.389 "product_name": "Malloc disk", 00:10:54.389 "block_size": 512, 00:10:54.389 "num_blocks": 262144, 00:10:54.389 "uuid": "5a743ebf-c250-4551-9de3-55d73dae2cc8", 00:10:54.389 "assigned_rate_limits": { 00:10:54.389 "rw_ios_per_sec": 0, 00:10:54.389 "rw_mbytes_per_sec": 0, 00:10:54.389 "r_mbytes_per_sec": 0, 00:10:54.389 "w_mbytes_per_sec": 0 00:10:54.389 }, 00:10:54.389 "claimed": false, 00:10:54.389 "zoned": false, 00:10:54.389 "supported_io_types": { 00:10:54.389 "read": true, 00:10:54.389 "write": true, 00:10:54.389 "unmap": true, 00:10:54.389 "flush": true, 00:10:54.389 "reset": true, 00:10:54.389 "nvme_admin": false, 00:10:54.389 "nvme_io": false, 00:10:54.389 "nvme_io_md": false, 00:10:54.389 "write_zeroes": true, 00:10:54.389 "zcopy": true, 00:10:54.389 "get_zone_info": false, 00:10:54.389 "zone_management": false, 00:10:54.389 "zone_append": false, 00:10:54.389 "compare": false, 00:10:54.389 "compare_and_write": false, 00:10:54.389 "abort": true, 00:10:54.389 "seek_hole": false, 00:10:54.389 "seek_data": false, 00:10:54.389 "copy": true, 00:10:54.389 "nvme_iov_md": false 00:10:54.389 }, 00:10:54.389 "memory_domains": [ 00:10:54.389 { 00:10:54.389 "dma_device_id": "system", 00:10:54.389 "dma_device_type": 1 00:10:54.389 }, 00:10:54.389 { 00:10:54.389 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:54.389 "dma_device_type": 2 00:10:54.390 } 00:10:54.390 ], 00:10:54.390 "driver_specific": {} 00:10:54.390 } 00:10:54.390 ] 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@905 -- # return 0 00:10:54.390 11:21:37 blockdev_general.bdev_error -- bdev/blockdev.sh@509 -- # rpc_cmd bdev_error_create Dev_1 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:54.390 true 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:54.390 11:21:37 blockdev_general.bdev_error -- bdev/blockdev.sh@510 -- # rpc_cmd bdev_malloc_create -b Dev_2 128 512 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:54.390 Dev_2 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:54.390 11:21:37 blockdev_general.bdev_error -- bdev/blockdev.sh@511 -- # waitforbdev Dev_2 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@897 -- # local bdev_name=Dev_2 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@899 -- # local i 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@902 -- # rpc_cmd bdev_wait_for_examine 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@904 -- # rpc_cmd bdev_get_bdevs -b Dev_2 -t 2000 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:54.390 [ 00:10:54.390 { 00:10:54.390 "name": "Dev_2", 00:10:54.390 "aliases": [ 00:10:54.390 "f242434d-b7ed-4b00-bb57-c8d449365152" 00:10:54.390 ], 00:10:54.390 "product_name": "Malloc disk", 00:10:54.390 "block_size": 512, 00:10:54.390 "num_blocks": 262144, 00:10:54.390 "uuid": "f242434d-b7ed-4b00-bb57-c8d449365152", 00:10:54.390 "assigned_rate_limits": { 00:10:54.390 "rw_ios_per_sec": 0, 00:10:54.390 "rw_mbytes_per_sec": 0, 00:10:54.390 "r_mbytes_per_sec": 0, 00:10:54.390 "w_mbytes_per_sec": 0 00:10:54.390 }, 00:10:54.390 "claimed": false, 00:10:54.390 "zoned": false, 00:10:54.390 "supported_io_types": { 00:10:54.390 "read": true, 00:10:54.390 "write": true, 00:10:54.390 "unmap": true, 00:10:54.390 "flush": true, 00:10:54.390 "reset": true, 00:10:54.390 "nvme_admin": false, 00:10:54.390 "nvme_io": false, 00:10:54.390 "nvme_io_md": false, 00:10:54.390 "write_zeroes": true, 00:10:54.390 "zcopy": true, 00:10:54.390 "get_zone_info": false, 00:10:54.390 "zone_management": false, 00:10:54.390 "zone_append": false, 00:10:54.390 "compare": false, 00:10:54.390 "compare_and_write": false, 00:10:54.390 "abort": true, 00:10:54.390 "seek_hole": false, 00:10:54.390 "seek_data": false, 00:10:54.390 "copy": true, 00:10:54.390 "nvme_iov_md": false 00:10:54.390 }, 00:10:54.390 "memory_domains": [ 00:10:54.390 { 00:10:54.390 "dma_device_id": "system", 00:10:54.390 "dma_device_type": 1 00:10:54.390 }, 00:10:54.390 { 00:10:54.390 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:54.390 "dma_device_type": 2 00:10:54.390 } 00:10:54.390 ], 00:10:54.390 "driver_specific": {} 00:10:54.390 } 00:10:54.390 ] 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@905 -- # return 0 00:10:54.390 11:21:37 blockdev_general.bdev_error -- bdev/blockdev.sh@512 -- # rpc_cmd bdev_error_inject_error EE_Dev_1 all failure -n 5 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:54.390 11:21:37 blockdev_general.bdev_error -- bdev/blockdev.sh@515 -- # NOT wait 859514 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@648 -- # local es=0 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@650 -- # valid_exec_arg wait 859514 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@636 -- # local arg=wait 00:10:54.390 11:21:37 blockdev_general.bdev_error -- bdev/blockdev.sh@514 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 1 perform_tests 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@640 -- # type -t wait 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:10:54.390 11:21:37 blockdev_general.bdev_error -- common/autotest_common.sh@651 -- # wait 859514 00:10:54.649 Running I/O for 5 seconds... 00:10:54.649 task offset: 23000 on job bdev=EE_Dev_1 fails 00:10:54.649 00:10:54.649 Latency(us) 00:10:54.649 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:10:54.649 Job: EE_Dev_1 (Core Mask 0x2, workload: randread, depth: 16, IO size: 4096) 00:10:54.649 Job: EE_Dev_1 ended in about 0.00 seconds with error 00:10:54.649 EE_Dev_1 : 0.00 29372.50 114.74 6675.57 0.00 366.97 132.67 655.36 00:10:54.649 Job: Dev_2 (Core Mask 0x2, workload: randread, depth: 16, IO size: 4096) 00:10:54.649 Dev_2 : 0.00 18181.82 71.02 0.00 0.00 652.99 138.02 1203.87 00:10:54.649 =================================================================================================================== 00:10:54.649 Total : 47554.31 185.76 6675.57 0.00 522.10 132.67 1203.87 00:10:54.649 [2024-07-15 11:21:38.044205] app.c:1053:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:10:54.649 request: 00:10:54.649 { 00:10:54.649 "method": "perform_tests", 00:10:54.649 "req_id": 1 00:10:54.649 } 00:10:54.649 Got JSON-RPC error response 00:10:54.649 response: 00:10:54.649 { 00:10:54.649 "code": -32603, 00:10:54.649 "message": "bdevperf failed with error Operation not permitted" 00:10:54.649 } 00:10:54.909 11:21:38 blockdev_general.bdev_error -- common/autotest_common.sh@651 -- # es=255 00:10:54.909 11:21:38 blockdev_general.bdev_error -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:10:54.909 11:21:38 blockdev_general.bdev_error -- common/autotest_common.sh@660 -- # es=127 00:10:54.909 11:21:38 blockdev_general.bdev_error -- common/autotest_common.sh@661 -- # case "$es" in 00:10:54.909 11:21:38 blockdev_general.bdev_error -- common/autotest_common.sh@668 -- # es=1 00:10:54.909 11:21:38 blockdev_general.bdev_error -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:10:54.909 00:10:54.909 real 0m9.167s 00:10:54.909 user 0m9.646s 00:10:54.909 sys 0m0.914s 00:10:54.909 11:21:38 blockdev_general.bdev_error -- common/autotest_common.sh@1124 -- # xtrace_disable 00:10:54.909 11:21:38 blockdev_general.bdev_error -- common/autotest_common.sh@10 -- # set +x 00:10:54.909 ************************************ 00:10:54.909 END TEST bdev_error 00:10:54.909 ************************************ 00:10:54.909 11:21:38 blockdev_general -- common/autotest_common.sh@1142 -- # return 0 00:10:54.909 11:21:38 blockdev_general -- bdev/blockdev.sh@791 -- # run_test bdev_stat stat_test_suite '' 00:10:54.909 11:21:38 blockdev_general -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:10:54.909 11:21:38 blockdev_general -- common/autotest_common.sh@1105 -- # xtrace_disable 00:10:54.909 11:21:38 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:10:54.909 ************************************ 00:10:54.909 START TEST bdev_stat 00:10:54.909 ************************************ 00:10:54.909 11:21:38 blockdev_general.bdev_stat -- common/autotest_common.sh@1123 -- # stat_test_suite '' 00:10:54.909 11:21:38 blockdev_general.bdev_stat -- bdev/blockdev.sh@592 -- # STAT_DEV=Malloc_STAT 00:10:54.909 11:21:38 blockdev_general.bdev_stat -- bdev/blockdev.sh@596 -- # STAT_PID=859744 00:10:54.909 11:21:38 blockdev_general.bdev_stat -- bdev/blockdev.sh@597 -- # echo 'Process Bdev IO statistics testing pid: 859744' 00:10:54.909 Process Bdev IO statistics testing pid: 859744 00:10:54.909 11:21:38 blockdev_general.bdev_stat -- bdev/blockdev.sh@595 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -z -m 0x3 -q 256 -o 4096 -w randread -t 10 -C '' 00:10:54.909 11:21:38 blockdev_general.bdev_stat -- bdev/blockdev.sh@598 -- # trap 'cleanup; killprocess $STAT_PID; exit 1' SIGINT SIGTERM EXIT 00:10:54.909 11:21:38 blockdev_general.bdev_stat -- bdev/blockdev.sh@599 -- # waitforlisten 859744 00:10:54.909 11:21:38 blockdev_general.bdev_stat -- common/autotest_common.sh@829 -- # '[' -z 859744 ']' 00:10:54.909 11:21:38 blockdev_general.bdev_stat -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:54.909 11:21:38 blockdev_general.bdev_stat -- common/autotest_common.sh@834 -- # local max_retries=100 00:10:54.910 11:21:38 blockdev_general.bdev_stat -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:54.910 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:54.910 11:21:38 blockdev_general.bdev_stat -- common/autotest_common.sh@838 -- # xtrace_disable 00:10:54.910 11:21:38 blockdev_general.bdev_stat -- common/autotest_common.sh@10 -- # set +x 00:10:54.910 [2024-07-15 11:21:38.493614] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:10:54.910 [2024-07-15 11:21:38.493689] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid859744 ] 00:10:55.168 [2024-07-15 11:21:38.625909] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:10:55.168 [2024-07-15 11:21:38.730046] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:10:55.168 [2024-07-15 11:21:38.730053] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- common/autotest_common.sh@862 -- # return 0 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- bdev/blockdev.sh@601 -- # rpc_cmd bdev_malloc_create -b Malloc_STAT 128 512 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- common/autotest_common.sh@10 -- # set +x 00:10:56.104 Malloc_STAT 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- bdev/blockdev.sh@602 -- # waitforbdev Malloc_STAT 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- common/autotest_common.sh@897 -- # local bdev_name=Malloc_STAT 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- common/autotest_common.sh@899 -- # local i 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- common/autotest_common.sh@902 -- # rpc_cmd bdev_wait_for_examine 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- common/autotest_common.sh@10 -- # set +x 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- common/autotest_common.sh@904 -- # rpc_cmd bdev_get_bdevs -b Malloc_STAT -t 2000 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- common/autotest_common.sh@10 -- # set +x 00:10:56.104 [ 00:10:56.104 { 00:10:56.104 "name": "Malloc_STAT", 00:10:56.104 "aliases": [ 00:10:56.104 "56d7c606-1c1f-4dbc-9a0d-cc8009c8140d" 00:10:56.104 ], 00:10:56.104 "product_name": "Malloc disk", 00:10:56.104 "block_size": 512, 00:10:56.104 "num_blocks": 262144, 00:10:56.104 "uuid": "56d7c606-1c1f-4dbc-9a0d-cc8009c8140d", 00:10:56.104 "assigned_rate_limits": { 00:10:56.104 "rw_ios_per_sec": 0, 00:10:56.104 "rw_mbytes_per_sec": 0, 00:10:56.104 "r_mbytes_per_sec": 0, 00:10:56.104 "w_mbytes_per_sec": 0 00:10:56.104 }, 00:10:56.104 "claimed": false, 00:10:56.104 "zoned": false, 00:10:56.104 "supported_io_types": { 00:10:56.104 "read": true, 00:10:56.104 "write": true, 00:10:56.104 "unmap": true, 00:10:56.104 "flush": true, 00:10:56.104 "reset": true, 00:10:56.104 "nvme_admin": false, 00:10:56.104 "nvme_io": false, 00:10:56.104 "nvme_io_md": false, 00:10:56.104 "write_zeroes": true, 00:10:56.104 "zcopy": true, 00:10:56.104 "get_zone_info": false, 00:10:56.104 "zone_management": false, 00:10:56.104 "zone_append": false, 00:10:56.104 "compare": false, 00:10:56.104 "compare_and_write": false, 00:10:56.104 "abort": true, 00:10:56.104 "seek_hole": false, 00:10:56.104 "seek_data": false, 00:10:56.104 "copy": true, 00:10:56.104 "nvme_iov_md": false 00:10:56.104 }, 00:10:56.104 "memory_domains": [ 00:10:56.104 { 00:10:56.104 "dma_device_id": "system", 00:10:56.104 "dma_device_type": 1 00:10:56.104 }, 00:10:56.104 { 00:10:56.104 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:56.104 "dma_device_type": 2 00:10:56.104 } 00:10:56.104 ], 00:10:56.104 "driver_specific": {} 00:10:56.104 } 00:10:56.104 ] 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- common/autotest_common.sh@905 -- # return 0 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- bdev/blockdev.sh@605 -- # sleep 2 00:10:56.104 11:21:39 blockdev_general.bdev_stat -- bdev/blockdev.sh@604 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:56.104 Running I/O for 10 seconds... 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@606 -- # stat_function_test Malloc_STAT 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@559 -- # local bdev_name=Malloc_STAT 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@560 -- # local iostats 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@561 -- # local io_count1 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@562 -- # local io_count2 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@563 -- # local iostats_per_channel 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@564 -- # local io_count_per_channel1 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@565 -- # local io_count_per_channel2 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@566 -- # local io_count_per_channel_all=0 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@568 -- # rpc_cmd bdev_get_iostat -b Malloc_STAT 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@10 -- # set +x 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@568 -- # iostats='{ 00:10:58.007 "tick_rate": 2300000000, 00:10:58.007 "ticks": 5342122825180784, 00:10:58.007 "bdevs": [ 00:10:58.007 { 00:10:58.007 "name": "Malloc_STAT", 00:10:58.007 "bytes_read": 773894656, 00:10:58.007 "num_read_ops": 188932, 00:10:58.007 "bytes_written": 0, 00:10:58.007 "num_write_ops": 0, 00:10:58.007 "bytes_unmapped": 0, 00:10:58.007 "num_unmap_ops": 0, 00:10:58.007 "bytes_copied": 0, 00:10:58.007 "num_copy_ops": 0, 00:10:58.007 "read_latency_ticks": 2228566071654, 00:10:58.007 "max_read_latency_ticks": 14853880, 00:10:58.007 "min_read_latency_ticks": 296988, 00:10:58.007 "write_latency_ticks": 0, 00:10:58.007 "max_write_latency_ticks": 0, 00:10:58.007 "min_write_latency_ticks": 0, 00:10:58.007 "unmap_latency_ticks": 0, 00:10:58.007 "max_unmap_latency_ticks": 0, 00:10:58.007 "min_unmap_latency_ticks": 0, 00:10:58.007 "copy_latency_ticks": 0, 00:10:58.007 "max_copy_latency_ticks": 0, 00:10:58.007 "min_copy_latency_ticks": 0, 00:10:58.007 "io_error": {} 00:10:58.007 } 00:10:58.007 ] 00:10:58.007 }' 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@569 -- # jq -r '.bdevs[0].num_read_ops' 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@569 -- # io_count1=188932 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@571 -- # rpc_cmd bdev_get_iostat -b Malloc_STAT -c 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@10 -- # set +x 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:58.007 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@571 -- # iostats_per_channel='{ 00:10:58.007 "tick_rate": 2300000000, 00:10:58.007 "ticks": 5342122968720204, 00:10:58.007 "name": "Malloc_STAT", 00:10:58.007 "channels": [ 00:10:58.007 { 00:10:58.007 "thread_id": 2, 00:10:58.007 "bytes_read": 394264576, 00:10:58.007 "num_read_ops": 96256, 00:10:58.007 "bytes_written": 0, 00:10:58.007 "num_write_ops": 0, 00:10:58.007 "bytes_unmapped": 0, 00:10:58.007 "num_unmap_ops": 0, 00:10:58.007 "bytes_copied": 0, 00:10:58.007 "num_copy_ops": 0, 00:10:58.007 "read_latency_ticks": 1150018197832, 00:10:58.007 "max_read_latency_ticks": 13567098, 00:10:58.007 "min_read_latency_ticks": 7582674, 00:10:58.007 "write_latency_ticks": 0, 00:10:58.007 "max_write_latency_ticks": 0, 00:10:58.007 "min_write_latency_ticks": 0, 00:10:58.007 "unmap_latency_ticks": 0, 00:10:58.007 "max_unmap_latency_ticks": 0, 00:10:58.007 "min_unmap_latency_ticks": 0, 00:10:58.008 "copy_latency_ticks": 0, 00:10:58.008 "max_copy_latency_ticks": 0, 00:10:58.008 "min_copy_latency_ticks": 0 00:10:58.008 }, 00:10:58.008 { 00:10:58.008 "thread_id": 3, 00:10:58.008 "bytes_read": 404750336, 00:10:58.008 "num_read_ops": 98816, 00:10:58.008 "bytes_written": 0, 00:10:58.008 "num_write_ops": 0, 00:10:58.008 "bytes_unmapped": 0, 00:10:58.008 "num_unmap_ops": 0, 00:10:58.008 "bytes_copied": 0, 00:10:58.008 "num_copy_ops": 0, 00:10:58.008 "read_latency_ticks": 1150927527860, 00:10:58.008 "max_read_latency_ticks": 14853880, 00:10:58.008 "min_read_latency_ticks": 7554452, 00:10:58.008 "write_latency_ticks": 0, 00:10:58.008 "max_write_latency_ticks": 0, 00:10:58.008 "min_write_latency_ticks": 0, 00:10:58.008 "unmap_latency_ticks": 0, 00:10:58.008 "max_unmap_latency_ticks": 0, 00:10:58.008 "min_unmap_latency_ticks": 0, 00:10:58.008 "copy_latency_ticks": 0, 00:10:58.008 "max_copy_latency_ticks": 0, 00:10:58.008 "min_copy_latency_ticks": 0 00:10:58.008 } 00:10:58.008 ] 00:10:58.008 }' 00:10:58.008 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@572 -- # jq -r '.channels[0].num_read_ops' 00:10:58.266 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@572 -- # io_count_per_channel1=96256 00:10:58.266 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@573 -- # io_count_per_channel_all=96256 00:10:58.266 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@574 -- # jq -r '.channels[1].num_read_ops' 00:10:58.266 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@574 -- # io_count_per_channel2=98816 00:10:58.266 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@575 -- # io_count_per_channel_all=195072 00:10:58.266 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@577 -- # rpc_cmd bdev_get_iostat -b Malloc_STAT 00:10:58.266 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:58.266 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@10 -- # set +x 00:10:58.266 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:58.266 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@577 -- # iostats='{ 00:10:58.266 "tick_rate": 2300000000, 00:10:58.266 "ticks": 5342123245463990, 00:10:58.266 "bdevs": [ 00:10:58.266 { 00:10:58.266 "name": "Malloc_STAT", 00:10:58.266 "bytes_read": 848343552, 00:10:58.266 "num_read_ops": 207108, 00:10:58.266 "bytes_written": 0, 00:10:58.266 "num_write_ops": 0, 00:10:58.266 "bytes_unmapped": 0, 00:10:58.266 "num_unmap_ops": 0, 00:10:58.266 "bytes_copied": 0, 00:10:58.266 "num_copy_ops": 0, 00:10:58.266 "read_latency_ticks": 2442675001344, 00:10:58.266 "max_read_latency_ticks": 14853880, 00:10:58.266 "min_read_latency_ticks": 296988, 00:10:58.266 "write_latency_ticks": 0, 00:10:58.266 "max_write_latency_ticks": 0, 00:10:58.266 "min_write_latency_ticks": 0, 00:10:58.266 "unmap_latency_ticks": 0, 00:10:58.266 "max_unmap_latency_ticks": 0, 00:10:58.266 "min_unmap_latency_ticks": 0, 00:10:58.266 "copy_latency_ticks": 0, 00:10:58.266 "max_copy_latency_ticks": 0, 00:10:58.266 "min_copy_latency_ticks": 0, 00:10:58.266 "io_error": {} 00:10:58.266 } 00:10:58.266 ] 00:10:58.266 }' 00:10:58.266 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@578 -- # jq -r '.bdevs[0].num_read_ops' 00:10:58.266 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@578 -- # io_count2=207108 00:10:58.266 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@583 -- # '[' 195072 -lt 188932 ']' 00:10:58.266 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@583 -- # '[' 195072 -gt 207108 ']' 00:10:58.266 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@608 -- # rpc_cmd bdev_malloc_delete Malloc_STAT 00:10:58.266 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:58.266 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@10 -- # set +x 00:10:58.266 00:10:58.266 Latency(us) 00:10:58.266 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:10:58.266 Job: Malloc_STAT (Core Mask 0x1, workload: randread, depth: 256, IO size: 4096) 00:10:58.266 Malloc_STAT : 2.15 49185.86 192.13 0.00 0.00 5191.83 1894.85 5926.73 00:10:58.267 Job: Malloc_STAT (Core Mask 0x2, workload: randread, depth: 256, IO size: 4096) 00:10:58.267 Malloc_STAT : 2.15 50493.50 197.24 0.00 0.00 5057.73 1795.12 6468.12 00:10:58.267 =================================================================================================================== 00:10:58.267 Total : 99679.35 389.37 0.00 0.00 5123.90 1795.12 6468.12 00:10:58.267 0 00:10:58.267 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:58.267 11:21:41 blockdev_general.bdev_stat -- bdev/blockdev.sh@609 -- # killprocess 859744 00:10:58.267 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@948 -- # '[' -z 859744 ']' 00:10:58.267 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@952 -- # kill -0 859744 00:10:58.267 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@953 -- # uname 00:10:58.267 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:10:58.267 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 859744 00:10:58.267 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:10:58.267 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:10:58.267 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@966 -- # echo 'killing process with pid 859744' 00:10:58.267 killing process with pid 859744 00:10:58.267 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@967 -- # kill 859744 00:10:58.267 Received shutdown signal, test time was about 2.230502 seconds 00:10:58.267 00:10:58.267 Latency(us) 00:10:58.267 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:10:58.267 =================================================================================================================== 00:10:58.267 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:10:58.267 11:21:41 blockdev_general.bdev_stat -- common/autotest_common.sh@972 -- # wait 859744 00:10:58.525 11:21:42 blockdev_general.bdev_stat -- bdev/blockdev.sh@610 -- # trap - SIGINT SIGTERM EXIT 00:10:58.525 00:10:58.525 real 0m3.609s 00:10:58.525 user 0m7.187s 00:10:58.525 sys 0m0.474s 00:10:58.525 11:21:42 blockdev_general.bdev_stat -- common/autotest_common.sh@1124 -- # xtrace_disable 00:10:58.525 11:21:42 blockdev_general.bdev_stat -- common/autotest_common.sh@10 -- # set +x 00:10:58.525 ************************************ 00:10:58.525 END TEST bdev_stat 00:10:58.525 ************************************ 00:10:58.525 11:21:42 blockdev_general -- common/autotest_common.sh@1142 -- # return 0 00:10:58.525 11:21:42 blockdev_general -- bdev/blockdev.sh@794 -- # [[ bdev == gpt ]] 00:10:58.525 11:21:42 blockdev_general -- bdev/blockdev.sh@798 -- # [[ bdev == crypto_sw ]] 00:10:58.525 11:21:42 blockdev_general -- bdev/blockdev.sh@810 -- # trap - SIGINT SIGTERM EXIT 00:10:58.526 11:21:42 blockdev_general -- bdev/blockdev.sh@811 -- # cleanup 00:10:58.526 11:21:42 blockdev_general -- bdev/blockdev.sh@23 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/aiofile 00:10:58.526 11:21:42 blockdev_general -- bdev/blockdev.sh@24 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json 00:10:58.526 11:21:42 blockdev_general -- bdev/blockdev.sh@26 -- # [[ bdev == rbd ]] 00:10:58.526 11:21:42 blockdev_general -- bdev/blockdev.sh@30 -- # [[ bdev == daos ]] 00:10:58.526 11:21:42 blockdev_general -- bdev/blockdev.sh@34 -- # [[ bdev = \g\p\t ]] 00:10:58.526 11:21:42 blockdev_general -- bdev/blockdev.sh@40 -- # [[ bdev == xnvme ]] 00:10:58.526 00:10:58.526 real 1m57.723s 00:10:58.526 user 7m11.351s 00:10:58.526 sys 0m23.795s 00:10:58.526 11:21:42 blockdev_general -- common/autotest_common.sh@1124 -- # xtrace_disable 00:10:58.526 11:21:42 blockdev_general -- common/autotest_common.sh@10 -- # set +x 00:10:58.526 ************************************ 00:10:58.526 END TEST blockdev_general 00:10:58.526 ************************************ 00:10:58.784 11:21:42 -- common/autotest_common.sh@1142 -- # return 0 00:10:58.784 11:21:42 -- spdk/autotest.sh@190 -- # run_test bdev_raid /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev_raid.sh 00:10:58.784 11:21:42 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:10:58.784 11:21:42 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:10:58.784 11:21:42 -- common/autotest_common.sh@10 -- # set +x 00:10:58.784 ************************************ 00:10:58.784 START TEST bdev_raid 00:10:58.784 ************************************ 00:10:58.784 11:21:42 bdev_raid -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev_raid.sh 00:10:58.784 * Looking for test storage... 00:10:58.785 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev 00:10:58.785 11:21:42 bdev_raid -- bdev/bdev_raid.sh@13 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbd_common.sh 00:10:58.785 11:21:42 bdev_raid -- bdev/nbd_common.sh@6 -- # set -e 00:10:58.785 11:21:42 bdev_raid -- bdev/bdev_raid.sh@15 -- # rpc_py='/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock' 00:10:58.785 11:21:42 bdev_raid -- bdev/bdev_raid.sh@851 -- # mkdir -p /raidtest 00:10:58.785 11:21:42 bdev_raid -- bdev/bdev_raid.sh@852 -- # trap 'cleanup; exit 1' EXIT 00:10:58.785 11:21:42 bdev_raid -- bdev/bdev_raid.sh@854 -- # base_blocklen=512 00:10:58.785 11:21:42 bdev_raid -- bdev/bdev_raid.sh@856 -- # uname -s 00:10:58.785 11:21:42 bdev_raid -- bdev/bdev_raid.sh@856 -- # '[' Linux = Linux ']' 00:10:58.785 11:21:42 bdev_raid -- bdev/bdev_raid.sh@856 -- # modprobe -n nbd 00:10:58.785 11:21:42 bdev_raid -- bdev/bdev_raid.sh@857 -- # has_nbd=true 00:10:58.785 11:21:42 bdev_raid -- bdev/bdev_raid.sh@858 -- # modprobe nbd 00:10:58.785 11:21:42 bdev_raid -- bdev/bdev_raid.sh@859 -- # run_test raid_function_test_raid0 raid_function_test raid0 00:10:58.785 11:21:42 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:10:58.785 11:21:42 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:10:58.785 11:21:42 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:58.785 ************************************ 00:10:58.785 START TEST raid_function_test_raid0 00:10:58.785 ************************************ 00:10:58.785 11:21:42 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1123 -- # raid_function_test raid0 00:10:58.785 11:21:42 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@80 -- # local raid_level=raid0 00:10:58.785 11:21:42 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@81 -- # local nbd=/dev/nbd0 00:10:58.785 11:21:42 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@82 -- # local raid_bdev 00:10:58.785 11:21:42 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@85 -- # raid_pid=860360 00:10:58.785 11:21:42 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@86 -- # echo 'Process raid pid: 860360' 00:10:58.785 Process raid pid: 860360 00:10:58.785 11:21:42 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:10:58.785 11:21:42 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@87 -- # waitforlisten 860360 /var/tmp/spdk-raid.sock 00:10:58.785 11:21:42 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@829 -- # '[' -z 860360 ']' 00:10:58.785 11:21:42 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:10:58.785 11:21:42 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@834 -- # local max_retries=100 00:10:58.785 11:21:42 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:10:58.785 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:10:58.785 11:21:42 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@838 -- # xtrace_disable 00:10:58.785 11:21:42 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:10:59.042 [2024-07-15 11:21:42.463218] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:10:59.042 [2024-07-15 11:21:42.463356] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:59.300 [2024-07-15 11:21:42.660924] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:59.300 [2024-07-15 11:21:42.761238] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:10:59.300 [2024-07-15 11:21:42.833100] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:59.300 [2024-07-15 11:21:42.833137] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:59.864 11:21:43 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:10:59.864 11:21:43 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@862 -- # return 0 00:10:59.864 11:21:43 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@89 -- # configure_raid_bdev raid0 00:10:59.864 11:21:43 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@66 -- # local raid_level=raid0 00:10:59.864 11:21:43 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@67 -- # rm -rf /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/rpcs.txt 00:10:59.864 11:21:43 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@69 -- # cat 00:10:59.864 11:21:43 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@74 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock 00:11:00.121 [2024-07-15 11:21:43.609153] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:11:00.121 [2024-07-15 11:21:43.610613] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:11:00.121 [2024-07-15 11:21:43.610669] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x218dbd0 00:11:00.121 [2024-07-15 11:21:43.610679] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:11:00.121 [2024-07-15 11:21:43.610864] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x218db10 00:11:00.121 [2024-07-15 11:21:43.610992] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x218dbd0 00:11:00.121 [2024-07-15 11:21:43.611002] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x218dbd0 00:11:00.121 [2024-07-15 11:21:43.611102] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:00.121 Base_1 00:11:00.121 Base_2 00:11:00.121 11:21:43 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@76 -- # rm -rf /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/rpcs.txt 00:11:00.121 11:21:43 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@90 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs online 00:11:00.121 11:21:43 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@90 -- # jq -r '.[0]["name"] | select(.)' 00:11:00.379 11:21:43 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@90 -- # raid_bdev=raid 00:11:00.379 11:21:43 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@91 -- # '[' raid = '' ']' 00:11:00.379 11:21:43 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@96 -- # nbd_start_disks /var/tmp/spdk-raid.sock raid /dev/nbd0 00:11:00.379 11:21:43 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:11:00.379 11:21:43 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:11:00.379 11:21:43 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:00.379 11:21:43 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:00.379 11:21:43 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:00.379 11:21:43 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@12 -- # local i 00:11:00.379 11:21:43 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:00.379 11:21:43 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:00.379 11:21:43 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk raid /dev/nbd0 00:11:00.638 [2024-07-15 11:21:44.102499] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x23418e0 00:11:00.638 /dev/nbd0 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@867 -- # local i 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@871 -- # break 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:00.638 1+0 records in 00:11:00.638 1+0 records out 00:11:00.638 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000242793 s, 16.9 MB/s 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@884 -- # size=4096 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@887 -- # return 0 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@97 -- # nbd_get_count /var/tmp/spdk-raid.sock 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:11:00.638 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_get_disks 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:11:00.896 { 00:11:00.896 "nbd_device": "/dev/nbd0", 00:11:00.896 "bdev_name": "raid" 00:11:00.896 } 00:11:00.896 ]' 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[ 00:11:00.896 { 00:11:00.896 "nbd_device": "/dev/nbd0", 00:11:00.896 "bdev_name": "raid" 00:11:00.896 } 00:11:00.896 ]' 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=1 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 1 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@97 -- # count=1 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@98 -- # '[' 1 -ne 1 ']' 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@102 -- # raid_unmap_data_verify /dev/nbd0 /var/tmp/spdk-raid.sock 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@18 -- # hash blkdiscard 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@19 -- # local nbd=/dev/nbd0 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@21 -- # local blksize 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@22 -- # lsblk -o LOG-SEC /dev/nbd0 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@22 -- # grep -v LOG-SEC 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@22 -- # cut -d ' ' -f 5 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@22 -- # blksize=512 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # local rw_blk_num=4096 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # local rw_len=2097152 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@25 -- # unmap_blk_offs=('0' '1028' '321') 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@25 -- # local unmap_blk_offs 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@26 -- # unmap_blk_nums=('128' '2035' '456') 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@26 -- # local unmap_blk_nums 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@27 -- # local unmap_off 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@28 -- # local unmap_len 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@31 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:11:00.896 4096+0 records in 00:11:00.896 4096+0 records out 00:11:00.896 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0305216 s, 68.7 MB/s 00:11:00.896 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@32 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:11:01.153 4096+0 records in 00:11:01.153 4096+0 records out 00:11:01.153 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.199472 s, 10.5 MB/s 00:11:01.153 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@33 -- # blockdev --flushbufs /dev/nbd0 00:11:01.153 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:11:01.153 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # (( i = 0 )) 00:11:01.153 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # (( i < 3 )) 00:11:01.153 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@39 -- # unmap_off=0 00:11:01.153 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@40 -- # unmap_len=65536 00:11:01.153 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@43 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:11:01.153 128+0 records in 00:11:01.153 128+0 records out 00:11:01.153 65536 bytes (66 kB, 64 KiB) copied, 0.000832293 s, 78.7 MB/s 00:11:01.153 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@46 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:11:01.153 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@47 -- # blockdev --flushbufs /dev/nbd0 00:11:01.153 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@50 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:11:01.153 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # (( i++ )) 00:11:01.153 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # (( i < 3 )) 00:11:01.153 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@39 -- # unmap_off=526336 00:11:01.153 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@40 -- # unmap_len=1041920 00:11:01.153 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@43 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:11:01.153 2035+0 records in 00:11:01.153 2035+0 records out 00:11:01.153 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.0108647 s, 95.9 MB/s 00:11:01.153 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@46 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:11:01.154 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@47 -- # blockdev --flushbufs /dev/nbd0 00:11:01.154 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@50 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:11:01.154 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # (( i++ )) 00:11:01.154 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # (( i < 3 )) 00:11:01.154 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@39 -- # unmap_off=164352 00:11:01.154 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@40 -- # unmap_len=233472 00:11:01.154 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@43 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:11:01.154 456+0 records in 00:11:01.154 456+0 records out 00:11:01.154 233472 bytes (233 kB, 228 KiB) copied, 0.00279866 s, 83.4 MB/s 00:11:01.154 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@46 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:11:01.411 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@47 -- # blockdev --flushbufs /dev/nbd0 00:11:01.411 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@50 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:11:01.411 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # (( i++ )) 00:11:01.411 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # (( i < 3 )) 00:11:01.411 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@54 -- # return 0 00:11:01.411 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@104 -- # nbd_stop_disks /var/tmp/spdk-raid.sock /dev/nbd0 00:11:01.411 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:11:01.411 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:01.411 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:01.411 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@51 -- # local i 00:11:01.411 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:01.411 11:21:44 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd0 00:11:01.669 [2024-07-15 11:21:45.024673] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:01.669 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:01.669 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:01.669 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:01.669 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:01.669 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:01.669 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:01.669 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@41 -- # break 00:11:01.669 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@45 -- # return 0 00:11:01.669 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@105 -- # nbd_get_count /var/tmp/spdk-raid.sock 00:11:01.669 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:11:01.669 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_get_disks 00:11:01.669 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:11:01.669 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[]' 00:11:01.669 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:11:01.669 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:11:01.927 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo '' 00:11:01.927 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:11:01.927 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # true 00:11:01.927 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=0 00:11:01.927 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 0 00:11:01.927 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@105 -- # count=0 00:11:01.927 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@106 -- # '[' 0 -ne 0 ']' 00:11:01.927 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@110 -- # killprocess 860360 00:11:01.927 11:21:45 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@948 -- # '[' -z 860360 ']' 00:11:01.927 11:21:45 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@952 -- # kill -0 860360 00:11:01.927 11:21:45 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@953 -- # uname 00:11:01.927 11:21:45 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:11:01.927 11:21:45 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 860360 00:11:01.927 11:21:45 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:11:01.927 11:21:45 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:11:01.927 11:21:45 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@966 -- # echo 'killing process with pid 860360' 00:11:01.927 killing process with pid 860360 00:11:01.927 11:21:45 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@967 -- # kill 860360 00:11:01.927 [2024-07-15 11:21:45.320979] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:01.927 [2024-07-15 11:21:45.321048] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:01.927 [2024-07-15 11:21:45.321090] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:01.927 [2024-07-15 11:21:45.321103] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x218dbd0 name raid, state offline 00:11:01.927 11:21:45 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@972 -- # wait 860360 00:11:01.927 [2024-07-15 11:21:45.337843] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:02.185 11:21:45 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@112 -- # return 0 00:11:02.185 00:11:02.185 real 0m3.195s 00:11:02.185 user 0m4.190s 00:11:02.185 sys 0m1.245s 00:11:02.185 11:21:45 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1124 -- # xtrace_disable 00:11:02.185 11:21:45 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:11:02.185 ************************************ 00:11:02.185 END TEST raid_function_test_raid0 00:11:02.185 ************************************ 00:11:02.185 11:21:45 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:11:02.185 11:21:45 bdev_raid -- bdev/bdev_raid.sh@860 -- # run_test raid_function_test_concat raid_function_test concat 00:11:02.185 11:21:45 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:11:02.185 11:21:45 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:11:02.185 11:21:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:02.185 ************************************ 00:11:02.185 START TEST raid_function_test_concat 00:11:02.185 ************************************ 00:11:02.185 11:21:45 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1123 -- # raid_function_test concat 00:11:02.185 11:21:45 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@80 -- # local raid_level=concat 00:11:02.185 11:21:45 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@81 -- # local nbd=/dev/nbd0 00:11:02.185 11:21:45 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@82 -- # local raid_bdev 00:11:02.185 11:21:45 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@85 -- # raid_pid=860803 00:11:02.185 11:21:45 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@86 -- # echo 'Process raid pid: 860803' 00:11:02.185 Process raid pid: 860803 00:11:02.185 11:21:45 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:11:02.185 11:21:45 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@87 -- # waitforlisten 860803 /var/tmp/spdk-raid.sock 00:11:02.185 11:21:45 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@829 -- # '[' -z 860803 ']' 00:11:02.185 11:21:45 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:11:02.185 11:21:45 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@834 -- # local max_retries=100 00:11:02.185 11:21:45 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:11:02.185 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:11:02.185 11:21:45 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@838 -- # xtrace_disable 00:11:02.185 11:21:45 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:11:02.185 [2024-07-15 11:21:45.702225] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:11:02.185 [2024-07-15 11:21:45.702298] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:02.443 [2024-07-15 11:21:45.831301] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:02.443 [2024-07-15 11:21:45.934811] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:11:02.443 [2024-07-15 11:21:45.998327] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:02.443 [2024-07-15 11:21:45.998354] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:03.008 11:21:46 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:11:03.008 11:21:46 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@862 -- # return 0 00:11:03.008 11:21:46 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@89 -- # configure_raid_bdev concat 00:11:03.008 11:21:46 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@66 -- # local raid_level=concat 00:11:03.008 11:21:46 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@67 -- # rm -rf /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/rpcs.txt 00:11:03.008 11:21:46 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@69 -- # cat 00:11:03.009 11:21:46 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@74 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock 00:11:03.267 [2024-07-15 11:21:46.687107] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:11:03.267 [2024-07-15 11:21:46.688578] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:11:03.267 [2024-07-15 11:21:46.688636] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1ea0bd0 00:11:03.267 [2024-07-15 11:21:46.688647] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:11:03.267 [2024-07-15 11:21:46.688837] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1ea0b10 00:11:03.267 [2024-07-15 11:21:46.688969] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1ea0bd0 00:11:03.267 [2024-07-15 11:21:46.688980] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x1ea0bd0 00:11:03.267 [2024-07-15 11:21:46.689081] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:03.267 Base_1 00:11:03.267 Base_2 00:11:03.267 11:21:46 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@76 -- # rm -rf /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/rpcs.txt 00:11:03.267 11:21:46 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@90 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs online 00:11:03.267 11:21:46 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@90 -- # jq -r '.[0]["name"] | select(.)' 00:11:03.526 11:21:46 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@90 -- # raid_bdev=raid 00:11:03.526 11:21:46 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@91 -- # '[' raid = '' ']' 00:11:03.526 11:21:46 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@96 -- # nbd_start_disks /var/tmp/spdk-raid.sock raid /dev/nbd0 00:11:03.526 11:21:46 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:11:03.526 11:21:46 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:11:03.526 11:21:46 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:03.526 11:21:46 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:03.526 11:21:46 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:03.526 11:21:46 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@12 -- # local i 00:11:03.526 11:21:46 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:03.526 11:21:46 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:03.526 11:21:46 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk raid /dev/nbd0 00:11:03.526 [2024-07-15 11:21:47.120292] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x20548e0 00:11:03.788 /dev/nbd0 00:11:03.788 11:21:47 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:03.789 11:21:47 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:03.789 11:21:47 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:11:03.789 11:21:47 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@867 -- # local i 00:11:03.789 11:21:47 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:11:03.789 11:21:47 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:11:03.789 11:21:47 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:11:03.789 11:21:47 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@871 -- # break 00:11:03.789 11:21:47 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:11:03.789 11:21:47 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:11:03.789 11:21:47 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:03.789 1+0 records in 00:11:03.789 1+0 records out 00:11:03.789 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000242084 s, 16.9 MB/s 00:11:03.789 11:21:47 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:11:03.789 11:21:47 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@884 -- # size=4096 00:11:03.789 11:21:47 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:11:03.789 11:21:47 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:11:03.789 11:21:47 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@887 -- # return 0 00:11:03.789 11:21:47 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:03.789 11:21:47 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:03.789 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@97 -- # nbd_get_count /var/tmp/spdk-raid.sock 00:11:03.789 11:21:47 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:11:03.789 11:21:47 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_get_disks 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:11:04.047 { 00:11:04.047 "nbd_device": "/dev/nbd0", 00:11:04.047 "bdev_name": "raid" 00:11:04.047 } 00:11:04.047 ]' 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[ 00:11:04.047 { 00:11:04.047 "nbd_device": "/dev/nbd0", 00:11:04.047 "bdev_name": "raid" 00:11:04.047 } 00:11:04.047 ]' 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=1 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 1 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@97 -- # count=1 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@98 -- # '[' 1 -ne 1 ']' 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@102 -- # raid_unmap_data_verify /dev/nbd0 /var/tmp/spdk-raid.sock 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@18 -- # hash blkdiscard 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@19 -- # local nbd=/dev/nbd0 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@21 -- # local blksize 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@22 -- # lsblk -o LOG-SEC /dev/nbd0 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@22 -- # cut -d ' ' -f 5 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@22 -- # grep -v LOG-SEC 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@22 -- # blksize=512 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # local rw_blk_num=4096 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # local rw_len=2097152 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@25 -- # unmap_blk_offs=('0' '1028' '321') 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@25 -- # local unmap_blk_offs 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@26 -- # unmap_blk_nums=('128' '2035' '456') 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@26 -- # local unmap_blk_nums 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@27 -- # local unmap_off 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@28 -- # local unmap_len 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@31 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:11:04.047 4096+0 records in 00:11:04.047 4096+0 records out 00:11:04.047 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0281485 s, 74.5 MB/s 00:11:04.047 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@32 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:11:04.306 4096+0 records in 00:11:04.306 4096+0 records out 00:11:04.306 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.206142 s, 10.2 MB/s 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@33 -- # blockdev --flushbufs /dev/nbd0 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # (( i = 0 )) 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # (( i < 3 )) 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@39 -- # unmap_off=0 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@40 -- # unmap_len=65536 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@43 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:11:04.306 128+0 records in 00:11:04.306 128+0 records out 00:11:04.306 65536 bytes (66 kB, 64 KiB) copied, 0.00083047 s, 78.9 MB/s 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@46 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@47 -- # blockdev --flushbufs /dev/nbd0 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@50 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # (( i++ )) 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # (( i < 3 )) 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@39 -- # unmap_off=526336 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@40 -- # unmap_len=1041920 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@43 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:11:04.306 2035+0 records in 00:11:04.306 2035+0 records out 00:11:04.306 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.0106525 s, 97.8 MB/s 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@46 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@47 -- # blockdev --flushbufs /dev/nbd0 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@50 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # (( i++ )) 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # (( i < 3 )) 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@39 -- # unmap_off=164352 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@40 -- # unmap_len=233472 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@43 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:11:04.306 456+0 records in 00:11:04.306 456+0 records out 00:11:04.306 233472 bytes (233 kB, 228 KiB) copied, 0.00273599 s, 85.3 MB/s 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@46 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@47 -- # blockdev --flushbufs /dev/nbd0 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@50 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # (( i++ )) 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # (( i < 3 )) 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@54 -- # return 0 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@104 -- # nbd_stop_disks /var/tmp/spdk-raid.sock /dev/nbd0 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@51 -- # local i 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:04.306 11:21:47 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd0 00:11:04.564 11:21:48 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:04.564 [2024-07-15 11:21:48.104630] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:04.564 11:21:48 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:04.564 11:21:48 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:04.564 11:21:48 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:04.564 11:21:48 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:04.564 11:21:48 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:04.564 11:21:48 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@41 -- # break 00:11:04.564 11:21:48 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@45 -- # return 0 00:11:04.564 11:21:48 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@105 -- # nbd_get_count /var/tmp/spdk-raid.sock 00:11:04.564 11:21:48 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:11:04.564 11:21:48 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_get_disks 00:11:04.822 11:21:48 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:11:04.822 11:21:48 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:11:04.822 11:21:48 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:11:04.822 11:21:48 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:11:04.822 11:21:48 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:11:04.822 11:21:48 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo '' 00:11:04.822 11:21:48 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # true 00:11:04.822 11:21:48 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=0 00:11:04.822 11:21:48 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 0 00:11:04.822 11:21:48 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@105 -- # count=0 00:11:04.822 11:21:48 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@106 -- # '[' 0 -ne 0 ']' 00:11:04.822 11:21:48 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@110 -- # killprocess 860803 00:11:04.822 11:21:48 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@948 -- # '[' -z 860803 ']' 00:11:04.822 11:21:48 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@952 -- # kill -0 860803 00:11:04.822 11:21:48 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@953 -- # uname 00:11:05.080 11:21:48 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:11:05.080 11:21:48 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 860803 00:11:05.080 11:21:48 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:11:05.080 11:21:48 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:11:05.080 11:21:48 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@966 -- # echo 'killing process with pid 860803' 00:11:05.080 killing process with pid 860803 00:11:05.080 11:21:48 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@967 -- # kill 860803 00:11:05.080 [2024-07-15 11:21:48.460810] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:05.080 [2024-07-15 11:21:48.460882] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:05.080 [2024-07-15 11:21:48.460934] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:05.080 [2024-07-15 11:21:48.460950] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1ea0bd0 name raid, state offline 00:11:05.080 11:21:48 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@972 -- # wait 860803 00:11:05.080 [2024-07-15 11:21:48.478033] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:05.339 11:21:48 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@112 -- # return 0 00:11:05.339 00:11:05.339 real 0m3.060s 00:11:05.339 user 0m3.940s 00:11:05.339 sys 0m1.198s 00:11:05.339 11:21:48 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1124 -- # xtrace_disable 00:11:05.339 11:21:48 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:11:05.339 ************************************ 00:11:05.339 END TEST raid_function_test_concat 00:11:05.339 ************************************ 00:11:05.339 11:21:48 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:11:05.339 11:21:48 bdev_raid -- bdev/bdev_raid.sh@863 -- # run_test raid0_resize_test raid0_resize_test 00:11:05.339 11:21:48 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:11:05.339 11:21:48 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:11:05.339 11:21:48 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:05.339 ************************************ 00:11:05.339 START TEST raid0_resize_test 00:11:05.339 ************************************ 00:11:05.339 11:21:48 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1123 -- # raid0_resize_test 00:11:05.339 11:21:48 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@347 -- # local blksize=512 00:11:05.339 11:21:48 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@348 -- # local bdev_size_mb=32 00:11:05.339 11:21:48 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@349 -- # local new_bdev_size_mb=64 00:11:05.339 11:21:48 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@350 -- # local blkcnt 00:11:05.339 11:21:48 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@351 -- # local raid_size_mb 00:11:05.339 11:21:48 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@352 -- # local new_raid_size_mb 00:11:05.339 11:21:48 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@355 -- # raid_pid=861354 00:11:05.339 11:21:48 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@356 -- # echo 'Process raid pid: 861354' 00:11:05.339 Process raid pid: 861354 00:11:05.339 11:21:48 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@357 -- # waitforlisten 861354 /var/tmp/spdk-raid.sock 00:11:05.339 11:21:48 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@354 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:11:05.339 11:21:48 bdev_raid.raid0_resize_test -- common/autotest_common.sh@829 -- # '[' -z 861354 ']' 00:11:05.339 11:21:48 bdev_raid.raid0_resize_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:11:05.339 11:21:48 bdev_raid.raid0_resize_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:11:05.339 11:21:48 bdev_raid.raid0_resize_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:11:05.339 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:11:05.339 11:21:48 bdev_raid.raid0_resize_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:11:05.339 11:21:48 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.339 [2024-07-15 11:21:48.843644] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:11:05.339 [2024-07-15 11:21:48.843701] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:05.597 [2024-07-15 11:21:48.955770] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:05.597 [2024-07-15 11:21:49.058869] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:11:05.597 [2024-07-15 11:21:49.125688] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:05.597 [2024-07-15 11:21:49.125728] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:06.162 11:21:49 bdev_raid.raid0_resize_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:11:06.162 11:21:49 bdev_raid.raid0_resize_test -- common/autotest_common.sh@862 -- # return 0 00:11:06.162 11:21:49 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_null_create Base_1 32 512 00:11:06.420 Base_1 00:11:06.420 11:21:49 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@360 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_null_create Base_2 32 512 00:11:06.420 Base_2 00:11:06.421 11:21:49 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@362 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r 0 -b 'Base_1 Base_2' -n Raid 00:11:06.679 [2024-07-15 11:21:50.209830] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:11:06.679 [2024-07-15 11:21:50.211219] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:11:06.679 [2024-07-15 11:21:50.211269] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xd95780 00:11:06.679 [2024-07-15 11:21:50.211278] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:11:06.679 [2024-07-15 11:21:50.211496] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x8e1020 00:11:06.679 [2024-07-15 11:21:50.211592] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xd95780 00:11:06.679 [2024-07-15 11:21:50.211602] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0xd95780 00:11:06.679 [2024-07-15 11:21:50.211718] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:06.679 11:21:50 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@365 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_null_resize Base_1 64 00:11:06.937 [2024-07-15 11:21:50.390272] bdev_raid.c:2262:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:11:06.937 [2024-07-15 11:21:50.390288] bdev_raid.c:2275:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:11:06.937 true 00:11:06.937 11:21:50 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@368 -- # jq '.[].num_blocks' 00:11:06.937 11:21:50 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@368 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Raid 00:11:07.195 [2024-07-15 11:21:50.566886] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:07.195 11:21:50 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@368 -- # blkcnt=131072 00:11:07.195 11:21:50 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@369 -- # raid_size_mb=64 00:11:07.195 11:21:50 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@370 -- # '[' 64 '!=' 64 ']' 00:11:07.195 11:21:50 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@376 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_null_resize Base_2 64 00:11:07.761 [2024-07-15 11:21:51.072052] bdev_raid.c:2262:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:11:07.761 [2024-07-15 11:21:51.072077] bdev_raid.c:2275:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:11:07.761 [2024-07-15 11:21:51.072104] bdev_raid.c:2289:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 131072 to 262144 00:11:07.761 true 00:11:07.761 11:21:51 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@379 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Raid 00:11:07.761 11:21:51 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@379 -- # jq '.[].num_blocks' 00:11:07.761 [2024-07-15 11:21:51.328894] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:07.761 11:21:51 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@379 -- # blkcnt=262144 00:11:07.761 11:21:51 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@380 -- # raid_size_mb=128 00:11:07.761 11:21:51 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@381 -- # '[' 128 '!=' 128 ']' 00:11:07.761 11:21:51 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@386 -- # killprocess 861354 00:11:07.761 11:21:51 bdev_raid.raid0_resize_test -- common/autotest_common.sh@948 -- # '[' -z 861354 ']' 00:11:07.761 11:21:51 bdev_raid.raid0_resize_test -- common/autotest_common.sh@952 -- # kill -0 861354 00:11:07.761 11:21:51 bdev_raid.raid0_resize_test -- common/autotest_common.sh@953 -- # uname 00:11:07.761 11:21:51 bdev_raid.raid0_resize_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:11:07.761 11:21:51 bdev_raid.raid0_resize_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 861354 00:11:08.020 11:21:51 bdev_raid.raid0_resize_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:11:08.020 11:21:51 bdev_raid.raid0_resize_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:11:08.020 11:21:51 bdev_raid.raid0_resize_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 861354' 00:11:08.020 killing process with pid 861354 00:11:08.020 11:21:51 bdev_raid.raid0_resize_test -- common/autotest_common.sh@967 -- # kill 861354 00:11:08.020 [2024-07-15 11:21:51.397271] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:08.020 [2024-07-15 11:21:51.397328] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:08.020 [2024-07-15 11:21:51.397370] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:08.020 [2024-07-15 11:21:51.397381] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xd95780 name Raid, state offline 00:11:08.020 11:21:51 bdev_raid.raid0_resize_test -- common/autotest_common.sh@972 -- # wait 861354 00:11:08.020 [2024-07-15 11:21:51.398762] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:08.020 11:21:51 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@388 -- # return 0 00:11:08.020 00:11:08.020 real 0m2.823s 00:11:08.020 user 0m4.275s 00:11:08.020 sys 0m0.629s 00:11:08.020 11:21:51 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:11:08.020 11:21:51 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.020 ************************************ 00:11:08.020 END TEST raid0_resize_test 00:11:08.020 ************************************ 00:11:08.279 11:21:51 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:11:08.279 11:21:51 bdev_raid -- bdev/bdev_raid.sh@865 -- # for n in {2..4} 00:11:08.279 11:21:51 bdev_raid -- bdev/bdev_raid.sh@866 -- # for level in raid0 concat raid1 00:11:08.279 11:21:51 bdev_raid -- bdev/bdev_raid.sh@867 -- # run_test raid_state_function_test raid_state_function_test raid0 2 false 00:11:08.279 11:21:51 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:11:08.279 11:21:51 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:11:08.279 11:21:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:08.279 ************************************ 00:11:08.279 START TEST raid_state_function_test 00:11:08.279 ************************************ 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1123 -- # raid_state_function_test raid0 2 false 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@220 -- # local raid_level=raid0 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=2 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # local superblock=false 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # local strip_size 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # '[' raid0 '!=' raid1 ']' 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # strip_size=64 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@232 -- # strip_size_create_arg='-z 64' 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # '[' false = true ']' 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@240 -- # superblock_create_arg= 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # raid_pid=861796 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 861796' 00:11:08.279 Process raid pid: 861796 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@246 -- # waitforlisten 861796 /var/tmp/spdk-raid.sock 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@829 -- # '[' -z 861796 ']' 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:11:08.279 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:11:08.279 11:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.279 [2024-07-15 11:21:51.766524] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:11:08.279 [2024-07-15 11:21:51.766593] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:08.537 [2024-07-15 11:21:51.895414] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:08.537 [2024-07-15 11:21:51.994630] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:11:08.537 [2024-07-15 11:21:52.059818] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:08.537 [2024-07-15 11:21:52.059859] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:09.472 11:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:11:09.472 11:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@862 -- # return 0 00:11:09.472 11:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:11:09.472 [2024-07-15 11:21:52.919016] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:09.472 [2024-07-15 11:21:52.919062] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:09.472 [2024-07-15 11:21:52.919073] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:09.472 [2024-07-15 11:21:52.919085] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:09.472 11:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:11:09.472 11:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:11:09.472 11:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:11:09.472 11:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:11:09.472 11:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:09.472 11:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:11:09.472 11:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:09.472 11:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:09.472 11:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:09.472 11:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:09.472 11:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:09.472 11:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:09.730 11:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:09.730 "name": "Existed_Raid", 00:11:09.730 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:09.730 "strip_size_kb": 64, 00:11:09.730 "state": "configuring", 00:11:09.730 "raid_level": "raid0", 00:11:09.730 "superblock": false, 00:11:09.730 "num_base_bdevs": 2, 00:11:09.730 "num_base_bdevs_discovered": 0, 00:11:09.730 "num_base_bdevs_operational": 2, 00:11:09.730 "base_bdevs_list": [ 00:11:09.730 { 00:11:09.730 "name": "BaseBdev1", 00:11:09.730 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:09.730 "is_configured": false, 00:11:09.730 "data_offset": 0, 00:11:09.730 "data_size": 0 00:11:09.730 }, 00:11:09.730 { 00:11:09.730 "name": "BaseBdev2", 00:11:09.730 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:09.730 "is_configured": false, 00:11:09.730 "data_offset": 0, 00:11:09.730 "data_size": 0 00:11:09.730 } 00:11:09.730 ] 00:11:09.730 }' 00:11:09.730 11:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:09.730 11:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.296 11:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:11:10.553 [2024-07-15 11:21:54.013795] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:10.553 [2024-07-15 11:21:54.013827] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1adca80 name Existed_Raid, state configuring 00:11:10.553 11:21:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:11:10.810 [2024-07-15 11:21:54.198298] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:10.810 [2024-07-15 11:21:54.198340] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:10.810 [2024-07-15 11:21:54.198350] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:10.810 [2024-07-15 11:21:54.198361] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:10.810 11:21:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:11:11.081 [2024-07-15 11:21:54.476909] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:11.081 BaseBdev1 00:11:11.081 11:21:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:11:11.081 11:21:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:11:11.081 11:21:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:11:11.081 11:21:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:11:11.081 11:21:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:11:11.081 11:21:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:11:11.081 11:21:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:11:11.367 11:21:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:11.626 [ 00:11:11.626 { 00:11:11.626 "name": "BaseBdev1", 00:11:11.626 "aliases": [ 00:11:11.626 "e90b9948-df84-4b51-982c-b819d1c52129" 00:11:11.626 ], 00:11:11.626 "product_name": "Malloc disk", 00:11:11.626 "block_size": 512, 00:11:11.626 "num_blocks": 65536, 00:11:11.626 "uuid": "e90b9948-df84-4b51-982c-b819d1c52129", 00:11:11.626 "assigned_rate_limits": { 00:11:11.626 "rw_ios_per_sec": 0, 00:11:11.626 "rw_mbytes_per_sec": 0, 00:11:11.626 "r_mbytes_per_sec": 0, 00:11:11.626 "w_mbytes_per_sec": 0 00:11:11.626 }, 00:11:11.626 "claimed": true, 00:11:11.626 "claim_type": "exclusive_write", 00:11:11.626 "zoned": false, 00:11:11.626 "supported_io_types": { 00:11:11.626 "read": true, 00:11:11.626 "write": true, 00:11:11.626 "unmap": true, 00:11:11.626 "flush": true, 00:11:11.626 "reset": true, 00:11:11.626 "nvme_admin": false, 00:11:11.626 "nvme_io": false, 00:11:11.626 "nvme_io_md": false, 00:11:11.626 "write_zeroes": true, 00:11:11.626 "zcopy": true, 00:11:11.626 "get_zone_info": false, 00:11:11.626 "zone_management": false, 00:11:11.626 "zone_append": false, 00:11:11.626 "compare": false, 00:11:11.626 "compare_and_write": false, 00:11:11.626 "abort": true, 00:11:11.626 "seek_hole": false, 00:11:11.626 "seek_data": false, 00:11:11.626 "copy": true, 00:11:11.626 "nvme_iov_md": false 00:11:11.626 }, 00:11:11.626 "memory_domains": [ 00:11:11.626 { 00:11:11.626 "dma_device_id": "system", 00:11:11.626 "dma_device_type": 1 00:11:11.626 }, 00:11:11.626 { 00:11:11.626 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:11.626 "dma_device_type": 2 00:11:11.626 } 00:11:11.626 ], 00:11:11.626 "driver_specific": {} 00:11:11.626 } 00:11:11.626 ] 00:11:11.626 11:21:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:11:11.626 11:21:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:11:11.626 11:21:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:11:11.626 11:21:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:11:11.626 11:21:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:11:11.626 11:21:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:11.626 11:21:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:11:11.626 11:21:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:11.626 11:21:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:11.626 11:21:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:11.626 11:21:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:11.626 11:21:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:11.626 11:21:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:11.884 11:21:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:11.884 "name": "Existed_Raid", 00:11:11.884 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:11.884 "strip_size_kb": 64, 00:11:11.884 "state": "configuring", 00:11:11.884 "raid_level": "raid0", 00:11:11.884 "superblock": false, 00:11:11.884 "num_base_bdevs": 2, 00:11:11.884 "num_base_bdevs_discovered": 1, 00:11:11.884 "num_base_bdevs_operational": 2, 00:11:11.884 "base_bdevs_list": [ 00:11:11.884 { 00:11:11.884 "name": "BaseBdev1", 00:11:11.884 "uuid": "e90b9948-df84-4b51-982c-b819d1c52129", 00:11:11.884 "is_configured": true, 00:11:11.884 "data_offset": 0, 00:11:11.884 "data_size": 65536 00:11:11.884 }, 00:11:11.884 { 00:11:11.884 "name": "BaseBdev2", 00:11:11.884 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:11.884 "is_configured": false, 00:11:11.884 "data_offset": 0, 00:11:11.884 "data_size": 0 00:11:11.884 } 00:11:11.884 ] 00:11:11.884 }' 00:11:11.884 11:21:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:11.884 11:21:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.449 11:21:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:11:12.449 [2024-07-15 11:21:55.992920] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:12.449 [2024-07-15 11:21:55.992970] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1adc350 name Existed_Raid, state configuring 00:11:12.449 11:21:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:11:12.707 [2024-07-15 11:21:56.165414] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:12.707 [2024-07-15 11:21:56.166906] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:12.707 [2024-07-15 11:21:56.166948] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:12.707 11:21:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:11:12.707 11:21:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:11:12.707 11:21:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:11:12.707 11:21:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:11:12.707 11:21:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:11:12.707 11:21:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:11:12.707 11:21:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:12.707 11:21:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:11:12.707 11:21:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:12.707 11:21:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:12.707 11:21:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:12.707 11:21:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:12.707 11:21:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:12.707 11:21:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:12.964 11:21:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:12.964 "name": "Existed_Raid", 00:11:12.964 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:12.964 "strip_size_kb": 64, 00:11:12.964 "state": "configuring", 00:11:12.964 "raid_level": "raid0", 00:11:12.964 "superblock": false, 00:11:12.964 "num_base_bdevs": 2, 00:11:12.964 "num_base_bdevs_discovered": 1, 00:11:12.964 "num_base_bdevs_operational": 2, 00:11:12.964 "base_bdevs_list": [ 00:11:12.964 { 00:11:12.964 "name": "BaseBdev1", 00:11:12.964 "uuid": "e90b9948-df84-4b51-982c-b819d1c52129", 00:11:12.964 "is_configured": true, 00:11:12.964 "data_offset": 0, 00:11:12.964 "data_size": 65536 00:11:12.964 }, 00:11:12.964 { 00:11:12.964 "name": "BaseBdev2", 00:11:12.964 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:12.964 "is_configured": false, 00:11:12.964 "data_offset": 0, 00:11:12.964 "data_size": 0 00:11:12.964 } 00:11:12.964 ] 00:11:12.964 }' 00:11:12.964 11:21:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:12.964 11:21:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.530 11:21:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:11:13.530 [2024-07-15 11:21:57.051147] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:13.530 [2024-07-15 11:21:57.051186] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1add000 00:11:13.530 [2024-07-15 11:21:57.051195] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:11:13.530 [2024-07-15 11:21:57.051388] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x19f70c0 00:11:13.530 [2024-07-15 11:21:57.051509] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1add000 00:11:13.530 [2024-07-15 11:21:57.051519] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x1add000 00:11:13.530 [2024-07-15 11:21:57.051680] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:13.530 BaseBdev2 00:11:13.530 11:21:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:11:13.530 11:21:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:11:13.530 11:21:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:11:13.530 11:21:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:11:13.530 11:21:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:11:13.530 11:21:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:11:13.530 11:21:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:11:13.787 11:21:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:14.045 [ 00:11:14.045 { 00:11:14.045 "name": "BaseBdev2", 00:11:14.045 "aliases": [ 00:11:14.045 "f110ab7e-706a-48be-9eb2-5c2d86924d2e" 00:11:14.045 ], 00:11:14.045 "product_name": "Malloc disk", 00:11:14.045 "block_size": 512, 00:11:14.045 "num_blocks": 65536, 00:11:14.045 "uuid": "f110ab7e-706a-48be-9eb2-5c2d86924d2e", 00:11:14.045 "assigned_rate_limits": { 00:11:14.045 "rw_ios_per_sec": 0, 00:11:14.045 "rw_mbytes_per_sec": 0, 00:11:14.045 "r_mbytes_per_sec": 0, 00:11:14.045 "w_mbytes_per_sec": 0 00:11:14.045 }, 00:11:14.045 "claimed": true, 00:11:14.045 "claim_type": "exclusive_write", 00:11:14.045 "zoned": false, 00:11:14.045 "supported_io_types": { 00:11:14.045 "read": true, 00:11:14.045 "write": true, 00:11:14.045 "unmap": true, 00:11:14.045 "flush": true, 00:11:14.045 "reset": true, 00:11:14.045 "nvme_admin": false, 00:11:14.045 "nvme_io": false, 00:11:14.045 "nvme_io_md": false, 00:11:14.045 "write_zeroes": true, 00:11:14.045 "zcopy": true, 00:11:14.045 "get_zone_info": false, 00:11:14.045 "zone_management": false, 00:11:14.045 "zone_append": false, 00:11:14.045 "compare": false, 00:11:14.045 "compare_and_write": false, 00:11:14.045 "abort": true, 00:11:14.046 "seek_hole": false, 00:11:14.046 "seek_data": false, 00:11:14.046 "copy": true, 00:11:14.046 "nvme_iov_md": false 00:11:14.046 }, 00:11:14.046 "memory_domains": [ 00:11:14.046 { 00:11:14.046 "dma_device_id": "system", 00:11:14.046 "dma_device_type": 1 00:11:14.046 }, 00:11:14.046 { 00:11:14.046 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:14.046 "dma_device_type": 2 00:11:14.046 } 00:11:14.046 ], 00:11:14.046 "driver_specific": {} 00:11:14.046 } 00:11:14.046 ] 00:11:14.046 11:21:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:11:14.046 11:21:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:11:14.046 11:21:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:11:14.046 11:21:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:11:14.046 11:21:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:11:14.046 11:21:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:11:14.046 11:21:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:11:14.046 11:21:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:14.046 11:21:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:11:14.046 11:21:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:14.046 11:21:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:14.046 11:21:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:14.046 11:21:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:14.046 11:21:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:14.046 11:21:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:14.046 11:21:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:14.046 "name": "Existed_Raid", 00:11:14.046 "uuid": "4aced7ec-8c7f-40be-b3a5-801f6c94bb34", 00:11:14.046 "strip_size_kb": 64, 00:11:14.046 "state": "online", 00:11:14.046 "raid_level": "raid0", 00:11:14.046 "superblock": false, 00:11:14.046 "num_base_bdevs": 2, 00:11:14.046 "num_base_bdevs_discovered": 2, 00:11:14.046 "num_base_bdevs_operational": 2, 00:11:14.046 "base_bdevs_list": [ 00:11:14.046 { 00:11:14.046 "name": "BaseBdev1", 00:11:14.046 "uuid": "e90b9948-df84-4b51-982c-b819d1c52129", 00:11:14.046 "is_configured": true, 00:11:14.046 "data_offset": 0, 00:11:14.046 "data_size": 65536 00:11:14.046 }, 00:11:14.046 { 00:11:14.046 "name": "BaseBdev2", 00:11:14.046 "uuid": "f110ab7e-706a-48be-9eb2-5c2d86924d2e", 00:11:14.046 "is_configured": true, 00:11:14.046 "data_offset": 0, 00:11:14.046 "data_size": 65536 00:11:14.046 } 00:11:14.046 ] 00:11:14.046 }' 00:11:14.046 11:21:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:14.046 11:21:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:14.610 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:11:14.610 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:11:14.610 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:11:14.610 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:11:14.610 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:11:14.610 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # local name 00:11:14.610 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:11:14.610 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:11:14.867 [2024-07-15 11:21:58.403009] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:14.867 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:11:14.867 "name": "Existed_Raid", 00:11:14.867 "aliases": [ 00:11:14.867 "4aced7ec-8c7f-40be-b3a5-801f6c94bb34" 00:11:14.867 ], 00:11:14.867 "product_name": "Raid Volume", 00:11:14.867 "block_size": 512, 00:11:14.867 "num_blocks": 131072, 00:11:14.867 "uuid": "4aced7ec-8c7f-40be-b3a5-801f6c94bb34", 00:11:14.867 "assigned_rate_limits": { 00:11:14.867 "rw_ios_per_sec": 0, 00:11:14.867 "rw_mbytes_per_sec": 0, 00:11:14.867 "r_mbytes_per_sec": 0, 00:11:14.867 "w_mbytes_per_sec": 0 00:11:14.867 }, 00:11:14.867 "claimed": false, 00:11:14.867 "zoned": false, 00:11:14.867 "supported_io_types": { 00:11:14.867 "read": true, 00:11:14.867 "write": true, 00:11:14.867 "unmap": true, 00:11:14.867 "flush": true, 00:11:14.867 "reset": true, 00:11:14.867 "nvme_admin": false, 00:11:14.867 "nvme_io": false, 00:11:14.867 "nvme_io_md": false, 00:11:14.867 "write_zeroes": true, 00:11:14.867 "zcopy": false, 00:11:14.867 "get_zone_info": false, 00:11:14.867 "zone_management": false, 00:11:14.867 "zone_append": false, 00:11:14.867 "compare": false, 00:11:14.867 "compare_and_write": false, 00:11:14.867 "abort": false, 00:11:14.867 "seek_hole": false, 00:11:14.867 "seek_data": false, 00:11:14.867 "copy": false, 00:11:14.867 "nvme_iov_md": false 00:11:14.867 }, 00:11:14.867 "memory_domains": [ 00:11:14.867 { 00:11:14.867 "dma_device_id": "system", 00:11:14.867 "dma_device_type": 1 00:11:14.867 }, 00:11:14.867 { 00:11:14.867 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:14.867 "dma_device_type": 2 00:11:14.867 }, 00:11:14.867 { 00:11:14.867 "dma_device_id": "system", 00:11:14.867 "dma_device_type": 1 00:11:14.867 }, 00:11:14.867 { 00:11:14.867 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:14.867 "dma_device_type": 2 00:11:14.867 } 00:11:14.867 ], 00:11:14.867 "driver_specific": { 00:11:14.867 "raid": { 00:11:14.867 "uuid": "4aced7ec-8c7f-40be-b3a5-801f6c94bb34", 00:11:14.867 "strip_size_kb": 64, 00:11:14.867 "state": "online", 00:11:14.867 "raid_level": "raid0", 00:11:14.867 "superblock": false, 00:11:14.867 "num_base_bdevs": 2, 00:11:14.867 "num_base_bdevs_discovered": 2, 00:11:14.867 "num_base_bdevs_operational": 2, 00:11:14.867 "base_bdevs_list": [ 00:11:14.867 { 00:11:14.867 "name": "BaseBdev1", 00:11:14.867 "uuid": "e90b9948-df84-4b51-982c-b819d1c52129", 00:11:14.867 "is_configured": true, 00:11:14.867 "data_offset": 0, 00:11:14.867 "data_size": 65536 00:11:14.867 }, 00:11:14.867 { 00:11:14.867 "name": "BaseBdev2", 00:11:14.867 "uuid": "f110ab7e-706a-48be-9eb2-5c2d86924d2e", 00:11:14.867 "is_configured": true, 00:11:14.867 "data_offset": 0, 00:11:14.867 "data_size": 65536 00:11:14.867 } 00:11:14.867 ] 00:11:14.867 } 00:11:14.867 } 00:11:14.867 }' 00:11:14.868 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:15.125 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:11:15.125 BaseBdev2' 00:11:15.125 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:11:15.125 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:11:15.125 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:11:15.383 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:11:15.383 "name": "BaseBdev1", 00:11:15.383 "aliases": [ 00:11:15.383 "e90b9948-df84-4b51-982c-b819d1c52129" 00:11:15.383 ], 00:11:15.383 "product_name": "Malloc disk", 00:11:15.383 "block_size": 512, 00:11:15.383 "num_blocks": 65536, 00:11:15.383 "uuid": "e90b9948-df84-4b51-982c-b819d1c52129", 00:11:15.383 "assigned_rate_limits": { 00:11:15.383 "rw_ios_per_sec": 0, 00:11:15.383 "rw_mbytes_per_sec": 0, 00:11:15.383 "r_mbytes_per_sec": 0, 00:11:15.383 "w_mbytes_per_sec": 0 00:11:15.383 }, 00:11:15.383 "claimed": true, 00:11:15.383 "claim_type": "exclusive_write", 00:11:15.383 "zoned": false, 00:11:15.383 "supported_io_types": { 00:11:15.383 "read": true, 00:11:15.383 "write": true, 00:11:15.383 "unmap": true, 00:11:15.383 "flush": true, 00:11:15.383 "reset": true, 00:11:15.383 "nvme_admin": false, 00:11:15.383 "nvme_io": false, 00:11:15.383 "nvme_io_md": false, 00:11:15.383 "write_zeroes": true, 00:11:15.383 "zcopy": true, 00:11:15.383 "get_zone_info": false, 00:11:15.383 "zone_management": false, 00:11:15.383 "zone_append": false, 00:11:15.383 "compare": false, 00:11:15.383 "compare_and_write": false, 00:11:15.383 "abort": true, 00:11:15.383 "seek_hole": false, 00:11:15.383 "seek_data": false, 00:11:15.383 "copy": true, 00:11:15.383 "nvme_iov_md": false 00:11:15.383 }, 00:11:15.383 "memory_domains": [ 00:11:15.383 { 00:11:15.383 "dma_device_id": "system", 00:11:15.383 "dma_device_type": 1 00:11:15.383 }, 00:11:15.383 { 00:11:15.383 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:15.383 "dma_device_type": 2 00:11:15.383 } 00:11:15.383 ], 00:11:15.383 "driver_specific": {} 00:11:15.383 }' 00:11:15.383 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:11:15.383 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:11:15.383 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:11:15.383 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:11:15.383 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:11:15.383 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:11:15.383 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:11:15.383 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:11:15.383 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:11:15.383 11:21:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:11:15.641 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:11:15.641 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:11:15.641 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:11:15.641 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:11:15.641 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:11:15.641 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:11:15.641 "name": "BaseBdev2", 00:11:15.641 "aliases": [ 00:11:15.641 "f110ab7e-706a-48be-9eb2-5c2d86924d2e" 00:11:15.641 ], 00:11:15.641 "product_name": "Malloc disk", 00:11:15.641 "block_size": 512, 00:11:15.641 "num_blocks": 65536, 00:11:15.641 "uuid": "f110ab7e-706a-48be-9eb2-5c2d86924d2e", 00:11:15.641 "assigned_rate_limits": { 00:11:15.641 "rw_ios_per_sec": 0, 00:11:15.641 "rw_mbytes_per_sec": 0, 00:11:15.641 "r_mbytes_per_sec": 0, 00:11:15.641 "w_mbytes_per_sec": 0 00:11:15.641 }, 00:11:15.641 "claimed": true, 00:11:15.641 "claim_type": "exclusive_write", 00:11:15.641 "zoned": false, 00:11:15.641 "supported_io_types": { 00:11:15.641 "read": true, 00:11:15.641 "write": true, 00:11:15.641 "unmap": true, 00:11:15.641 "flush": true, 00:11:15.641 "reset": true, 00:11:15.641 "nvme_admin": false, 00:11:15.641 "nvme_io": false, 00:11:15.641 "nvme_io_md": false, 00:11:15.641 "write_zeroes": true, 00:11:15.641 "zcopy": true, 00:11:15.641 "get_zone_info": false, 00:11:15.641 "zone_management": false, 00:11:15.641 "zone_append": false, 00:11:15.641 "compare": false, 00:11:15.641 "compare_and_write": false, 00:11:15.641 "abort": true, 00:11:15.641 "seek_hole": false, 00:11:15.641 "seek_data": false, 00:11:15.641 "copy": true, 00:11:15.641 "nvme_iov_md": false 00:11:15.641 }, 00:11:15.641 "memory_domains": [ 00:11:15.641 { 00:11:15.641 "dma_device_id": "system", 00:11:15.641 "dma_device_type": 1 00:11:15.641 }, 00:11:15.641 { 00:11:15.641 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:15.641 "dma_device_type": 2 00:11:15.641 } 00:11:15.641 ], 00:11:15.641 "driver_specific": {} 00:11:15.641 }' 00:11:15.641 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:11:15.899 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:11:15.899 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:11:15.899 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:11:15.899 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:11:15.899 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:11:15.899 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:11:15.899 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:11:15.899 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:11:15.899 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:11:16.157 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:11:16.157 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:11:16.157 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:11:16.415 [2024-07-15 11:21:59.794477] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:16.415 [2024-07-15 11:21:59.794506] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:16.415 [2024-07-15 11:21:59.794546] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:16.415 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@275 -- # local expected_state 00:11:16.415 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # has_redundancy raid0 00:11:16.415 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:11:16.415 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # return 1 00:11:16.415 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@277 -- # expected_state=offline 00:11:16.415 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:11:16.415 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:11:16.415 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=offline 00:11:16.415 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:11:16.415 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:16.415 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:11:16.415 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:16.415 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:16.415 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:16.415 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:16.415 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:16.415 11:21:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:16.672 11:22:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:16.672 "name": "Existed_Raid", 00:11:16.672 "uuid": "4aced7ec-8c7f-40be-b3a5-801f6c94bb34", 00:11:16.672 "strip_size_kb": 64, 00:11:16.673 "state": "offline", 00:11:16.673 "raid_level": "raid0", 00:11:16.673 "superblock": false, 00:11:16.673 "num_base_bdevs": 2, 00:11:16.673 "num_base_bdevs_discovered": 1, 00:11:16.673 "num_base_bdevs_operational": 1, 00:11:16.673 "base_bdevs_list": [ 00:11:16.673 { 00:11:16.673 "name": null, 00:11:16.673 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:16.673 "is_configured": false, 00:11:16.673 "data_offset": 0, 00:11:16.673 "data_size": 65536 00:11:16.673 }, 00:11:16.673 { 00:11:16.673 "name": "BaseBdev2", 00:11:16.673 "uuid": "f110ab7e-706a-48be-9eb2-5c2d86924d2e", 00:11:16.673 "is_configured": true, 00:11:16.673 "data_offset": 0, 00:11:16.673 "data_size": 65536 00:11:16.673 } 00:11:16.673 ] 00:11:16.673 }' 00:11:16.673 11:22:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:16.673 11:22:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:17.243 11:22:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:11:17.243 11:22:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:11:17.243 11:22:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:17.243 11:22:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:11:17.500 11:22:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:11:17.500 11:22:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:17.500 11:22:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:11:17.758 [2024-07-15 11:22:01.102982] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:17.758 [2024-07-15 11:22:01.103038] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1add000 name Existed_Raid, state offline 00:11:17.758 11:22:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:11:17.758 11:22:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:11:17.758 11:22:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:17.758 11:22:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:11:18.016 11:22:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:11:18.016 11:22:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:11:18.016 11:22:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # '[' 2 -gt 2 ']' 00:11:18.016 11:22:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@341 -- # killprocess 861796 00:11:18.016 11:22:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@948 -- # '[' -z 861796 ']' 00:11:18.016 11:22:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@952 -- # kill -0 861796 00:11:18.016 11:22:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@953 -- # uname 00:11:18.016 11:22:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:11:18.016 11:22:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 861796 00:11:18.016 11:22:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:11:18.016 11:22:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:11:18.016 11:22:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 861796' 00:11:18.016 killing process with pid 861796 00:11:18.016 11:22:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@967 -- # kill 861796 00:11:18.016 [2024-07-15 11:22:01.425008] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:18.016 11:22:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # wait 861796 00:11:18.016 [2024-07-15 11:22:01.426005] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@343 -- # return 0 00:11:18.274 00:11:18.274 real 0m9.956s 00:11:18.274 user 0m17.663s 00:11:18.274 sys 0m1.851s 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:18.274 ************************************ 00:11:18.274 END TEST raid_state_function_test 00:11:18.274 ************************************ 00:11:18.274 11:22:01 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:11:18.274 11:22:01 bdev_raid -- bdev/bdev_raid.sh@868 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 2 true 00:11:18.274 11:22:01 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:11:18.274 11:22:01 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:11:18.274 11:22:01 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:18.274 ************************************ 00:11:18.274 START TEST raid_state_function_test_sb 00:11:18.274 ************************************ 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1123 -- # raid_state_function_test raid0 2 true 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@220 -- # local raid_level=raid0 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=2 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # local superblock=true 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # local strip_size 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # '[' raid0 '!=' raid1 ']' 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # strip_size=64 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@232 -- # strip_size_create_arg='-z 64' 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # '[' true = true ']' 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@238 -- # superblock_create_arg=-s 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # raid_pid=863269 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 863269' 00:11:18.274 Process raid pid: 863269 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@246 -- # waitforlisten 863269 /var/tmp/spdk-raid.sock 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@829 -- # '[' -z 863269 ']' 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@834 -- # local max_retries=100 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:11:18.274 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # xtrace_disable 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:11:18.274 11:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.274 [2024-07-15 11:22:01.796726] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:11:18.274 [2024-07-15 11:22:01.796789] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:18.531 [2024-07-15 11:22:01.927948] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:18.531 [2024-07-15 11:22:02.036423] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:11:18.531 [2024-07-15 11:22:02.107221] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:18.531 [2024-07-15 11:22:02.107252] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:18.789 11:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:11:18.789 11:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@862 -- # return 0 00:11:18.789 11:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r raid0 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:11:19.047 [2024-07-15 11:22:02.476952] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:19.047 [2024-07-15 11:22:02.476992] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:19.047 [2024-07-15 11:22:02.477003] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:19.047 [2024-07-15 11:22:02.477015] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:19.047 11:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:11:19.047 11:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:11:19.047 11:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:11:19.047 11:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:11:19.047 11:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:19.047 11:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:11:19.047 11:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:19.047 11:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:19.047 11:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:19.047 11:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:19.047 11:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:19.047 11:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:19.305 11:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:19.305 "name": "Existed_Raid", 00:11:19.305 "uuid": "cf75ffb1-a751-484d-b95b-b24cff0fda28", 00:11:19.305 "strip_size_kb": 64, 00:11:19.305 "state": "configuring", 00:11:19.305 "raid_level": "raid0", 00:11:19.305 "superblock": true, 00:11:19.305 "num_base_bdevs": 2, 00:11:19.305 "num_base_bdevs_discovered": 0, 00:11:19.305 "num_base_bdevs_operational": 2, 00:11:19.305 "base_bdevs_list": [ 00:11:19.305 { 00:11:19.305 "name": "BaseBdev1", 00:11:19.305 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:19.305 "is_configured": false, 00:11:19.305 "data_offset": 0, 00:11:19.305 "data_size": 0 00:11:19.305 }, 00:11:19.305 { 00:11:19.305 "name": "BaseBdev2", 00:11:19.305 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:19.305 "is_configured": false, 00:11:19.305 "data_offset": 0, 00:11:19.305 "data_size": 0 00:11:19.305 } 00:11:19.305 ] 00:11:19.305 }' 00:11:19.305 11:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:19.305 11:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:19.869 11:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:11:19.869 [2024-07-15 11:22:03.455395] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:19.869 [2024-07-15 11:22:03.455423] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xf96a80 name Existed_Raid, state configuring 00:11:20.127 11:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r raid0 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:11:20.127 [2024-07-15 11:22:03.696065] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:20.127 [2024-07-15 11:22:03.696094] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:20.127 [2024-07-15 11:22:03.696104] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:20.127 [2024-07-15 11:22:03.696115] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:20.127 11:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:11:20.385 [2024-07-15 11:22:03.886436] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:20.385 BaseBdev1 00:11:20.385 11:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:11:20.385 11:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:11:20.385 11:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:11:20.385 11:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:11:20.385 11:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:11:20.385 11:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:11:20.385 11:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:11:20.643 11:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:20.901 [ 00:11:20.901 { 00:11:20.901 "name": "BaseBdev1", 00:11:20.901 "aliases": [ 00:11:20.901 "fe69eeff-fd63-4eab-b05e-ecf9a0954f86" 00:11:20.901 ], 00:11:20.901 "product_name": "Malloc disk", 00:11:20.901 "block_size": 512, 00:11:20.901 "num_blocks": 65536, 00:11:20.901 "uuid": "fe69eeff-fd63-4eab-b05e-ecf9a0954f86", 00:11:20.901 "assigned_rate_limits": { 00:11:20.901 "rw_ios_per_sec": 0, 00:11:20.901 "rw_mbytes_per_sec": 0, 00:11:20.901 "r_mbytes_per_sec": 0, 00:11:20.901 "w_mbytes_per_sec": 0 00:11:20.901 }, 00:11:20.901 "claimed": true, 00:11:20.901 "claim_type": "exclusive_write", 00:11:20.901 "zoned": false, 00:11:20.901 "supported_io_types": { 00:11:20.901 "read": true, 00:11:20.901 "write": true, 00:11:20.901 "unmap": true, 00:11:20.901 "flush": true, 00:11:20.901 "reset": true, 00:11:20.901 "nvme_admin": false, 00:11:20.901 "nvme_io": false, 00:11:20.901 "nvme_io_md": false, 00:11:20.901 "write_zeroes": true, 00:11:20.901 "zcopy": true, 00:11:20.901 "get_zone_info": false, 00:11:20.901 "zone_management": false, 00:11:20.901 "zone_append": false, 00:11:20.901 "compare": false, 00:11:20.901 "compare_and_write": false, 00:11:20.901 "abort": true, 00:11:20.901 "seek_hole": false, 00:11:20.901 "seek_data": false, 00:11:20.901 "copy": true, 00:11:20.901 "nvme_iov_md": false 00:11:20.901 }, 00:11:20.901 "memory_domains": [ 00:11:20.901 { 00:11:20.901 "dma_device_id": "system", 00:11:20.901 "dma_device_type": 1 00:11:20.901 }, 00:11:20.901 { 00:11:20.901 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:20.901 "dma_device_type": 2 00:11:20.901 } 00:11:20.901 ], 00:11:20.901 "driver_specific": {} 00:11:20.901 } 00:11:20.901 ] 00:11:20.901 11:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:11:20.901 11:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:11:20.901 11:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:11:20.901 11:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:11:20.901 11:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:11:20.901 11:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:20.901 11:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:11:20.902 11:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:20.902 11:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:20.902 11:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:20.902 11:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:20.902 11:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:20.902 11:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:21.159 11:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:21.159 "name": "Existed_Raid", 00:11:21.159 "uuid": "1aa7f6bd-b2d0-41f9-9e6d-24548ae84a64", 00:11:21.159 "strip_size_kb": 64, 00:11:21.159 "state": "configuring", 00:11:21.159 "raid_level": "raid0", 00:11:21.159 "superblock": true, 00:11:21.159 "num_base_bdevs": 2, 00:11:21.159 "num_base_bdevs_discovered": 1, 00:11:21.159 "num_base_bdevs_operational": 2, 00:11:21.159 "base_bdevs_list": [ 00:11:21.159 { 00:11:21.159 "name": "BaseBdev1", 00:11:21.159 "uuid": "fe69eeff-fd63-4eab-b05e-ecf9a0954f86", 00:11:21.159 "is_configured": true, 00:11:21.159 "data_offset": 2048, 00:11:21.159 "data_size": 63488 00:11:21.159 }, 00:11:21.159 { 00:11:21.159 "name": "BaseBdev2", 00:11:21.159 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:21.159 "is_configured": false, 00:11:21.159 "data_offset": 0, 00:11:21.159 "data_size": 0 00:11:21.159 } 00:11:21.159 ] 00:11:21.159 }' 00:11:21.159 11:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:21.159 11:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:21.723 11:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:11:21.980 [2024-07-15 11:22:05.338270] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:21.980 [2024-07-15 11:22:05.338306] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xf96350 name Existed_Raid, state configuring 00:11:21.980 11:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r raid0 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:11:22.238 [2024-07-15 11:22:05.586968] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:22.238 [2024-07-15 11:22:05.588517] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:22.238 [2024-07-15 11:22:05.588548] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:22.238 11:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:11:22.238 11:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:11:22.238 11:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:11:22.238 11:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:11:22.238 11:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:11:22.238 11:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:11:22.238 11:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:22.238 11:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:11:22.238 11:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:22.238 11:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:22.238 11:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:22.238 11:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:22.238 11:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:22.238 11:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:22.238 11:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:22.238 "name": "Existed_Raid", 00:11:22.238 "uuid": "bd8a5ac4-17b8-41b9-abf5-2015f5354666", 00:11:22.238 "strip_size_kb": 64, 00:11:22.238 "state": "configuring", 00:11:22.238 "raid_level": "raid0", 00:11:22.238 "superblock": true, 00:11:22.238 "num_base_bdevs": 2, 00:11:22.238 "num_base_bdevs_discovered": 1, 00:11:22.238 "num_base_bdevs_operational": 2, 00:11:22.238 "base_bdevs_list": [ 00:11:22.238 { 00:11:22.238 "name": "BaseBdev1", 00:11:22.238 "uuid": "fe69eeff-fd63-4eab-b05e-ecf9a0954f86", 00:11:22.238 "is_configured": true, 00:11:22.238 "data_offset": 2048, 00:11:22.238 "data_size": 63488 00:11:22.238 }, 00:11:22.238 { 00:11:22.238 "name": "BaseBdev2", 00:11:22.238 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:22.238 "is_configured": false, 00:11:22.238 "data_offset": 0, 00:11:22.238 "data_size": 0 00:11:22.238 } 00:11:22.238 ] 00:11:22.238 }' 00:11:22.238 11:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:22.238 11:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:22.803 11:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:11:23.059 [2024-07-15 11:22:06.610207] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:23.059 [2024-07-15 11:22:06.610360] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xf97000 00:11:23.059 [2024-07-15 11:22:06.610373] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:11:23.059 [2024-07-15 11:22:06.610546] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xeb10c0 00:11:23.059 [2024-07-15 11:22:06.610660] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xf97000 00:11:23.059 [2024-07-15 11:22:06.610670] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0xf97000 00:11:23.059 [2024-07-15 11:22:06.610761] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:23.059 BaseBdev2 00:11:23.059 11:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:11:23.059 11:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:11:23.059 11:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:11:23.059 11:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:11:23.059 11:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:11:23.059 11:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:11:23.059 11:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:11:23.347 11:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:23.605 [ 00:11:23.605 { 00:11:23.605 "name": "BaseBdev2", 00:11:23.605 "aliases": [ 00:11:23.605 "81b1e68c-732d-4642-93df-6ac689e70490" 00:11:23.605 ], 00:11:23.605 "product_name": "Malloc disk", 00:11:23.605 "block_size": 512, 00:11:23.605 "num_blocks": 65536, 00:11:23.605 "uuid": "81b1e68c-732d-4642-93df-6ac689e70490", 00:11:23.605 "assigned_rate_limits": { 00:11:23.605 "rw_ios_per_sec": 0, 00:11:23.605 "rw_mbytes_per_sec": 0, 00:11:23.605 "r_mbytes_per_sec": 0, 00:11:23.605 "w_mbytes_per_sec": 0 00:11:23.605 }, 00:11:23.605 "claimed": true, 00:11:23.605 "claim_type": "exclusive_write", 00:11:23.605 "zoned": false, 00:11:23.605 "supported_io_types": { 00:11:23.605 "read": true, 00:11:23.605 "write": true, 00:11:23.605 "unmap": true, 00:11:23.605 "flush": true, 00:11:23.605 "reset": true, 00:11:23.605 "nvme_admin": false, 00:11:23.605 "nvme_io": false, 00:11:23.605 "nvme_io_md": false, 00:11:23.605 "write_zeroes": true, 00:11:23.605 "zcopy": true, 00:11:23.605 "get_zone_info": false, 00:11:23.605 "zone_management": false, 00:11:23.605 "zone_append": false, 00:11:23.605 "compare": false, 00:11:23.605 "compare_and_write": false, 00:11:23.605 "abort": true, 00:11:23.605 "seek_hole": false, 00:11:23.605 "seek_data": false, 00:11:23.605 "copy": true, 00:11:23.605 "nvme_iov_md": false 00:11:23.605 }, 00:11:23.605 "memory_domains": [ 00:11:23.605 { 00:11:23.605 "dma_device_id": "system", 00:11:23.605 "dma_device_type": 1 00:11:23.605 }, 00:11:23.605 { 00:11:23.605 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:23.605 "dma_device_type": 2 00:11:23.605 } 00:11:23.605 ], 00:11:23.605 "driver_specific": {} 00:11:23.605 } 00:11:23.605 ] 00:11:23.605 11:22:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:11:23.605 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:11:23.605 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:11:23.605 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:11:23.605 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:11:23.605 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:11:23.605 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:11:23.605 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:23.605 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:11:23.605 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:23.605 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:23.605 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:23.605 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:23.605 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:23.605 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:23.863 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:23.863 "name": "Existed_Raid", 00:11:23.863 "uuid": "bd8a5ac4-17b8-41b9-abf5-2015f5354666", 00:11:23.863 "strip_size_kb": 64, 00:11:23.863 "state": "online", 00:11:23.863 "raid_level": "raid0", 00:11:23.863 "superblock": true, 00:11:23.863 "num_base_bdevs": 2, 00:11:23.863 "num_base_bdevs_discovered": 2, 00:11:23.863 "num_base_bdevs_operational": 2, 00:11:23.863 "base_bdevs_list": [ 00:11:23.863 { 00:11:23.863 "name": "BaseBdev1", 00:11:23.863 "uuid": "fe69eeff-fd63-4eab-b05e-ecf9a0954f86", 00:11:23.863 "is_configured": true, 00:11:23.863 "data_offset": 2048, 00:11:23.863 "data_size": 63488 00:11:23.863 }, 00:11:23.863 { 00:11:23.863 "name": "BaseBdev2", 00:11:23.863 "uuid": "81b1e68c-732d-4642-93df-6ac689e70490", 00:11:23.863 "is_configured": true, 00:11:23.863 "data_offset": 2048, 00:11:23.863 "data_size": 63488 00:11:23.863 } 00:11:23.863 ] 00:11:23.863 }' 00:11:23.863 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:23.863 11:22:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:24.426 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:11:24.426 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:11:24.426 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:11:24.426 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:11:24.426 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:11:24.426 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # local name 00:11:24.426 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:11:24.427 11:22:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:11:24.683 [2024-07-15 11:22:08.174622] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:24.683 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:11:24.683 "name": "Existed_Raid", 00:11:24.683 "aliases": [ 00:11:24.683 "bd8a5ac4-17b8-41b9-abf5-2015f5354666" 00:11:24.683 ], 00:11:24.683 "product_name": "Raid Volume", 00:11:24.683 "block_size": 512, 00:11:24.683 "num_blocks": 126976, 00:11:24.683 "uuid": "bd8a5ac4-17b8-41b9-abf5-2015f5354666", 00:11:24.683 "assigned_rate_limits": { 00:11:24.683 "rw_ios_per_sec": 0, 00:11:24.683 "rw_mbytes_per_sec": 0, 00:11:24.683 "r_mbytes_per_sec": 0, 00:11:24.683 "w_mbytes_per_sec": 0 00:11:24.683 }, 00:11:24.683 "claimed": false, 00:11:24.683 "zoned": false, 00:11:24.683 "supported_io_types": { 00:11:24.683 "read": true, 00:11:24.683 "write": true, 00:11:24.683 "unmap": true, 00:11:24.683 "flush": true, 00:11:24.683 "reset": true, 00:11:24.683 "nvme_admin": false, 00:11:24.683 "nvme_io": false, 00:11:24.683 "nvme_io_md": false, 00:11:24.683 "write_zeroes": true, 00:11:24.683 "zcopy": false, 00:11:24.683 "get_zone_info": false, 00:11:24.683 "zone_management": false, 00:11:24.683 "zone_append": false, 00:11:24.683 "compare": false, 00:11:24.683 "compare_and_write": false, 00:11:24.683 "abort": false, 00:11:24.683 "seek_hole": false, 00:11:24.683 "seek_data": false, 00:11:24.683 "copy": false, 00:11:24.683 "nvme_iov_md": false 00:11:24.683 }, 00:11:24.683 "memory_domains": [ 00:11:24.683 { 00:11:24.683 "dma_device_id": "system", 00:11:24.683 "dma_device_type": 1 00:11:24.683 }, 00:11:24.683 { 00:11:24.683 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:24.683 "dma_device_type": 2 00:11:24.683 }, 00:11:24.683 { 00:11:24.683 "dma_device_id": "system", 00:11:24.683 "dma_device_type": 1 00:11:24.683 }, 00:11:24.683 { 00:11:24.683 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:24.683 "dma_device_type": 2 00:11:24.683 } 00:11:24.683 ], 00:11:24.683 "driver_specific": { 00:11:24.683 "raid": { 00:11:24.683 "uuid": "bd8a5ac4-17b8-41b9-abf5-2015f5354666", 00:11:24.683 "strip_size_kb": 64, 00:11:24.683 "state": "online", 00:11:24.683 "raid_level": "raid0", 00:11:24.683 "superblock": true, 00:11:24.683 "num_base_bdevs": 2, 00:11:24.683 "num_base_bdevs_discovered": 2, 00:11:24.683 "num_base_bdevs_operational": 2, 00:11:24.683 "base_bdevs_list": [ 00:11:24.683 { 00:11:24.683 "name": "BaseBdev1", 00:11:24.683 "uuid": "fe69eeff-fd63-4eab-b05e-ecf9a0954f86", 00:11:24.683 "is_configured": true, 00:11:24.683 "data_offset": 2048, 00:11:24.683 "data_size": 63488 00:11:24.683 }, 00:11:24.683 { 00:11:24.683 "name": "BaseBdev2", 00:11:24.683 "uuid": "81b1e68c-732d-4642-93df-6ac689e70490", 00:11:24.683 "is_configured": true, 00:11:24.683 "data_offset": 2048, 00:11:24.683 "data_size": 63488 00:11:24.683 } 00:11:24.683 ] 00:11:24.683 } 00:11:24.683 } 00:11:24.683 }' 00:11:24.683 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:24.683 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:11:24.683 BaseBdev2' 00:11:24.683 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:11:24.683 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:11:24.683 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:11:24.940 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:11:24.940 "name": "BaseBdev1", 00:11:24.940 "aliases": [ 00:11:24.940 "fe69eeff-fd63-4eab-b05e-ecf9a0954f86" 00:11:24.940 ], 00:11:24.940 "product_name": "Malloc disk", 00:11:24.940 "block_size": 512, 00:11:24.940 "num_blocks": 65536, 00:11:24.940 "uuid": "fe69eeff-fd63-4eab-b05e-ecf9a0954f86", 00:11:24.940 "assigned_rate_limits": { 00:11:24.940 "rw_ios_per_sec": 0, 00:11:24.940 "rw_mbytes_per_sec": 0, 00:11:24.940 "r_mbytes_per_sec": 0, 00:11:24.940 "w_mbytes_per_sec": 0 00:11:24.940 }, 00:11:24.940 "claimed": true, 00:11:24.940 "claim_type": "exclusive_write", 00:11:24.940 "zoned": false, 00:11:24.940 "supported_io_types": { 00:11:24.940 "read": true, 00:11:24.940 "write": true, 00:11:24.940 "unmap": true, 00:11:24.940 "flush": true, 00:11:24.940 "reset": true, 00:11:24.940 "nvme_admin": false, 00:11:24.940 "nvme_io": false, 00:11:24.940 "nvme_io_md": false, 00:11:24.940 "write_zeroes": true, 00:11:24.940 "zcopy": true, 00:11:24.940 "get_zone_info": false, 00:11:24.940 "zone_management": false, 00:11:24.940 "zone_append": false, 00:11:24.940 "compare": false, 00:11:24.940 "compare_and_write": false, 00:11:24.940 "abort": true, 00:11:24.940 "seek_hole": false, 00:11:24.940 "seek_data": false, 00:11:24.940 "copy": true, 00:11:24.940 "nvme_iov_md": false 00:11:24.940 }, 00:11:24.940 "memory_domains": [ 00:11:24.940 { 00:11:24.940 "dma_device_id": "system", 00:11:24.940 "dma_device_type": 1 00:11:24.940 }, 00:11:24.940 { 00:11:24.940 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:24.940 "dma_device_type": 2 00:11:24.940 } 00:11:24.940 ], 00:11:24.940 "driver_specific": {} 00:11:24.940 }' 00:11:24.940 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:11:24.940 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:11:25.197 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:11:25.197 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:11:25.197 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:11:25.197 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:11:25.197 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:11:25.197 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:11:25.197 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:11:25.197 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:11:25.455 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:11:25.455 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:11:25.455 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:11:25.455 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:11:25.455 11:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:11:25.712 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:11:25.712 "name": "BaseBdev2", 00:11:25.712 "aliases": [ 00:11:25.712 "81b1e68c-732d-4642-93df-6ac689e70490" 00:11:25.712 ], 00:11:25.712 "product_name": "Malloc disk", 00:11:25.712 "block_size": 512, 00:11:25.712 "num_blocks": 65536, 00:11:25.712 "uuid": "81b1e68c-732d-4642-93df-6ac689e70490", 00:11:25.712 "assigned_rate_limits": { 00:11:25.712 "rw_ios_per_sec": 0, 00:11:25.712 "rw_mbytes_per_sec": 0, 00:11:25.712 "r_mbytes_per_sec": 0, 00:11:25.712 "w_mbytes_per_sec": 0 00:11:25.712 }, 00:11:25.712 "claimed": true, 00:11:25.712 "claim_type": "exclusive_write", 00:11:25.712 "zoned": false, 00:11:25.712 "supported_io_types": { 00:11:25.712 "read": true, 00:11:25.712 "write": true, 00:11:25.712 "unmap": true, 00:11:25.712 "flush": true, 00:11:25.712 "reset": true, 00:11:25.712 "nvme_admin": false, 00:11:25.712 "nvme_io": false, 00:11:25.712 "nvme_io_md": false, 00:11:25.712 "write_zeroes": true, 00:11:25.712 "zcopy": true, 00:11:25.712 "get_zone_info": false, 00:11:25.712 "zone_management": false, 00:11:25.712 "zone_append": false, 00:11:25.712 "compare": false, 00:11:25.712 "compare_and_write": false, 00:11:25.712 "abort": true, 00:11:25.712 "seek_hole": false, 00:11:25.712 "seek_data": false, 00:11:25.712 "copy": true, 00:11:25.712 "nvme_iov_md": false 00:11:25.712 }, 00:11:25.712 "memory_domains": [ 00:11:25.712 { 00:11:25.712 "dma_device_id": "system", 00:11:25.712 "dma_device_type": 1 00:11:25.712 }, 00:11:25.712 { 00:11:25.712 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:25.712 "dma_device_type": 2 00:11:25.712 } 00:11:25.712 ], 00:11:25.712 "driver_specific": {} 00:11:25.712 }' 00:11:25.712 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:11:25.712 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:11:25.712 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:11:25.713 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:11:25.713 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:11:25.713 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:11:25.713 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:11:25.970 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:11:25.970 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:11:25.970 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:11:25.970 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:11:25.970 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:11:25.970 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:11:26.228 [2024-07-15 11:22:09.658328] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:26.228 [2024-07-15 11:22:09.658353] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:26.228 [2024-07-15 11:22:09.658394] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:26.228 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@275 -- # local expected_state 00:11:26.228 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # has_redundancy raid0 00:11:26.228 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # case $1 in 00:11:26.228 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # return 1 00:11:26.228 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@277 -- # expected_state=offline 00:11:26.228 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:11:26.228 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:11:26.228 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=offline 00:11:26.228 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:11:26.228 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:26.228 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:11:26.228 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:26.228 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:26.228 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:26.228 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:26.228 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:26.228 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:26.503 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:26.503 "name": "Existed_Raid", 00:11:26.503 "uuid": "bd8a5ac4-17b8-41b9-abf5-2015f5354666", 00:11:26.503 "strip_size_kb": 64, 00:11:26.503 "state": "offline", 00:11:26.503 "raid_level": "raid0", 00:11:26.503 "superblock": true, 00:11:26.503 "num_base_bdevs": 2, 00:11:26.503 "num_base_bdevs_discovered": 1, 00:11:26.503 "num_base_bdevs_operational": 1, 00:11:26.503 "base_bdevs_list": [ 00:11:26.503 { 00:11:26.503 "name": null, 00:11:26.503 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:26.503 "is_configured": false, 00:11:26.503 "data_offset": 2048, 00:11:26.503 "data_size": 63488 00:11:26.503 }, 00:11:26.503 { 00:11:26.503 "name": "BaseBdev2", 00:11:26.503 "uuid": "81b1e68c-732d-4642-93df-6ac689e70490", 00:11:26.503 "is_configured": true, 00:11:26.503 "data_offset": 2048, 00:11:26.503 "data_size": 63488 00:11:26.503 } 00:11:26.503 ] 00:11:26.503 }' 00:11:26.503 11:22:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:26.503 11:22:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.068 11:22:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:11:27.068 11:22:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:11:27.068 11:22:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:27.068 11:22:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:11:27.326 11:22:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:11:27.326 11:22:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:27.326 11:22:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:11:27.602 [2024-07-15 11:22:10.934792] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:27.602 [2024-07-15 11:22:10.934843] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xf97000 name Existed_Raid, state offline 00:11:27.602 11:22:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:11:27.602 11:22:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:11:27.602 11:22:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:27.602 11:22:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:11:27.859 11:22:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:11:27.859 11:22:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:11:27.859 11:22:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # '[' 2 -gt 2 ']' 00:11:27.859 11:22:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@341 -- # killprocess 863269 00:11:27.860 11:22:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@948 -- # '[' -z 863269 ']' 00:11:27.860 11:22:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@952 -- # kill -0 863269 00:11:27.860 11:22:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@953 -- # uname 00:11:27.860 11:22:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:11:27.860 11:22:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 863269 00:11:27.860 11:22:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:11:27.860 11:22:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:11:27.860 11:22:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@966 -- # echo 'killing process with pid 863269' 00:11:27.860 killing process with pid 863269 00:11:27.860 11:22:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@967 -- # kill 863269 00:11:27.860 [2024-07-15 11:22:11.281480] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:27.860 11:22:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # wait 863269 00:11:27.860 [2024-07-15 11:22:11.282382] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:28.117 11:22:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@343 -- # return 0 00:11:28.117 00:11:28.117 real 0m9.771s 00:11:28.117 user 0m17.620s 00:11:28.117 sys 0m2.023s 00:11:28.117 11:22:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1124 -- # xtrace_disable 00:11:28.117 11:22:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.117 ************************************ 00:11:28.117 END TEST raid_state_function_test_sb 00:11:28.117 ************************************ 00:11:28.117 11:22:11 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:11:28.117 11:22:11 bdev_raid -- bdev/bdev_raid.sh@869 -- # run_test raid_superblock_test raid_superblock_test raid0 2 00:11:28.117 11:22:11 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:11:28.117 11:22:11 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:11:28.117 11:22:11 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:28.117 ************************************ 00:11:28.117 START TEST raid_superblock_test 00:11:28.117 ************************************ 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1123 -- # raid_superblock_test raid0 2 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@392 -- # local raid_level=raid0 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local num_base_bdevs=2 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # base_bdevs_malloc=() 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local base_bdevs_malloc 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_pt=() 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_pt 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt_uuid=() 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt_uuid 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local raid_bdev_name=raid_bdev1 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local strip_size 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size_create_arg 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local raid_bdev_uuid 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@403 -- # '[' raid0 '!=' raid1 ']' 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # strip_size=64 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size_create_arg='-z 64' 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # raid_pid=864767 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@410 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -L bdev_raid 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # waitforlisten 864767 /var/tmp/spdk-raid.sock 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@829 -- # '[' -z 864767 ']' 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:11:28.117 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:11:28.117 11:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:28.117 [2024-07-15 11:22:11.643245] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:11:28.117 [2024-07-15 11:22:11.643311] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid864767 ] 00:11:28.375 [2024-07-15 11:22:11.775342] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:28.375 [2024-07-15 11:22:11.879885] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:11:28.375 [2024-07-15 11:22:11.942429] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:28.375 [2024-07-15 11:22:11.942466] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:29.309 11:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:11:29.309 11:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@862 -- # return 0 00:11:29.309 11:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i = 1 )) 00:11:29.309 11:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:11:29.309 11:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc1 00:11:29.309 11:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt1 00:11:29.309 11:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:11:29.309 11:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:29.309 11:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:11:29.309 11:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:29.309 11:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc1 00:11:29.309 malloc1 00:11:29.309 11:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:11:29.566 [2024-07-15 11:22:13.039010] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:11:29.566 [2024-07-15 11:22:13.039057] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:29.566 [2024-07-15 11:22:13.039077] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1b13570 00:11:29.566 [2024-07-15 11:22:13.039090] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:29.566 [2024-07-15 11:22:13.040897] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:29.566 [2024-07-15 11:22:13.040938] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:11:29.566 pt1 00:11:29.566 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:11:29.566 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:11:29.566 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc2 00:11:29.566 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt2 00:11:29.566 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:11:29.567 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:29.567 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:11:29.567 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:29.567 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc2 00:11:29.823 malloc2 00:11:29.823 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:30.080 [2024-07-15 11:22:13.521973] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:30.080 [2024-07-15 11:22:13.522021] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:30.080 [2024-07-15 11:22:13.522039] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1b14970 00:11:30.080 [2024-07-15 11:22:13.522051] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:30.080 [2024-07-15 11:22:13.523655] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:30.080 [2024-07-15 11:22:13.523683] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:30.080 pt2 00:11:30.080 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:11:30.080 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:11:30.080 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@429 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'pt1 pt2' -n raid_bdev1 -s 00:11:30.336 [2024-07-15 11:22:13.762622] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:11:30.336 [2024-07-15 11:22:13.763962] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:30.336 [2024-07-15 11:22:13.764102] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1cb7270 00:11:30.336 [2024-07-15 11:22:13.764115] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:11:30.336 [2024-07-15 11:22:13.764314] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1cacc10 00:11:30.336 [2024-07-15 11:22:13.764460] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1cb7270 00:11:30.336 [2024-07-15 11:22:13.764470] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x1cb7270 00:11:30.336 [2024-07-15 11:22:13.764570] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:30.336 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:11:30.336 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:11:30.336 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:11:30.336 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:11:30.336 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:30.336 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:11:30.336 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:30.336 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:30.336 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:30.336 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:30.336 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:30.336 11:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:30.593 11:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:30.593 "name": "raid_bdev1", 00:11:30.593 "uuid": "d547bd9c-c94e-4369-ad2a-cef7844502ec", 00:11:30.593 "strip_size_kb": 64, 00:11:30.593 "state": "online", 00:11:30.593 "raid_level": "raid0", 00:11:30.593 "superblock": true, 00:11:30.593 "num_base_bdevs": 2, 00:11:30.593 "num_base_bdevs_discovered": 2, 00:11:30.593 "num_base_bdevs_operational": 2, 00:11:30.593 "base_bdevs_list": [ 00:11:30.593 { 00:11:30.593 "name": "pt1", 00:11:30.593 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:30.593 "is_configured": true, 00:11:30.593 "data_offset": 2048, 00:11:30.593 "data_size": 63488 00:11:30.593 }, 00:11:30.593 { 00:11:30.593 "name": "pt2", 00:11:30.593 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:30.593 "is_configured": true, 00:11:30.593 "data_offset": 2048, 00:11:30.593 "data_size": 63488 00:11:30.593 } 00:11:30.593 ] 00:11:30.593 }' 00:11:30.593 11:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:30.593 11:22:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.171 11:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_properties raid_bdev1 00:11:31.171 11:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:11:31.171 11:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:11:31.171 11:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:11:31.171 11:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:11:31.171 11:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # local name 00:11:31.171 11:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:11:31.171 11:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:11:31.427 [2024-07-15 11:22:14.853743] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:31.427 11:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:11:31.427 "name": "raid_bdev1", 00:11:31.427 "aliases": [ 00:11:31.427 "d547bd9c-c94e-4369-ad2a-cef7844502ec" 00:11:31.427 ], 00:11:31.428 "product_name": "Raid Volume", 00:11:31.428 "block_size": 512, 00:11:31.428 "num_blocks": 126976, 00:11:31.428 "uuid": "d547bd9c-c94e-4369-ad2a-cef7844502ec", 00:11:31.428 "assigned_rate_limits": { 00:11:31.428 "rw_ios_per_sec": 0, 00:11:31.428 "rw_mbytes_per_sec": 0, 00:11:31.428 "r_mbytes_per_sec": 0, 00:11:31.428 "w_mbytes_per_sec": 0 00:11:31.428 }, 00:11:31.428 "claimed": false, 00:11:31.428 "zoned": false, 00:11:31.428 "supported_io_types": { 00:11:31.428 "read": true, 00:11:31.428 "write": true, 00:11:31.428 "unmap": true, 00:11:31.428 "flush": true, 00:11:31.428 "reset": true, 00:11:31.428 "nvme_admin": false, 00:11:31.428 "nvme_io": false, 00:11:31.428 "nvme_io_md": false, 00:11:31.428 "write_zeroes": true, 00:11:31.428 "zcopy": false, 00:11:31.428 "get_zone_info": false, 00:11:31.428 "zone_management": false, 00:11:31.428 "zone_append": false, 00:11:31.428 "compare": false, 00:11:31.428 "compare_and_write": false, 00:11:31.428 "abort": false, 00:11:31.428 "seek_hole": false, 00:11:31.428 "seek_data": false, 00:11:31.428 "copy": false, 00:11:31.428 "nvme_iov_md": false 00:11:31.428 }, 00:11:31.428 "memory_domains": [ 00:11:31.428 { 00:11:31.428 "dma_device_id": "system", 00:11:31.428 "dma_device_type": 1 00:11:31.428 }, 00:11:31.428 { 00:11:31.428 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:31.428 "dma_device_type": 2 00:11:31.428 }, 00:11:31.428 { 00:11:31.428 "dma_device_id": "system", 00:11:31.428 "dma_device_type": 1 00:11:31.428 }, 00:11:31.428 { 00:11:31.428 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:31.428 "dma_device_type": 2 00:11:31.428 } 00:11:31.428 ], 00:11:31.428 "driver_specific": { 00:11:31.428 "raid": { 00:11:31.428 "uuid": "d547bd9c-c94e-4369-ad2a-cef7844502ec", 00:11:31.428 "strip_size_kb": 64, 00:11:31.428 "state": "online", 00:11:31.428 "raid_level": "raid0", 00:11:31.428 "superblock": true, 00:11:31.428 "num_base_bdevs": 2, 00:11:31.428 "num_base_bdevs_discovered": 2, 00:11:31.428 "num_base_bdevs_operational": 2, 00:11:31.428 "base_bdevs_list": [ 00:11:31.428 { 00:11:31.428 "name": "pt1", 00:11:31.428 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:31.428 "is_configured": true, 00:11:31.428 "data_offset": 2048, 00:11:31.428 "data_size": 63488 00:11:31.428 }, 00:11:31.428 { 00:11:31.428 "name": "pt2", 00:11:31.428 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:31.428 "is_configured": true, 00:11:31.428 "data_offset": 2048, 00:11:31.428 "data_size": 63488 00:11:31.428 } 00:11:31.428 ] 00:11:31.428 } 00:11:31.428 } 00:11:31.428 }' 00:11:31.428 11:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:31.428 11:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:11:31.428 pt2' 00:11:31.428 11:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:11:31.428 11:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:11:31.428 11:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:11:31.685 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:11:31.685 "name": "pt1", 00:11:31.685 "aliases": [ 00:11:31.685 "00000000-0000-0000-0000-000000000001" 00:11:31.685 ], 00:11:31.685 "product_name": "passthru", 00:11:31.685 "block_size": 512, 00:11:31.685 "num_blocks": 65536, 00:11:31.685 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:31.685 "assigned_rate_limits": { 00:11:31.685 "rw_ios_per_sec": 0, 00:11:31.685 "rw_mbytes_per_sec": 0, 00:11:31.685 "r_mbytes_per_sec": 0, 00:11:31.685 "w_mbytes_per_sec": 0 00:11:31.685 }, 00:11:31.685 "claimed": true, 00:11:31.685 "claim_type": "exclusive_write", 00:11:31.685 "zoned": false, 00:11:31.685 "supported_io_types": { 00:11:31.685 "read": true, 00:11:31.685 "write": true, 00:11:31.685 "unmap": true, 00:11:31.685 "flush": true, 00:11:31.685 "reset": true, 00:11:31.685 "nvme_admin": false, 00:11:31.685 "nvme_io": false, 00:11:31.685 "nvme_io_md": false, 00:11:31.685 "write_zeroes": true, 00:11:31.685 "zcopy": true, 00:11:31.685 "get_zone_info": false, 00:11:31.685 "zone_management": false, 00:11:31.685 "zone_append": false, 00:11:31.685 "compare": false, 00:11:31.685 "compare_and_write": false, 00:11:31.685 "abort": true, 00:11:31.685 "seek_hole": false, 00:11:31.685 "seek_data": false, 00:11:31.685 "copy": true, 00:11:31.685 "nvme_iov_md": false 00:11:31.685 }, 00:11:31.685 "memory_domains": [ 00:11:31.685 { 00:11:31.685 "dma_device_id": "system", 00:11:31.685 "dma_device_type": 1 00:11:31.685 }, 00:11:31.685 { 00:11:31.685 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:31.685 "dma_device_type": 2 00:11:31.685 } 00:11:31.685 ], 00:11:31.685 "driver_specific": { 00:11:31.685 "passthru": { 00:11:31.685 "name": "pt1", 00:11:31.685 "base_bdev_name": "malloc1" 00:11:31.685 } 00:11:31.685 } 00:11:31.685 }' 00:11:31.685 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:11:31.685 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:11:31.685 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:11:31.685 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:11:31.685 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:11:31.685 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:11:31.941 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:11:31.941 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:11:31.941 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:11:31.941 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:11:31.941 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:11:31.941 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:11:31.941 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:11:31.941 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:11:31.941 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:11:32.198 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:11:32.198 "name": "pt2", 00:11:32.198 "aliases": [ 00:11:32.198 "00000000-0000-0000-0000-000000000002" 00:11:32.198 ], 00:11:32.198 "product_name": "passthru", 00:11:32.198 "block_size": 512, 00:11:32.198 "num_blocks": 65536, 00:11:32.198 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:32.198 "assigned_rate_limits": { 00:11:32.198 "rw_ios_per_sec": 0, 00:11:32.198 "rw_mbytes_per_sec": 0, 00:11:32.198 "r_mbytes_per_sec": 0, 00:11:32.198 "w_mbytes_per_sec": 0 00:11:32.198 }, 00:11:32.198 "claimed": true, 00:11:32.198 "claim_type": "exclusive_write", 00:11:32.198 "zoned": false, 00:11:32.198 "supported_io_types": { 00:11:32.198 "read": true, 00:11:32.198 "write": true, 00:11:32.198 "unmap": true, 00:11:32.198 "flush": true, 00:11:32.198 "reset": true, 00:11:32.198 "nvme_admin": false, 00:11:32.198 "nvme_io": false, 00:11:32.198 "nvme_io_md": false, 00:11:32.198 "write_zeroes": true, 00:11:32.198 "zcopy": true, 00:11:32.198 "get_zone_info": false, 00:11:32.198 "zone_management": false, 00:11:32.198 "zone_append": false, 00:11:32.198 "compare": false, 00:11:32.198 "compare_and_write": false, 00:11:32.198 "abort": true, 00:11:32.198 "seek_hole": false, 00:11:32.198 "seek_data": false, 00:11:32.198 "copy": true, 00:11:32.198 "nvme_iov_md": false 00:11:32.198 }, 00:11:32.198 "memory_domains": [ 00:11:32.198 { 00:11:32.198 "dma_device_id": "system", 00:11:32.198 "dma_device_type": 1 00:11:32.198 }, 00:11:32.198 { 00:11:32.198 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:32.198 "dma_device_type": 2 00:11:32.198 } 00:11:32.198 ], 00:11:32.198 "driver_specific": { 00:11:32.198 "passthru": { 00:11:32.198 "name": "pt2", 00:11:32.198 "base_bdev_name": "malloc2" 00:11:32.198 } 00:11:32.198 } 00:11:32.198 }' 00:11:32.198 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:11:32.198 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:11:32.198 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:11:32.456 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:11:32.456 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:11:32.456 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:11:32.456 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:11:32.456 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:11:32.456 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:11:32.456 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:11:32.456 11:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:11:32.456 11:22:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:11:32.456 11:22:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:11:32.456 11:22:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # jq -r '.[] | .uuid' 00:11:32.713 [2024-07-15 11:22:16.249420] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:32.713 11:22:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # raid_bdev_uuid=d547bd9c-c94e-4369-ad2a-cef7844502ec 00:11:32.713 11:22:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # '[' -z d547bd9c-c94e-4369-ad2a-cef7844502ec ']' 00:11:32.713 11:22:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@440 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:11:32.970 [2024-07-15 11:22:16.493824] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:32.970 [2024-07-15 11:22:16.493843] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:32.970 [2024-07-15 11:22:16.493895] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:32.970 [2024-07-15 11:22:16.493946] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:32.971 [2024-07-15 11:22:16.493958] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1cb7270 name raid_bdev1, state offline 00:11:32.971 11:22:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:32.971 11:22:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # jq -r '.[]' 00:11:33.228 11:22:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # raid_bdev= 00:11:33.228 11:22:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # '[' -n '' ']' 00:11:33.228 11:22:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:11:33.228 11:22:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt1 00:11:33.486 11:22:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:11:33.486 11:22:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:11:33.744 11:22:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs 00:11:33.744 11:22:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:11:34.002 11:22:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # '[' false == true ']' 00:11:34.002 11:22:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@456 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'malloc1 malloc2' -n raid_bdev1 00:11:34.002 11:22:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@648 -- # local es=0 00:11:34.002 11:22:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'malloc1 malloc2' -n raid_bdev1 00:11:34.002 11:22:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:11:34.002 11:22:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:11:34.002 11:22:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:11:34.002 11:22:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:11:34.002 11:22:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:11:34.003 11:22:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:11:34.003 11:22:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:11:34.003 11:22:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py ]] 00:11:34.003 11:22:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'malloc1 malloc2' -n raid_bdev1 00:11:34.260 [2024-07-15 11:22:17.709012] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:11:34.260 [2024-07-15 11:22:17.710347] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:11:34.260 [2024-07-15 11:22:17.710401] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:11:34.260 [2024-07-15 11:22:17.710438] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:11:34.260 [2024-07-15 11:22:17.710457] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:34.260 [2024-07-15 11:22:17.710467] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1cb6ff0 name raid_bdev1, state configuring 00:11:34.260 request: 00:11:34.260 { 00:11:34.260 "name": "raid_bdev1", 00:11:34.260 "raid_level": "raid0", 00:11:34.260 "base_bdevs": [ 00:11:34.260 "malloc1", 00:11:34.260 "malloc2" 00:11:34.260 ], 00:11:34.260 "strip_size_kb": 64, 00:11:34.260 "superblock": false, 00:11:34.260 "method": "bdev_raid_create", 00:11:34.260 "req_id": 1 00:11:34.260 } 00:11:34.260 Got JSON-RPC error response 00:11:34.260 response: 00:11:34.260 { 00:11:34.260 "code": -17, 00:11:34.260 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:11:34.260 } 00:11:34.260 11:22:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@651 -- # es=1 00:11:34.260 11:22:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:11:34.260 11:22:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:11:34.260 11:22:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:11:34.260 11:22:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # jq -r '.[]' 00:11:34.260 11:22:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:34.518 11:22:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # raid_bdev= 00:11:34.518 11:22:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # '[' -n '' ']' 00:11:34.518 11:22:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@464 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:11:34.775 [2024-07-15 11:22:18.194229] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:11:34.775 [2024-07-15 11:22:18.194268] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:34.775 [2024-07-15 11:22:18.194290] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1b137a0 00:11:34.775 [2024-07-15 11:22:18.194302] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:34.775 [2024-07-15 11:22:18.195931] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:34.775 [2024-07-15 11:22:18.195957] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:11:34.775 [2024-07-15 11:22:18.196024] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:11:34.775 [2024-07-15 11:22:18.196051] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:11:34.775 pt1 00:11:34.775 11:22:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@467 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 2 00:11:34.775 11:22:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:11:34.775 11:22:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:11:34.775 11:22:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:11:34.775 11:22:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:34.775 11:22:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:11:34.775 11:22:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:34.775 11:22:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:34.775 11:22:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:34.775 11:22:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:34.775 11:22:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:34.775 11:22:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:35.033 11:22:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:35.033 "name": "raid_bdev1", 00:11:35.033 "uuid": "d547bd9c-c94e-4369-ad2a-cef7844502ec", 00:11:35.033 "strip_size_kb": 64, 00:11:35.033 "state": "configuring", 00:11:35.033 "raid_level": "raid0", 00:11:35.033 "superblock": true, 00:11:35.033 "num_base_bdevs": 2, 00:11:35.033 "num_base_bdevs_discovered": 1, 00:11:35.033 "num_base_bdevs_operational": 2, 00:11:35.033 "base_bdevs_list": [ 00:11:35.033 { 00:11:35.033 "name": "pt1", 00:11:35.033 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:35.033 "is_configured": true, 00:11:35.033 "data_offset": 2048, 00:11:35.033 "data_size": 63488 00:11:35.033 }, 00:11:35.033 { 00:11:35.033 "name": null, 00:11:35.033 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:35.033 "is_configured": false, 00:11:35.033 "data_offset": 2048, 00:11:35.033 "data_size": 63488 00:11:35.033 } 00:11:35.033 ] 00:11:35.033 }' 00:11:35.033 11:22:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:35.033 11:22:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:35.598 11:22:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@469 -- # '[' 2 -gt 2 ']' 00:11:35.598 11:22:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i = 1 )) 00:11:35.598 11:22:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:11:35.598 11:22:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:35.855 [2024-07-15 11:22:19.261073] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:35.855 [2024-07-15 11:22:19.261121] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:35.855 [2024-07-15 11:22:19.261139] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1cad820 00:11:35.855 [2024-07-15 11:22:19.261151] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:35.855 [2024-07-15 11:22:19.261489] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:35.855 [2024-07-15 11:22:19.261506] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:35.855 [2024-07-15 11:22:19.261566] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:11:35.855 [2024-07-15 11:22:19.261584] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:35.855 [2024-07-15 11:22:19.261679] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1b09ec0 00:11:35.855 [2024-07-15 11:22:19.261689] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:11:35.855 [2024-07-15 11:22:19.261854] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1b0c530 00:11:35.855 [2024-07-15 11:22:19.261980] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1b09ec0 00:11:35.855 [2024-07-15 11:22:19.261991] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x1b09ec0 00:11:35.855 [2024-07-15 11:22:19.262089] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:35.855 pt2 00:11:35.855 11:22:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:11:35.855 11:22:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:11:35.855 11:22:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@482 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:11:35.855 11:22:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:11:35.855 11:22:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:11:35.855 11:22:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:11:35.855 11:22:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:35.855 11:22:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:11:35.855 11:22:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:35.855 11:22:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:35.855 11:22:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:35.855 11:22:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:35.855 11:22:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:35.855 11:22:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:36.112 11:22:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:36.112 "name": "raid_bdev1", 00:11:36.112 "uuid": "d547bd9c-c94e-4369-ad2a-cef7844502ec", 00:11:36.112 "strip_size_kb": 64, 00:11:36.112 "state": "online", 00:11:36.112 "raid_level": "raid0", 00:11:36.112 "superblock": true, 00:11:36.112 "num_base_bdevs": 2, 00:11:36.112 "num_base_bdevs_discovered": 2, 00:11:36.112 "num_base_bdevs_operational": 2, 00:11:36.112 "base_bdevs_list": [ 00:11:36.112 { 00:11:36.112 "name": "pt1", 00:11:36.112 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:36.112 "is_configured": true, 00:11:36.112 "data_offset": 2048, 00:11:36.112 "data_size": 63488 00:11:36.112 }, 00:11:36.112 { 00:11:36.112 "name": "pt2", 00:11:36.112 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:36.112 "is_configured": true, 00:11:36.112 "data_offset": 2048, 00:11:36.112 "data_size": 63488 00:11:36.112 } 00:11:36.112 ] 00:11:36.112 }' 00:11:36.112 11:22:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:36.112 11:22:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:36.678 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_properties raid_bdev1 00:11:36.678 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:11:36.678 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:11:36.678 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:11:36.678 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:11:36.678 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # local name 00:11:36.678 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:11:36.678 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:11:36.936 [2024-07-15 11:22:20.344219] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:36.936 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:11:36.936 "name": "raid_bdev1", 00:11:36.936 "aliases": [ 00:11:36.936 "d547bd9c-c94e-4369-ad2a-cef7844502ec" 00:11:36.936 ], 00:11:36.936 "product_name": "Raid Volume", 00:11:36.936 "block_size": 512, 00:11:36.936 "num_blocks": 126976, 00:11:36.936 "uuid": "d547bd9c-c94e-4369-ad2a-cef7844502ec", 00:11:36.936 "assigned_rate_limits": { 00:11:36.936 "rw_ios_per_sec": 0, 00:11:36.936 "rw_mbytes_per_sec": 0, 00:11:36.936 "r_mbytes_per_sec": 0, 00:11:36.936 "w_mbytes_per_sec": 0 00:11:36.936 }, 00:11:36.936 "claimed": false, 00:11:36.936 "zoned": false, 00:11:36.936 "supported_io_types": { 00:11:36.936 "read": true, 00:11:36.936 "write": true, 00:11:36.936 "unmap": true, 00:11:36.936 "flush": true, 00:11:36.936 "reset": true, 00:11:36.936 "nvme_admin": false, 00:11:36.936 "nvme_io": false, 00:11:36.936 "nvme_io_md": false, 00:11:36.936 "write_zeroes": true, 00:11:36.936 "zcopy": false, 00:11:36.936 "get_zone_info": false, 00:11:36.936 "zone_management": false, 00:11:36.936 "zone_append": false, 00:11:36.936 "compare": false, 00:11:36.936 "compare_and_write": false, 00:11:36.936 "abort": false, 00:11:36.936 "seek_hole": false, 00:11:36.936 "seek_data": false, 00:11:36.936 "copy": false, 00:11:36.936 "nvme_iov_md": false 00:11:36.936 }, 00:11:36.936 "memory_domains": [ 00:11:36.936 { 00:11:36.936 "dma_device_id": "system", 00:11:36.936 "dma_device_type": 1 00:11:36.936 }, 00:11:36.936 { 00:11:36.936 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:36.936 "dma_device_type": 2 00:11:36.936 }, 00:11:36.936 { 00:11:36.936 "dma_device_id": "system", 00:11:36.936 "dma_device_type": 1 00:11:36.936 }, 00:11:36.936 { 00:11:36.936 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:36.936 "dma_device_type": 2 00:11:36.936 } 00:11:36.936 ], 00:11:36.936 "driver_specific": { 00:11:36.936 "raid": { 00:11:36.936 "uuid": "d547bd9c-c94e-4369-ad2a-cef7844502ec", 00:11:36.936 "strip_size_kb": 64, 00:11:36.936 "state": "online", 00:11:36.936 "raid_level": "raid0", 00:11:36.936 "superblock": true, 00:11:36.936 "num_base_bdevs": 2, 00:11:36.936 "num_base_bdevs_discovered": 2, 00:11:36.936 "num_base_bdevs_operational": 2, 00:11:36.936 "base_bdevs_list": [ 00:11:36.936 { 00:11:36.936 "name": "pt1", 00:11:36.936 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:36.936 "is_configured": true, 00:11:36.936 "data_offset": 2048, 00:11:36.936 "data_size": 63488 00:11:36.936 }, 00:11:36.936 { 00:11:36.936 "name": "pt2", 00:11:36.936 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:36.936 "is_configured": true, 00:11:36.936 "data_offset": 2048, 00:11:36.936 "data_size": 63488 00:11:36.936 } 00:11:36.936 ] 00:11:36.936 } 00:11:36.936 } 00:11:36.936 }' 00:11:36.936 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:36.936 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:11:36.936 pt2' 00:11:36.936 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:11:36.936 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:11:36.936 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:11:37.195 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:11:37.195 "name": "pt1", 00:11:37.195 "aliases": [ 00:11:37.195 "00000000-0000-0000-0000-000000000001" 00:11:37.195 ], 00:11:37.195 "product_name": "passthru", 00:11:37.195 "block_size": 512, 00:11:37.195 "num_blocks": 65536, 00:11:37.195 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:37.195 "assigned_rate_limits": { 00:11:37.195 "rw_ios_per_sec": 0, 00:11:37.195 "rw_mbytes_per_sec": 0, 00:11:37.195 "r_mbytes_per_sec": 0, 00:11:37.195 "w_mbytes_per_sec": 0 00:11:37.195 }, 00:11:37.195 "claimed": true, 00:11:37.195 "claim_type": "exclusive_write", 00:11:37.195 "zoned": false, 00:11:37.195 "supported_io_types": { 00:11:37.195 "read": true, 00:11:37.195 "write": true, 00:11:37.195 "unmap": true, 00:11:37.195 "flush": true, 00:11:37.195 "reset": true, 00:11:37.195 "nvme_admin": false, 00:11:37.195 "nvme_io": false, 00:11:37.195 "nvme_io_md": false, 00:11:37.195 "write_zeroes": true, 00:11:37.195 "zcopy": true, 00:11:37.195 "get_zone_info": false, 00:11:37.195 "zone_management": false, 00:11:37.195 "zone_append": false, 00:11:37.195 "compare": false, 00:11:37.195 "compare_and_write": false, 00:11:37.195 "abort": true, 00:11:37.195 "seek_hole": false, 00:11:37.195 "seek_data": false, 00:11:37.195 "copy": true, 00:11:37.195 "nvme_iov_md": false 00:11:37.195 }, 00:11:37.195 "memory_domains": [ 00:11:37.195 { 00:11:37.195 "dma_device_id": "system", 00:11:37.195 "dma_device_type": 1 00:11:37.195 }, 00:11:37.195 { 00:11:37.195 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:37.195 "dma_device_type": 2 00:11:37.195 } 00:11:37.195 ], 00:11:37.195 "driver_specific": { 00:11:37.195 "passthru": { 00:11:37.195 "name": "pt1", 00:11:37.195 "base_bdev_name": "malloc1" 00:11:37.195 } 00:11:37.195 } 00:11:37.195 }' 00:11:37.195 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:11:37.195 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:11:37.195 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:11:37.195 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:11:37.452 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:11:37.452 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:11:37.452 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:11:37.452 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:11:37.452 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:11:37.452 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:11:37.452 11:22:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:11:37.452 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:11:37.452 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:11:37.452 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:11:37.452 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:11:37.710 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:11:37.710 "name": "pt2", 00:11:37.710 "aliases": [ 00:11:37.710 "00000000-0000-0000-0000-000000000002" 00:11:37.710 ], 00:11:37.710 "product_name": "passthru", 00:11:37.710 "block_size": 512, 00:11:37.710 "num_blocks": 65536, 00:11:37.710 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:37.710 "assigned_rate_limits": { 00:11:37.710 "rw_ios_per_sec": 0, 00:11:37.710 "rw_mbytes_per_sec": 0, 00:11:37.710 "r_mbytes_per_sec": 0, 00:11:37.710 "w_mbytes_per_sec": 0 00:11:37.711 }, 00:11:37.711 "claimed": true, 00:11:37.711 "claim_type": "exclusive_write", 00:11:37.711 "zoned": false, 00:11:37.711 "supported_io_types": { 00:11:37.711 "read": true, 00:11:37.711 "write": true, 00:11:37.711 "unmap": true, 00:11:37.711 "flush": true, 00:11:37.711 "reset": true, 00:11:37.711 "nvme_admin": false, 00:11:37.711 "nvme_io": false, 00:11:37.711 "nvme_io_md": false, 00:11:37.711 "write_zeroes": true, 00:11:37.711 "zcopy": true, 00:11:37.711 "get_zone_info": false, 00:11:37.711 "zone_management": false, 00:11:37.711 "zone_append": false, 00:11:37.711 "compare": false, 00:11:37.711 "compare_and_write": false, 00:11:37.711 "abort": true, 00:11:37.711 "seek_hole": false, 00:11:37.711 "seek_data": false, 00:11:37.711 "copy": true, 00:11:37.711 "nvme_iov_md": false 00:11:37.711 }, 00:11:37.711 "memory_domains": [ 00:11:37.711 { 00:11:37.711 "dma_device_id": "system", 00:11:37.711 "dma_device_type": 1 00:11:37.711 }, 00:11:37.711 { 00:11:37.711 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:37.711 "dma_device_type": 2 00:11:37.711 } 00:11:37.711 ], 00:11:37.711 "driver_specific": { 00:11:37.711 "passthru": { 00:11:37.711 "name": "pt2", 00:11:37.711 "base_bdev_name": "malloc2" 00:11:37.711 } 00:11:37.711 } 00:11:37.711 }' 00:11:37.711 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:11:37.968 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:11:37.968 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:11:37.968 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:11:37.968 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:11:37.968 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:11:37.968 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:11:37.968 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:11:37.968 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:11:37.968 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:11:38.226 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:11:38.226 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:11:38.226 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # jq -r '.[] | .uuid' 00:11:38.226 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:11:38.484 [2024-07-15 11:22:21.856223] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:38.484 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # '[' d547bd9c-c94e-4369-ad2a-cef7844502ec '!=' d547bd9c-c94e-4369-ad2a-cef7844502ec ']' 00:11:38.484 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@490 -- # has_redundancy raid0 00:11:38.484 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:11:38.484 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@215 -- # return 1 00:11:38.484 11:22:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@562 -- # killprocess 864767 00:11:38.484 11:22:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@948 -- # '[' -z 864767 ']' 00:11:38.484 11:22:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@952 -- # kill -0 864767 00:11:38.484 11:22:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@953 -- # uname 00:11:38.484 11:22:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:11:38.484 11:22:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 864767 00:11:38.485 11:22:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:11:38.485 11:22:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:11:38.485 11:22:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 864767' 00:11:38.485 killing process with pid 864767 00:11:38.485 11:22:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@967 -- # kill 864767 00:11:38.485 [2024-07-15 11:22:21.931021] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:38.485 [2024-07-15 11:22:21.931074] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:38.485 [2024-07-15 11:22:21.931116] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:38.485 [2024-07-15 11:22:21.931128] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1b09ec0 name raid_bdev1, state offline 00:11:38.485 11:22:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # wait 864767 00:11:38.485 [2024-07-15 11:22:21.948153] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:38.761 11:22:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@564 -- # return 0 00:11:38.761 00:11:38.761 real 0m10.574s 00:11:38.761 user 0m18.797s 00:11:38.761 sys 0m2.020s 00:11:38.761 11:22:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:11:38.761 11:22:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.761 ************************************ 00:11:38.761 END TEST raid_superblock_test 00:11:38.761 ************************************ 00:11:38.761 11:22:22 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:11:38.761 11:22:22 bdev_raid -- bdev/bdev_raid.sh@870 -- # run_test raid_read_error_test raid_io_error_test raid0 2 read 00:11:38.761 11:22:22 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:11:38.761 11:22:22 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:11:38.761 11:22:22 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:38.761 ************************************ 00:11:38.761 START TEST raid_read_error_test 00:11:38.761 ************************************ 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1123 -- # raid_io_error_test raid0 2 read 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@788 -- # local raid_level=raid0 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@789 -- # local num_base_bdevs=2 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local error_io_type=read 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i = 1 )) 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev1 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev2 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local base_bdevs 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local raid_bdev_name=raid_bdev1 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local strip_size 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local create_arg 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local bdevperf_log 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local fail_per_s 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # '[' raid0 '!=' raid1 ']' 00:11:38.761 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@799 -- # strip_size=64 00:11:38.762 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # create_arg+=' -z 64' 00:11:38.762 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@805 -- # mktemp -p /raidtest 00:11:38.762 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@805 -- # bdevperf_log=/raidtest/tmp.dtS1aW9q7G 00:11:38.762 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@808 -- # raid_pid=866365 00:11:38.762 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # waitforlisten 866365 /var/tmp/spdk-raid.sock 00:11:38.762 11:22:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:11:38.762 11:22:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@829 -- # '[' -z 866365 ']' 00:11:38.762 11:22:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:11:38.762 11:22:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:11:38.762 11:22:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:11:38.762 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:11:38.762 11:22:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:11:38.762 11:22:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.762 [2024-07-15 11:22:22.328715] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:11:38.762 [2024-07-15 11:22:22.328787] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid866365 ] 00:11:39.019 [2024-07-15 11:22:22.462248] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:39.019 [2024-07-15 11:22:22.563705] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:11:39.277 [2024-07-15 11:22:22.630238] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:39.277 [2024-07-15 11:22:22.630279] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:39.840 11:22:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:11:39.840 11:22:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@862 -- # return 0 00:11:39.840 11:22:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:11:39.840 11:22:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:40.097 BaseBdev1_malloc 00:11:40.097 11:22:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev1_malloc 00:11:40.354 true 00:11:40.354 11:22:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:11:40.617 [2024-07-15 11:22:23.984647] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:11:40.617 [2024-07-15 11:22:23.984692] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:40.617 [2024-07-15 11:22:23.984711] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x186c0d0 00:11:40.617 [2024-07-15 11:22:23.984724] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:40.617 [2024-07-15 11:22:23.986447] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:40.617 [2024-07-15 11:22:23.986476] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:40.617 BaseBdev1 00:11:40.617 11:22:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:11:40.617 11:22:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:40.876 BaseBdev2_malloc 00:11:40.876 11:22:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev2_malloc 00:11:41.133 true 00:11:41.133 11:22:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:11:41.133 [2024-07-15 11:22:24.723319] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:11:41.133 [2024-07-15 11:22:24.723368] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:41.133 [2024-07-15 11:22:24.723389] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1870910 00:11:41.133 [2024-07-15 11:22:24.723402] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:41.133 [2024-07-15 11:22:24.724849] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:41.133 [2024-07-15 11:22:24.724879] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:41.391 BaseBdev2 00:11:41.391 11:22:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@819 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'BaseBdev1 BaseBdev2' -n raid_bdev1 -s 00:11:41.391 [2024-07-15 11:22:24.972032] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:41.391 [2024-07-15 11:22:24.973396] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:41.391 [2024-07-15 11:22:24.973602] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1872320 00:11:41.391 [2024-07-15 11:22:24.973615] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:11:41.391 [2024-07-15 11:22:24.973817] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1871270 00:11:41.391 [2024-07-15 11:22:24.973979] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1872320 00:11:41.391 [2024-07-15 11:22:24.973990] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x1872320 00:11:41.391 [2024-07-15 11:22:24.974105] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:41.648 11:22:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@820 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:11:41.648 11:22:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:11:41.648 11:22:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:11:41.648 11:22:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:11:41.648 11:22:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:41.648 11:22:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:11:41.648 11:22:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:41.648 11:22:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:41.648 11:22:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:41.648 11:22:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:41.648 11:22:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:41.648 11:22:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:41.648 11:22:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:41.648 "name": "raid_bdev1", 00:11:41.648 "uuid": "d5ba917c-51b5-48a6-ab6f-d61bc127795b", 00:11:41.648 "strip_size_kb": 64, 00:11:41.648 "state": "online", 00:11:41.648 "raid_level": "raid0", 00:11:41.648 "superblock": true, 00:11:41.648 "num_base_bdevs": 2, 00:11:41.648 "num_base_bdevs_discovered": 2, 00:11:41.648 "num_base_bdevs_operational": 2, 00:11:41.648 "base_bdevs_list": [ 00:11:41.648 { 00:11:41.648 "name": "BaseBdev1", 00:11:41.648 "uuid": "5eefc097-c37b-52da-8acb-6ec55e71e48f", 00:11:41.648 "is_configured": true, 00:11:41.648 "data_offset": 2048, 00:11:41.648 "data_size": 63488 00:11:41.648 }, 00:11:41.648 { 00:11:41.648 "name": "BaseBdev2", 00:11:41.648 "uuid": "132bdde5-2d44-55cc-9d72-bca255d04d2a", 00:11:41.648 "is_configured": true, 00:11:41.648 "data_offset": 2048, 00:11:41.648 "data_size": 63488 00:11:41.648 } 00:11:41.648 ] 00:11:41.648 }' 00:11:41.648 11:22:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:41.648 11:22:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:42.579 11:22:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@824 -- # sleep 1 00:11:42.579 11:22:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@823 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:11:42.579 [2024-07-15 11:22:25.934851] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x186d9b0 00:11:43.518 11:22:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@827 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:11:43.518 11:22:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # local expected_num_base_bdevs 00:11:43.518 11:22:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@830 -- # [[ raid0 = \r\a\i\d\1 ]] 00:11:43.518 11:22:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=2 00:11:43.518 11:22:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:11:43.518 11:22:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:11:43.518 11:22:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:11:43.518 11:22:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:11:43.518 11:22:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:43.518 11:22:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:11:43.518 11:22:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:43.518 11:22:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:43.518 11:22:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:43.518 11:22:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:43.518 11:22:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:43.518 11:22:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:43.842 11:22:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:43.842 "name": "raid_bdev1", 00:11:43.842 "uuid": "d5ba917c-51b5-48a6-ab6f-d61bc127795b", 00:11:43.842 "strip_size_kb": 64, 00:11:43.842 "state": "online", 00:11:43.842 "raid_level": "raid0", 00:11:43.842 "superblock": true, 00:11:43.842 "num_base_bdevs": 2, 00:11:43.842 "num_base_bdevs_discovered": 2, 00:11:43.842 "num_base_bdevs_operational": 2, 00:11:43.842 "base_bdevs_list": [ 00:11:43.842 { 00:11:43.842 "name": "BaseBdev1", 00:11:43.842 "uuid": "5eefc097-c37b-52da-8acb-6ec55e71e48f", 00:11:43.842 "is_configured": true, 00:11:43.842 "data_offset": 2048, 00:11:43.842 "data_size": 63488 00:11:43.842 }, 00:11:43.842 { 00:11:43.842 "name": "BaseBdev2", 00:11:43.842 "uuid": "132bdde5-2d44-55cc-9d72-bca255d04d2a", 00:11:43.842 "is_configured": true, 00:11:43.842 "data_offset": 2048, 00:11:43.842 "data_size": 63488 00:11:43.842 } 00:11:43.842 ] 00:11:43.842 }' 00:11:43.842 11:22:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:43.842 11:22:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:44.407 11:22:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:11:44.665 [2024-07-15 11:22:28.147801] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:44.665 [2024-07-15 11:22:28.147843] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:44.666 [2024-07-15 11:22:28.151055] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:44.666 [2024-07-15 11:22:28.151086] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:44.666 [2024-07-15 11:22:28.151115] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:44.666 [2024-07-15 11:22:28.151126] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1872320 name raid_bdev1, state offline 00:11:44.666 0 00:11:44.666 11:22:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # killprocess 866365 00:11:44.666 11:22:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@948 -- # '[' -z 866365 ']' 00:11:44.666 11:22:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@952 -- # kill -0 866365 00:11:44.666 11:22:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@953 -- # uname 00:11:44.666 11:22:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:11:44.666 11:22:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 866365 00:11:44.666 11:22:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:11:44.666 11:22:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:11:44.666 11:22:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 866365' 00:11:44.666 killing process with pid 866365 00:11:44.666 11:22:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@967 -- # kill 866365 00:11:44.666 [2024-07-15 11:22:28.226636] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:44.666 11:22:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # wait 866365 00:11:44.666 [2024-07-15 11:22:28.238835] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:44.925 11:22:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # grep -v Job /raidtest/tmp.dtS1aW9q7G 00:11:44.925 11:22:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # grep raid_bdev1 00:11:44.925 11:22:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # awk '{print $6}' 00:11:44.925 11:22:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # fail_per_s=0.45 00:11:44.925 11:22:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@844 -- # has_redundancy raid0 00:11:44.925 11:22:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:11:44.925 11:22:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@215 -- # return 1 00:11:44.925 11:22:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.45 != \0\.\0\0 ]] 00:11:44.925 00:11:44.925 real 0m6.233s 00:11:44.925 user 0m9.661s 00:11:44.925 sys 0m1.143s 00:11:44.925 11:22:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:11:44.925 11:22:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:44.925 ************************************ 00:11:44.925 END TEST raid_read_error_test 00:11:44.925 ************************************ 00:11:45.184 11:22:28 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:11:45.185 11:22:28 bdev_raid -- bdev/bdev_raid.sh@871 -- # run_test raid_write_error_test raid_io_error_test raid0 2 write 00:11:45.185 11:22:28 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:11:45.185 11:22:28 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:11:45.185 11:22:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:45.185 ************************************ 00:11:45.185 START TEST raid_write_error_test 00:11:45.185 ************************************ 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1123 -- # raid_io_error_test raid0 2 write 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@788 -- # local raid_level=raid0 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@789 -- # local num_base_bdevs=2 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local error_io_type=write 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i = 1 )) 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev1 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev2 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local base_bdevs 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local raid_bdev_name=raid_bdev1 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local strip_size 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local create_arg 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local bdevperf_log 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local fail_per_s 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # '[' raid0 '!=' raid1 ']' 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@799 -- # strip_size=64 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # create_arg+=' -z 64' 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@805 -- # mktemp -p /raidtest 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@805 -- # bdevperf_log=/raidtest/tmp.qobjhTobnb 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@808 -- # raid_pid=867334 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # waitforlisten 867334 /var/tmp/spdk-raid.sock 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@829 -- # '[' -z 867334 ']' 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:11:45.185 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:11:45.185 11:22:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.185 [2024-07-15 11:22:28.652030] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:11:45.185 [2024-07-15 11:22:28.652107] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid867334 ] 00:11:45.445 [2024-07-15 11:22:28.782846] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:45.445 [2024-07-15 11:22:28.885729] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:11:45.445 [2024-07-15 11:22:28.953958] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:45.445 [2024-07-15 11:22:28.954000] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:46.010 11:22:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:11:46.010 11:22:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@862 -- # return 0 00:11:46.010 11:22:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:11:46.010 11:22:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:46.267 BaseBdev1_malloc 00:11:46.267 11:22:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev1_malloc 00:11:46.525 true 00:11:46.526 11:22:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:11:46.784 [2024-07-15 11:22:30.228679] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:11:46.784 [2024-07-15 11:22:30.228726] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:46.784 [2024-07-15 11:22:30.228748] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x164a0d0 00:11:46.784 [2024-07-15 11:22:30.228761] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:46.784 [2024-07-15 11:22:30.230622] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:46.784 [2024-07-15 11:22:30.230653] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:46.784 BaseBdev1 00:11:46.784 11:22:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:11:46.784 11:22:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:47.042 BaseBdev2_malloc 00:11:47.042 11:22:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev2_malloc 00:11:47.300 true 00:11:47.300 11:22:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:11:47.558 [2024-07-15 11:22:30.903089] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:11:47.558 [2024-07-15 11:22:30.903138] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:47.558 [2024-07-15 11:22:30.903160] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x164e910 00:11:47.558 [2024-07-15 11:22:30.903173] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:47.558 [2024-07-15 11:22:30.904697] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:47.558 [2024-07-15 11:22:30.904725] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:47.558 BaseBdev2 00:11:47.558 11:22:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@819 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'BaseBdev1 BaseBdev2' -n raid_bdev1 -s 00:11:47.558 [2024-07-15 11:22:31.151769] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:47.818 [2024-07-15 11:22:31.153021] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:47.818 [2024-07-15 11:22:31.153220] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1650320 00:11:47.818 [2024-07-15 11:22:31.153233] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:11:47.818 [2024-07-15 11:22:31.153420] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x164f270 00:11:47.818 [2024-07-15 11:22:31.153563] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1650320 00:11:47.818 [2024-07-15 11:22:31.153573] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x1650320 00:11:47.818 [2024-07-15 11:22:31.153673] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:47.818 11:22:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@820 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:11:47.818 11:22:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:11:47.818 11:22:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:11:47.818 11:22:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:11:47.818 11:22:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:47.818 11:22:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:11:47.818 11:22:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:47.818 11:22:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:47.818 11:22:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:47.818 11:22:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:47.818 11:22:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:47.818 11:22:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:48.076 11:22:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:48.076 "name": "raid_bdev1", 00:11:48.076 "uuid": "fc6f8f0b-a2a6-4b30-89c3-981d92ab9338", 00:11:48.076 "strip_size_kb": 64, 00:11:48.076 "state": "online", 00:11:48.076 "raid_level": "raid0", 00:11:48.076 "superblock": true, 00:11:48.076 "num_base_bdevs": 2, 00:11:48.076 "num_base_bdevs_discovered": 2, 00:11:48.076 "num_base_bdevs_operational": 2, 00:11:48.076 "base_bdevs_list": [ 00:11:48.076 { 00:11:48.076 "name": "BaseBdev1", 00:11:48.076 "uuid": "7549f577-1bfc-55de-b817-4626b4f97587", 00:11:48.076 "is_configured": true, 00:11:48.076 "data_offset": 2048, 00:11:48.076 "data_size": 63488 00:11:48.076 }, 00:11:48.076 { 00:11:48.076 "name": "BaseBdev2", 00:11:48.076 "uuid": "bee26899-3a8d-5bc5-8142-38acdcce0928", 00:11:48.076 "is_configured": true, 00:11:48.076 "data_offset": 2048, 00:11:48.076 "data_size": 63488 00:11:48.076 } 00:11:48.076 ] 00:11:48.076 }' 00:11:48.076 11:22:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:48.076 11:22:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:48.643 11:22:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@823 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:11:48.643 11:22:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@824 -- # sleep 1 00:11:48.643 [2024-07-15 11:22:32.118627] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x164b9b0 00:11:49.578 11:22:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@827 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:11:49.836 11:22:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # local expected_num_base_bdevs 00:11:49.836 11:22:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@830 -- # [[ raid0 = \r\a\i\d\1 ]] 00:11:49.836 11:22:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=2 00:11:49.836 11:22:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:11:49.836 11:22:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:11:49.836 11:22:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:11:49.836 11:22:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:11:49.836 11:22:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:49.836 11:22:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:11:49.836 11:22:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:49.836 11:22:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:49.836 11:22:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:49.836 11:22:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:49.836 11:22:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:49.836 11:22:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:50.093 11:22:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:50.093 "name": "raid_bdev1", 00:11:50.093 "uuid": "fc6f8f0b-a2a6-4b30-89c3-981d92ab9338", 00:11:50.093 "strip_size_kb": 64, 00:11:50.093 "state": "online", 00:11:50.093 "raid_level": "raid0", 00:11:50.093 "superblock": true, 00:11:50.093 "num_base_bdevs": 2, 00:11:50.093 "num_base_bdevs_discovered": 2, 00:11:50.093 "num_base_bdevs_operational": 2, 00:11:50.093 "base_bdevs_list": [ 00:11:50.093 { 00:11:50.093 "name": "BaseBdev1", 00:11:50.093 "uuid": "7549f577-1bfc-55de-b817-4626b4f97587", 00:11:50.093 "is_configured": true, 00:11:50.093 "data_offset": 2048, 00:11:50.093 "data_size": 63488 00:11:50.093 }, 00:11:50.093 { 00:11:50.093 "name": "BaseBdev2", 00:11:50.093 "uuid": "bee26899-3a8d-5bc5-8142-38acdcce0928", 00:11:50.093 "is_configured": true, 00:11:50.093 "data_offset": 2048, 00:11:50.093 "data_size": 63488 00:11:50.093 } 00:11:50.093 ] 00:11:50.093 }' 00:11:50.093 11:22:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:50.093 11:22:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.657 11:22:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:11:50.657 [2024-07-15 11:22:34.250781] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:50.657 [2024-07-15 11:22:34.250821] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:50.915 [2024-07-15 11:22:34.254075] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:50.915 [2024-07-15 11:22:34.254112] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:50.915 [2024-07-15 11:22:34.254141] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:50.915 [2024-07-15 11:22:34.254153] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1650320 name raid_bdev1, state offline 00:11:50.915 0 00:11:50.915 11:22:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # killprocess 867334 00:11:50.915 11:22:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@948 -- # '[' -z 867334 ']' 00:11:50.915 11:22:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@952 -- # kill -0 867334 00:11:50.915 11:22:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@953 -- # uname 00:11:50.915 11:22:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:11:50.915 11:22:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 867334 00:11:50.915 11:22:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:11:50.915 11:22:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:11:50.915 11:22:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 867334' 00:11:50.915 killing process with pid 867334 00:11:50.915 11:22:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@967 -- # kill 867334 00:11:50.915 [2024-07-15 11:22:34.324870] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:50.915 11:22:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # wait 867334 00:11:50.915 [2024-07-15 11:22:34.337090] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:51.173 11:22:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # grep -v Job /raidtest/tmp.qobjhTobnb 00:11:51.173 11:22:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # grep raid_bdev1 00:11:51.173 11:22:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # awk '{print $6}' 00:11:51.173 11:22:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # fail_per_s=0.47 00:11:51.173 11:22:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@844 -- # has_redundancy raid0 00:11:51.173 11:22:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:11:51.173 11:22:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@215 -- # return 1 00:11:51.173 11:22:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.47 != \0\.\0\0 ]] 00:11:51.173 00:11:51.173 real 0m6.013s 00:11:51.173 user 0m9.320s 00:11:51.173 sys 0m1.058s 00:11:51.173 11:22:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:11:51.173 11:22:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.173 ************************************ 00:11:51.173 END TEST raid_write_error_test 00:11:51.173 ************************************ 00:11:51.173 11:22:34 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:11:51.173 11:22:34 bdev_raid -- bdev/bdev_raid.sh@866 -- # for level in raid0 concat raid1 00:11:51.173 11:22:34 bdev_raid -- bdev/bdev_raid.sh@867 -- # run_test raid_state_function_test raid_state_function_test concat 2 false 00:11:51.173 11:22:34 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:11:51.173 11:22:34 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:11:51.173 11:22:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:51.173 ************************************ 00:11:51.173 START TEST raid_state_function_test 00:11:51.173 ************************************ 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1123 -- # raid_state_function_test concat 2 false 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@220 -- # local raid_level=concat 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=2 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # local superblock=false 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # local strip_size 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # '[' concat '!=' raid1 ']' 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # strip_size=64 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@232 -- # strip_size_create_arg='-z 64' 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # '[' false = true ']' 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@240 -- # superblock_create_arg= 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # raid_pid=868269 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 868269' 00:11:51.173 Process raid pid: 868269 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@246 -- # waitforlisten 868269 /var/tmp/spdk-raid.sock 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@829 -- # '[' -z 868269 ']' 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:11:51.173 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:11:51.173 11:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.173 [2024-07-15 11:22:34.731431] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:11:51.173 [2024-07-15 11:22:34.731495] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:51.432 [2024-07-15 11:22:34.863462] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:51.432 [2024-07-15 11:22:34.969094] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:11:51.690 [2024-07-15 11:22:35.037022] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:51.690 [2024-07-15 11:22:35.037051] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:52.254 11:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:11:52.254 11:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@862 -- # return 0 00:11:52.254 11:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:11:52.511 [2024-07-15 11:22:35.887867] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:52.511 [2024-07-15 11:22:35.887914] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:52.511 [2024-07-15 11:22:35.887935] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:52.511 [2024-07-15 11:22:35.887948] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:52.511 11:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:11:52.511 11:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:11:52.511 11:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:11:52.511 11:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:11:52.511 11:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:52.511 11:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:11:52.511 11:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:52.511 11:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:52.511 11:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:52.511 11:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:52.511 11:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:52.511 11:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:52.769 11:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:52.769 "name": "Existed_Raid", 00:11:52.769 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:52.769 "strip_size_kb": 64, 00:11:52.769 "state": "configuring", 00:11:52.769 "raid_level": "concat", 00:11:52.769 "superblock": false, 00:11:52.769 "num_base_bdevs": 2, 00:11:52.769 "num_base_bdevs_discovered": 0, 00:11:52.769 "num_base_bdevs_operational": 2, 00:11:52.769 "base_bdevs_list": [ 00:11:52.769 { 00:11:52.769 "name": "BaseBdev1", 00:11:52.769 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:52.769 "is_configured": false, 00:11:52.769 "data_offset": 0, 00:11:52.769 "data_size": 0 00:11:52.769 }, 00:11:52.769 { 00:11:52.769 "name": "BaseBdev2", 00:11:52.769 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:52.769 "is_configured": false, 00:11:52.769 "data_offset": 0, 00:11:52.769 "data_size": 0 00:11:52.769 } 00:11:52.769 ] 00:11:52.769 }' 00:11:52.769 11:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:52.769 11:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.334 11:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:11:53.592 [2024-07-15 11:22:36.982635] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:53.592 [2024-07-15 11:22:36.982668] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x18fea80 name Existed_Raid, state configuring 00:11:53.592 11:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:11:53.850 [2024-07-15 11:22:37.223299] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:53.850 [2024-07-15 11:22:37.223337] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:53.850 [2024-07-15 11:22:37.223349] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:53.850 [2024-07-15 11:22:37.223370] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:53.850 11:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:11:54.108 [2024-07-15 11:22:37.477774] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:54.108 BaseBdev1 00:11:54.108 11:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:11:54.108 11:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:11:54.108 11:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:11:54.108 11:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:11:54.108 11:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:11:54.108 11:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:11:54.108 11:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:11:54.366 11:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:54.624 [ 00:11:54.624 { 00:11:54.624 "name": "BaseBdev1", 00:11:54.624 "aliases": [ 00:11:54.624 "1b4550bf-feaa-4c4e-b213-bd52b1077e95" 00:11:54.624 ], 00:11:54.624 "product_name": "Malloc disk", 00:11:54.624 "block_size": 512, 00:11:54.624 "num_blocks": 65536, 00:11:54.624 "uuid": "1b4550bf-feaa-4c4e-b213-bd52b1077e95", 00:11:54.624 "assigned_rate_limits": { 00:11:54.624 "rw_ios_per_sec": 0, 00:11:54.624 "rw_mbytes_per_sec": 0, 00:11:54.624 "r_mbytes_per_sec": 0, 00:11:54.624 "w_mbytes_per_sec": 0 00:11:54.624 }, 00:11:54.624 "claimed": true, 00:11:54.624 "claim_type": "exclusive_write", 00:11:54.624 "zoned": false, 00:11:54.624 "supported_io_types": { 00:11:54.624 "read": true, 00:11:54.624 "write": true, 00:11:54.624 "unmap": true, 00:11:54.624 "flush": true, 00:11:54.624 "reset": true, 00:11:54.624 "nvme_admin": false, 00:11:54.624 "nvme_io": false, 00:11:54.624 "nvme_io_md": false, 00:11:54.624 "write_zeroes": true, 00:11:54.624 "zcopy": true, 00:11:54.624 "get_zone_info": false, 00:11:54.624 "zone_management": false, 00:11:54.624 "zone_append": false, 00:11:54.624 "compare": false, 00:11:54.624 "compare_and_write": false, 00:11:54.624 "abort": true, 00:11:54.624 "seek_hole": false, 00:11:54.624 "seek_data": false, 00:11:54.624 "copy": true, 00:11:54.624 "nvme_iov_md": false 00:11:54.624 }, 00:11:54.624 "memory_domains": [ 00:11:54.624 { 00:11:54.624 "dma_device_id": "system", 00:11:54.624 "dma_device_type": 1 00:11:54.624 }, 00:11:54.624 { 00:11:54.624 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:54.624 "dma_device_type": 2 00:11:54.624 } 00:11:54.624 ], 00:11:54.624 "driver_specific": {} 00:11:54.624 } 00:11:54.624 ] 00:11:54.624 11:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:11:54.624 11:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:11:54.624 11:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:11:54.624 11:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:11:54.624 11:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:11:54.624 11:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:54.624 11:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:11:54.624 11:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:54.624 11:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:54.625 11:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:54.625 11:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:54.625 11:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:54.625 11:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:54.882 11:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:54.882 "name": "Existed_Raid", 00:11:54.882 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:54.882 "strip_size_kb": 64, 00:11:54.882 "state": "configuring", 00:11:54.882 "raid_level": "concat", 00:11:54.882 "superblock": false, 00:11:54.882 "num_base_bdevs": 2, 00:11:54.882 "num_base_bdevs_discovered": 1, 00:11:54.882 "num_base_bdevs_operational": 2, 00:11:54.882 "base_bdevs_list": [ 00:11:54.882 { 00:11:54.882 "name": "BaseBdev1", 00:11:54.882 "uuid": "1b4550bf-feaa-4c4e-b213-bd52b1077e95", 00:11:54.882 "is_configured": true, 00:11:54.882 "data_offset": 0, 00:11:54.882 "data_size": 65536 00:11:54.882 }, 00:11:54.882 { 00:11:54.882 "name": "BaseBdev2", 00:11:54.882 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:54.882 "is_configured": false, 00:11:54.882 "data_offset": 0, 00:11:54.882 "data_size": 0 00:11:54.882 } 00:11:54.882 ] 00:11:54.882 }' 00:11:54.882 11:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:54.882 11:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.448 11:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:11:55.706 [2024-07-15 11:22:39.049939] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:55.706 [2024-07-15 11:22:39.049976] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x18fe350 name Existed_Raid, state configuring 00:11:55.706 11:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:11:55.706 [2024-07-15 11:22:39.294600] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:55.706 [2024-07-15 11:22:39.296079] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:55.706 [2024-07-15 11:22:39.296110] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:55.964 11:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:11:55.964 11:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:11:55.964 11:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:11:55.964 11:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:11:55.964 11:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:11:55.964 11:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:11:55.964 11:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:55.964 11:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:11:55.964 11:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:55.964 11:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:55.964 11:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:55.964 11:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:55.964 11:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:55.964 11:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:56.223 11:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:56.223 "name": "Existed_Raid", 00:11:56.223 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:56.223 "strip_size_kb": 64, 00:11:56.223 "state": "configuring", 00:11:56.223 "raid_level": "concat", 00:11:56.223 "superblock": false, 00:11:56.223 "num_base_bdevs": 2, 00:11:56.223 "num_base_bdevs_discovered": 1, 00:11:56.223 "num_base_bdevs_operational": 2, 00:11:56.223 "base_bdevs_list": [ 00:11:56.223 { 00:11:56.223 "name": "BaseBdev1", 00:11:56.223 "uuid": "1b4550bf-feaa-4c4e-b213-bd52b1077e95", 00:11:56.223 "is_configured": true, 00:11:56.223 "data_offset": 0, 00:11:56.223 "data_size": 65536 00:11:56.223 }, 00:11:56.223 { 00:11:56.223 "name": "BaseBdev2", 00:11:56.223 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:56.223 "is_configured": false, 00:11:56.223 "data_offset": 0, 00:11:56.223 "data_size": 0 00:11:56.223 } 00:11:56.223 ] 00:11:56.223 }' 00:11:56.223 11:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:56.223 11:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:56.790 11:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:11:57.048 [2024-07-15 11:22:40.394018] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:57.048 [2024-07-15 11:22:40.394056] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x18ff000 00:11:57.048 [2024-07-15 11:22:40.394065] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:11:57.048 [2024-07-15 11:22:40.394258] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x18190c0 00:11:57.048 [2024-07-15 11:22:40.394378] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x18ff000 00:11:57.048 [2024-07-15 11:22:40.394388] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x18ff000 00:11:57.048 [2024-07-15 11:22:40.394551] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:57.048 BaseBdev2 00:11:57.048 11:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:11:57.048 11:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:11:57.048 11:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:11:57.048 11:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:11:57.048 11:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:11:57.048 11:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:11:57.048 11:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:11:57.307 11:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:57.307 [ 00:11:57.307 { 00:11:57.307 "name": "BaseBdev2", 00:11:57.307 "aliases": [ 00:11:57.307 "fe3b9e2e-65e1-4bfe-8816-d6aaf74cce47" 00:11:57.307 ], 00:11:57.307 "product_name": "Malloc disk", 00:11:57.307 "block_size": 512, 00:11:57.307 "num_blocks": 65536, 00:11:57.307 "uuid": "fe3b9e2e-65e1-4bfe-8816-d6aaf74cce47", 00:11:57.307 "assigned_rate_limits": { 00:11:57.307 "rw_ios_per_sec": 0, 00:11:57.307 "rw_mbytes_per_sec": 0, 00:11:57.307 "r_mbytes_per_sec": 0, 00:11:57.307 "w_mbytes_per_sec": 0 00:11:57.307 }, 00:11:57.307 "claimed": true, 00:11:57.307 "claim_type": "exclusive_write", 00:11:57.307 "zoned": false, 00:11:57.307 "supported_io_types": { 00:11:57.307 "read": true, 00:11:57.307 "write": true, 00:11:57.307 "unmap": true, 00:11:57.307 "flush": true, 00:11:57.307 "reset": true, 00:11:57.307 "nvme_admin": false, 00:11:57.307 "nvme_io": false, 00:11:57.307 "nvme_io_md": false, 00:11:57.307 "write_zeroes": true, 00:11:57.307 "zcopy": true, 00:11:57.307 "get_zone_info": false, 00:11:57.307 "zone_management": false, 00:11:57.307 "zone_append": false, 00:11:57.307 "compare": false, 00:11:57.307 "compare_and_write": false, 00:11:57.307 "abort": true, 00:11:57.307 "seek_hole": false, 00:11:57.307 "seek_data": false, 00:11:57.307 "copy": true, 00:11:57.307 "nvme_iov_md": false 00:11:57.307 }, 00:11:57.307 "memory_domains": [ 00:11:57.307 { 00:11:57.307 "dma_device_id": "system", 00:11:57.307 "dma_device_type": 1 00:11:57.307 }, 00:11:57.307 { 00:11:57.307 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:57.307 "dma_device_type": 2 00:11:57.307 } 00:11:57.307 ], 00:11:57.307 "driver_specific": {} 00:11:57.307 } 00:11:57.307 ] 00:11:57.567 11:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:11:57.567 11:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:11:57.567 11:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:11:57.567 11:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:11:57.567 11:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:11:57.567 11:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:11:57.567 11:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:11:57.567 11:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:57.567 11:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:11:57.567 11:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:57.567 11:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:57.567 11:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:57.567 11:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:57.567 11:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:57.567 11:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:57.567 11:22:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:11:57.567 "name": "Existed_Raid", 00:11:57.567 "uuid": "d5c71e57-ba20-4675-ba67-6f230fa4a5dc", 00:11:57.567 "strip_size_kb": 64, 00:11:57.567 "state": "online", 00:11:57.567 "raid_level": "concat", 00:11:57.567 "superblock": false, 00:11:57.567 "num_base_bdevs": 2, 00:11:57.567 "num_base_bdevs_discovered": 2, 00:11:57.567 "num_base_bdevs_operational": 2, 00:11:57.567 "base_bdevs_list": [ 00:11:57.567 { 00:11:57.567 "name": "BaseBdev1", 00:11:57.567 "uuid": "1b4550bf-feaa-4c4e-b213-bd52b1077e95", 00:11:57.567 "is_configured": true, 00:11:57.567 "data_offset": 0, 00:11:57.567 "data_size": 65536 00:11:57.567 }, 00:11:57.567 { 00:11:57.567 "name": "BaseBdev2", 00:11:57.567 "uuid": "fe3b9e2e-65e1-4bfe-8816-d6aaf74cce47", 00:11:57.567 "is_configured": true, 00:11:57.567 "data_offset": 0, 00:11:57.567 "data_size": 65536 00:11:57.567 } 00:11:57.567 ] 00:11:57.567 }' 00:11:57.567 11:22:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:11:57.567 11:22:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:58.134 11:22:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:11:58.134 11:22:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:11:58.134 11:22:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:11:58.134 11:22:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:11:58.134 11:22:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:11:58.134 11:22:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # local name 00:11:58.134 11:22:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:11:58.134 11:22:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:11:58.392 [2024-07-15 11:22:41.934385] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:58.392 11:22:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:11:58.392 "name": "Existed_Raid", 00:11:58.392 "aliases": [ 00:11:58.392 "d5c71e57-ba20-4675-ba67-6f230fa4a5dc" 00:11:58.392 ], 00:11:58.392 "product_name": "Raid Volume", 00:11:58.392 "block_size": 512, 00:11:58.392 "num_blocks": 131072, 00:11:58.392 "uuid": "d5c71e57-ba20-4675-ba67-6f230fa4a5dc", 00:11:58.392 "assigned_rate_limits": { 00:11:58.392 "rw_ios_per_sec": 0, 00:11:58.392 "rw_mbytes_per_sec": 0, 00:11:58.392 "r_mbytes_per_sec": 0, 00:11:58.392 "w_mbytes_per_sec": 0 00:11:58.392 }, 00:11:58.392 "claimed": false, 00:11:58.392 "zoned": false, 00:11:58.392 "supported_io_types": { 00:11:58.392 "read": true, 00:11:58.392 "write": true, 00:11:58.392 "unmap": true, 00:11:58.392 "flush": true, 00:11:58.392 "reset": true, 00:11:58.392 "nvme_admin": false, 00:11:58.392 "nvme_io": false, 00:11:58.392 "nvme_io_md": false, 00:11:58.392 "write_zeroes": true, 00:11:58.392 "zcopy": false, 00:11:58.392 "get_zone_info": false, 00:11:58.392 "zone_management": false, 00:11:58.392 "zone_append": false, 00:11:58.392 "compare": false, 00:11:58.392 "compare_and_write": false, 00:11:58.392 "abort": false, 00:11:58.392 "seek_hole": false, 00:11:58.392 "seek_data": false, 00:11:58.392 "copy": false, 00:11:58.392 "nvme_iov_md": false 00:11:58.392 }, 00:11:58.392 "memory_domains": [ 00:11:58.392 { 00:11:58.392 "dma_device_id": "system", 00:11:58.392 "dma_device_type": 1 00:11:58.392 }, 00:11:58.392 { 00:11:58.392 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:58.392 "dma_device_type": 2 00:11:58.392 }, 00:11:58.392 { 00:11:58.392 "dma_device_id": "system", 00:11:58.392 "dma_device_type": 1 00:11:58.392 }, 00:11:58.392 { 00:11:58.392 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:58.392 "dma_device_type": 2 00:11:58.392 } 00:11:58.392 ], 00:11:58.392 "driver_specific": { 00:11:58.392 "raid": { 00:11:58.392 "uuid": "d5c71e57-ba20-4675-ba67-6f230fa4a5dc", 00:11:58.392 "strip_size_kb": 64, 00:11:58.392 "state": "online", 00:11:58.392 "raid_level": "concat", 00:11:58.392 "superblock": false, 00:11:58.392 "num_base_bdevs": 2, 00:11:58.392 "num_base_bdevs_discovered": 2, 00:11:58.392 "num_base_bdevs_operational": 2, 00:11:58.392 "base_bdevs_list": [ 00:11:58.392 { 00:11:58.392 "name": "BaseBdev1", 00:11:58.392 "uuid": "1b4550bf-feaa-4c4e-b213-bd52b1077e95", 00:11:58.392 "is_configured": true, 00:11:58.392 "data_offset": 0, 00:11:58.392 "data_size": 65536 00:11:58.392 }, 00:11:58.392 { 00:11:58.392 "name": "BaseBdev2", 00:11:58.392 "uuid": "fe3b9e2e-65e1-4bfe-8816-d6aaf74cce47", 00:11:58.392 "is_configured": true, 00:11:58.392 "data_offset": 0, 00:11:58.392 "data_size": 65536 00:11:58.392 } 00:11:58.392 ] 00:11:58.392 } 00:11:58.392 } 00:11:58.392 }' 00:11:58.392 11:22:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:58.651 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:11:58.651 BaseBdev2' 00:11:58.651 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:11:58.651 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:11:58.651 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:11:58.651 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:11:58.651 "name": "BaseBdev1", 00:11:58.651 "aliases": [ 00:11:58.651 "1b4550bf-feaa-4c4e-b213-bd52b1077e95" 00:11:58.651 ], 00:11:58.651 "product_name": "Malloc disk", 00:11:58.651 "block_size": 512, 00:11:58.651 "num_blocks": 65536, 00:11:58.651 "uuid": "1b4550bf-feaa-4c4e-b213-bd52b1077e95", 00:11:58.651 "assigned_rate_limits": { 00:11:58.651 "rw_ios_per_sec": 0, 00:11:58.651 "rw_mbytes_per_sec": 0, 00:11:58.651 "r_mbytes_per_sec": 0, 00:11:58.651 "w_mbytes_per_sec": 0 00:11:58.651 }, 00:11:58.651 "claimed": true, 00:11:58.651 "claim_type": "exclusive_write", 00:11:58.651 "zoned": false, 00:11:58.651 "supported_io_types": { 00:11:58.651 "read": true, 00:11:58.651 "write": true, 00:11:58.651 "unmap": true, 00:11:58.651 "flush": true, 00:11:58.651 "reset": true, 00:11:58.651 "nvme_admin": false, 00:11:58.651 "nvme_io": false, 00:11:58.651 "nvme_io_md": false, 00:11:58.651 "write_zeroes": true, 00:11:58.651 "zcopy": true, 00:11:58.651 "get_zone_info": false, 00:11:58.651 "zone_management": false, 00:11:58.651 "zone_append": false, 00:11:58.651 "compare": false, 00:11:58.651 "compare_and_write": false, 00:11:58.651 "abort": true, 00:11:58.651 "seek_hole": false, 00:11:58.651 "seek_data": false, 00:11:58.651 "copy": true, 00:11:58.651 "nvme_iov_md": false 00:11:58.651 }, 00:11:58.651 "memory_domains": [ 00:11:58.651 { 00:11:58.651 "dma_device_id": "system", 00:11:58.651 "dma_device_type": 1 00:11:58.651 }, 00:11:58.651 { 00:11:58.651 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:58.651 "dma_device_type": 2 00:11:58.651 } 00:11:58.651 ], 00:11:58.651 "driver_specific": {} 00:11:58.651 }' 00:11:58.651 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:11:58.651 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:11:58.651 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:11:58.651 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:11:58.910 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:11:58.910 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:11:58.910 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:11:58.910 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:11:58.910 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:11:58.910 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:11:58.910 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:11:59.169 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:11:59.169 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:11:59.169 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:11:59.169 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:11:59.427 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:11:59.427 "name": "BaseBdev2", 00:11:59.427 "aliases": [ 00:11:59.427 "fe3b9e2e-65e1-4bfe-8816-d6aaf74cce47" 00:11:59.427 ], 00:11:59.427 "product_name": "Malloc disk", 00:11:59.427 "block_size": 512, 00:11:59.427 "num_blocks": 65536, 00:11:59.427 "uuid": "fe3b9e2e-65e1-4bfe-8816-d6aaf74cce47", 00:11:59.427 "assigned_rate_limits": { 00:11:59.427 "rw_ios_per_sec": 0, 00:11:59.427 "rw_mbytes_per_sec": 0, 00:11:59.427 "r_mbytes_per_sec": 0, 00:11:59.427 "w_mbytes_per_sec": 0 00:11:59.427 }, 00:11:59.427 "claimed": true, 00:11:59.427 "claim_type": "exclusive_write", 00:11:59.427 "zoned": false, 00:11:59.427 "supported_io_types": { 00:11:59.427 "read": true, 00:11:59.427 "write": true, 00:11:59.427 "unmap": true, 00:11:59.427 "flush": true, 00:11:59.427 "reset": true, 00:11:59.427 "nvme_admin": false, 00:11:59.427 "nvme_io": false, 00:11:59.427 "nvme_io_md": false, 00:11:59.427 "write_zeroes": true, 00:11:59.427 "zcopy": true, 00:11:59.427 "get_zone_info": false, 00:11:59.428 "zone_management": false, 00:11:59.428 "zone_append": false, 00:11:59.428 "compare": false, 00:11:59.428 "compare_and_write": false, 00:11:59.428 "abort": true, 00:11:59.428 "seek_hole": false, 00:11:59.428 "seek_data": false, 00:11:59.428 "copy": true, 00:11:59.428 "nvme_iov_md": false 00:11:59.428 }, 00:11:59.428 "memory_domains": [ 00:11:59.428 { 00:11:59.428 "dma_device_id": "system", 00:11:59.428 "dma_device_type": 1 00:11:59.428 }, 00:11:59.428 { 00:11:59.428 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:59.428 "dma_device_type": 2 00:11:59.428 } 00:11:59.428 ], 00:11:59.428 "driver_specific": {} 00:11:59.428 }' 00:11:59.428 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:11:59.428 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:11:59.428 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:11:59.428 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:11:59.428 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:11:59.428 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:11:59.428 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:11:59.428 11:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:11:59.686 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:11:59.686 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:11:59.686 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:11:59.686 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:11:59.686 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:11:59.944 [2024-07-15 11:22:43.341918] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:59.944 [2024-07-15 11:22:43.341954] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:59.944 [2024-07-15 11:22:43.341994] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:59.944 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@275 -- # local expected_state 00:11:59.944 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # has_redundancy concat 00:11:59.944 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:11:59.944 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # return 1 00:11:59.944 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@277 -- # expected_state=offline 00:11:59.944 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:11:59.944 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:11:59.944 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=offline 00:11:59.944 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:11:59.944 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:11:59.944 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:11:59.944 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:11:59.944 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:11:59.944 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:11:59.944 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:11:59.944 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:11:59.944 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:00.202 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:00.202 "name": "Existed_Raid", 00:12:00.202 "uuid": "d5c71e57-ba20-4675-ba67-6f230fa4a5dc", 00:12:00.202 "strip_size_kb": 64, 00:12:00.202 "state": "offline", 00:12:00.202 "raid_level": "concat", 00:12:00.202 "superblock": false, 00:12:00.202 "num_base_bdevs": 2, 00:12:00.202 "num_base_bdevs_discovered": 1, 00:12:00.202 "num_base_bdevs_operational": 1, 00:12:00.202 "base_bdevs_list": [ 00:12:00.202 { 00:12:00.202 "name": null, 00:12:00.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:00.202 "is_configured": false, 00:12:00.202 "data_offset": 0, 00:12:00.202 "data_size": 65536 00:12:00.202 }, 00:12:00.202 { 00:12:00.202 "name": "BaseBdev2", 00:12:00.202 "uuid": "fe3b9e2e-65e1-4bfe-8816-d6aaf74cce47", 00:12:00.202 "is_configured": true, 00:12:00.202 "data_offset": 0, 00:12:00.202 "data_size": 65536 00:12:00.202 } 00:12:00.202 ] 00:12:00.202 }' 00:12:00.202 11:22:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:00.202 11:22:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:01.176 11:22:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:12:01.176 11:22:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:12:01.176 11:22:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:01.176 11:22:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:12:01.176 11:22:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:12:01.176 11:22:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:12:01.176 11:22:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:12:01.435 [2024-07-15 11:22:44.959484] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:01.435 [2024-07-15 11:22:44.959540] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x18ff000 name Existed_Raid, state offline 00:12:01.435 11:22:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:12:01.435 11:22:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:12:01.435 11:22:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:01.435 11:22:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:12:01.694 11:22:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:12:01.694 11:22:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:12:01.694 11:22:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # '[' 2 -gt 2 ']' 00:12:01.694 11:22:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@341 -- # killprocess 868269 00:12:01.694 11:22:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@948 -- # '[' -z 868269 ']' 00:12:01.694 11:22:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@952 -- # kill -0 868269 00:12:01.694 11:22:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@953 -- # uname 00:12:01.694 11:22:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:12:01.694 11:22:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 868269 00:12:01.694 11:22:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:12:01.694 11:22:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:12:01.694 11:22:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 868269' 00:12:01.694 killing process with pid 868269 00:12:01.694 11:22:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@967 -- # kill 868269 00:12:01.694 [2024-07-15 11:22:45.277788] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:01.694 11:22:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # wait 868269 00:12:01.694 [2024-07-15 11:22:45.278780] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:01.980 11:22:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@343 -- # return 0 00:12:01.980 00:12:01.980 real 0m10.845s 00:12:01.980 user 0m19.309s 00:12:01.980 sys 0m1.971s 00:12:01.980 11:22:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:12:01.980 11:22:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:01.980 ************************************ 00:12:01.980 END TEST raid_state_function_test 00:12:01.980 ************************************ 00:12:01.980 11:22:45 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:12:01.980 11:22:45 bdev_raid -- bdev/bdev_raid.sh@868 -- # run_test raid_state_function_test_sb raid_state_function_test concat 2 true 00:12:01.980 11:22:45 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:12:01.980 11:22:45 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:12:01.980 11:22:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:02.239 ************************************ 00:12:02.239 START TEST raid_state_function_test_sb 00:12:02.239 ************************************ 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1123 -- # raid_state_function_test concat 2 true 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@220 -- # local raid_level=concat 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=2 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # local superblock=true 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # local strip_size 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # '[' concat '!=' raid1 ']' 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # strip_size=64 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@232 -- # strip_size_create_arg='-z 64' 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # '[' true = true ']' 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@238 -- # superblock_create_arg=-s 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # raid_pid=869942 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 869942' 00:12:02.239 Process raid pid: 869942 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@246 -- # waitforlisten 869942 /var/tmp/spdk-raid.sock 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@829 -- # '[' -z 869942 ']' 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@834 -- # local max_retries=100 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:12:02.239 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # xtrace_disable 00:12:02.239 11:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.239 [2024-07-15 11:22:45.635107] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:12:02.239 [2024-07-15 11:22:45.635169] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:02.239 [2024-07-15 11:22:45.765227] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:02.498 [2024-07-15 11:22:45.870360] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:12:02.498 [2024-07-15 11:22:45.926376] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:02.498 [2024-07-15 11:22:45.926403] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:02.756 11:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:12:02.756 11:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@862 -- # return 0 00:12:02.756 11:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r concat -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:12:02.756 [2024-07-15 11:22:46.324519] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:12:02.756 [2024-07-15 11:22:46.324558] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:12:02.756 [2024-07-15 11:22:46.324569] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:12:02.756 [2024-07-15 11:22:46.324581] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:12:02.756 11:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:12:02.756 11:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:12:02.756 11:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:12:02.756 11:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:12:02.756 11:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:12:02.756 11:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:12:02.756 11:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:02.756 11:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:02.756 11:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:02.756 11:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:02.756 11:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:02.756 11:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:03.014 11:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:03.014 "name": "Existed_Raid", 00:12:03.014 "uuid": "79ce73fd-362c-44ea-be35-fb5d1f75e0f2", 00:12:03.014 "strip_size_kb": 64, 00:12:03.014 "state": "configuring", 00:12:03.014 "raid_level": "concat", 00:12:03.014 "superblock": true, 00:12:03.014 "num_base_bdevs": 2, 00:12:03.014 "num_base_bdevs_discovered": 0, 00:12:03.014 "num_base_bdevs_operational": 2, 00:12:03.014 "base_bdevs_list": [ 00:12:03.014 { 00:12:03.014 "name": "BaseBdev1", 00:12:03.014 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:03.014 "is_configured": false, 00:12:03.014 "data_offset": 0, 00:12:03.014 "data_size": 0 00:12:03.014 }, 00:12:03.014 { 00:12:03.014 "name": "BaseBdev2", 00:12:03.014 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:03.014 "is_configured": false, 00:12:03.014 "data_offset": 0, 00:12:03.014 "data_size": 0 00:12:03.014 } 00:12:03.014 ] 00:12:03.014 }' 00:12:03.014 11:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:03.014 11:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.578 11:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:12:03.836 [2024-07-15 11:22:47.391209] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:12:03.836 [2024-07-15 11:22:47.391235] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1c9fa80 name Existed_Raid, state configuring 00:12:03.836 11:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r concat -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:12:04.094 [2024-07-15 11:22:47.635884] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:12:04.094 [2024-07-15 11:22:47.635918] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:12:04.094 [2024-07-15 11:22:47.635933] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:12:04.094 [2024-07-15 11:22:47.635945] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:12:04.094 11:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:12:04.351 [2024-07-15 11:22:47.894459] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:04.351 BaseBdev1 00:12:04.351 11:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:12:04.351 11:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:12:04.351 11:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:12:04.351 11:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:12:04.351 11:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:12:04.351 11:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:12:04.351 11:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:12:04.609 11:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:12:04.867 [ 00:12:04.867 { 00:12:04.867 "name": "BaseBdev1", 00:12:04.867 "aliases": [ 00:12:04.867 "1c002ed4-820d-4b70-933c-bc810aee61f5" 00:12:04.867 ], 00:12:04.867 "product_name": "Malloc disk", 00:12:04.867 "block_size": 512, 00:12:04.867 "num_blocks": 65536, 00:12:04.867 "uuid": "1c002ed4-820d-4b70-933c-bc810aee61f5", 00:12:04.867 "assigned_rate_limits": { 00:12:04.867 "rw_ios_per_sec": 0, 00:12:04.867 "rw_mbytes_per_sec": 0, 00:12:04.867 "r_mbytes_per_sec": 0, 00:12:04.867 "w_mbytes_per_sec": 0 00:12:04.867 }, 00:12:04.867 "claimed": true, 00:12:04.867 "claim_type": "exclusive_write", 00:12:04.867 "zoned": false, 00:12:04.867 "supported_io_types": { 00:12:04.867 "read": true, 00:12:04.867 "write": true, 00:12:04.867 "unmap": true, 00:12:04.867 "flush": true, 00:12:04.867 "reset": true, 00:12:04.867 "nvme_admin": false, 00:12:04.867 "nvme_io": false, 00:12:04.867 "nvme_io_md": false, 00:12:04.867 "write_zeroes": true, 00:12:04.867 "zcopy": true, 00:12:04.867 "get_zone_info": false, 00:12:04.867 "zone_management": false, 00:12:04.867 "zone_append": false, 00:12:04.867 "compare": false, 00:12:04.867 "compare_and_write": false, 00:12:04.867 "abort": true, 00:12:04.867 "seek_hole": false, 00:12:04.867 "seek_data": false, 00:12:04.867 "copy": true, 00:12:04.867 "nvme_iov_md": false 00:12:04.867 }, 00:12:04.867 "memory_domains": [ 00:12:04.867 { 00:12:04.867 "dma_device_id": "system", 00:12:04.867 "dma_device_type": 1 00:12:04.867 }, 00:12:04.867 { 00:12:04.867 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:04.867 "dma_device_type": 2 00:12:04.867 } 00:12:04.867 ], 00:12:04.867 "driver_specific": {} 00:12:04.867 } 00:12:04.867 ] 00:12:04.867 11:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:12:04.867 11:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:12:04.867 11:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:12:04.867 11:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:12:04.867 11:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:12:04.868 11:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:12:04.868 11:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:12:04.868 11:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:04.868 11:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:04.868 11:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:04.868 11:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:04.868 11:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:04.868 11:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:05.126 11:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:05.126 "name": "Existed_Raid", 00:12:05.126 "uuid": "1c82c8c4-b805-4c35-9277-fc6398bdc903", 00:12:05.126 "strip_size_kb": 64, 00:12:05.126 "state": "configuring", 00:12:05.126 "raid_level": "concat", 00:12:05.126 "superblock": true, 00:12:05.126 "num_base_bdevs": 2, 00:12:05.126 "num_base_bdevs_discovered": 1, 00:12:05.126 "num_base_bdevs_operational": 2, 00:12:05.126 "base_bdevs_list": [ 00:12:05.126 { 00:12:05.126 "name": "BaseBdev1", 00:12:05.126 "uuid": "1c002ed4-820d-4b70-933c-bc810aee61f5", 00:12:05.126 "is_configured": true, 00:12:05.126 "data_offset": 2048, 00:12:05.126 "data_size": 63488 00:12:05.126 }, 00:12:05.126 { 00:12:05.126 "name": "BaseBdev2", 00:12:05.126 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:05.126 "is_configured": false, 00:12:05.126 "data_offset": 0, 00:12:05.126 "data_size": 0 00:12:05.126 } 00:12:05.126 ] 00:12:05.126 }' 00:12:05.126 11:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:05.126 11:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:05.692 11:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:12:05.950 [2024-07-15 11:22:49.482788] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:12:05.950 [2024-07-15 11:22:49.482823] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1c9f350 name Existed_Raid, state configuring 00:12:05.950 11:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r concat -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:12:06.208 [2024-07-15 11:22:49.659303] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:06.208 [2024-07-15 11:22:49.660784] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:12:06.208 [2024-07-15 11:22:49.660817] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:12:06.208 11:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:12:06.208 11:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:12:06.208 11:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:12:06.208 11:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:12:06.208 11:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:12:06.208 11:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:12:06.208 11:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:12:06.208 11:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:12:06.208 11:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:06.208 11:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:06.208 11:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:06.208 11:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:06.208 11:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:06.208 11:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:06.466 11:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:06.466 "name": "Existed_Raid", 00:12:06.466 "uuid": "c36b4397-3ae3-48c7-925a-dd3a1baa60b4", 00:12:06.466 "strip_size_kb": 64, 00:12:06.466 "state": "configuring", 00:12:06.466 "raid_level": "concat", 00:12:06.466 "superblock": true, 00:12:06.466 "num_base_bdevs": 2, 00:12:06.466 "num_base_bdevs_discovered": 1, 00:12:06.466 "num_base_bdevs_operational": 2, 00:12:06.466 "base_bdevs_list": [ 00:12:06.466 { 00:12:06.466 "name": "BaseBdev1", 00:12:06.466 "uuid": "1c002ed4-820d-4b70-933c-bc810aee61f5", 00:12:06.466 "is_configured": true, 00:12:06.466 "data_offset": 2048, 00:12:06.466 "data_size": 63488 00:12:06.466 }, 00:12:06.466 { 00:12:06.466 "name": "BaseBdev2", 00:12:06.466 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:06.466 "is_configured": false, 00:12:06.466 "data_offset": 0, 00:12:06.466 "data_size": 0 00:12:06.466 } 00:12:06.466 ] 00:12:06.466 }' 00:12:06.466 11:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:06.466 11:22:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:07.401 11:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:12:07.658 [2024-07-15 11:22:51.027465] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:07.658 [2024-07-15 11:22:51.027620] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1ca0000 00:12:07.658 [2024-07-15 11:22:51.027633] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:12:07.658 [2024-07-15 11:22:51.027803] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1bba0c0 00:12:07.658 [2024-07-15 11:22:51.027916] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1ca0000 00:12:07.658 [2024-07-15 11:22:51.027933] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x1ca0000 00:12:07.658 [2024-07-15 11:22:51.028024] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:07.658 BaseBdev2 00:12:07.658 11:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:12:07.658 11:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:12:07.658 11:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:12:07.658 11:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:12:07.658 11:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:12:07.658 11:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:12:07.658 11:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:12:07.917 11:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:12:08.175 [ 00:12:08.175 { 00:12:08.175 "name": "BaseBdev2", 00:12:08.175 "aliases": [ 00:12:08.175 "61d88843-e5fe-4a34-8c16-9e6127d68ca3" 00:12:08.175 ], 00:12:08.175 "product_name": "Malloc disk", 00:12:08.175 "block_size": 512, 00:12:08.175 "num_blocks": 65536, 00:12:08.175 "uuid": "61d88843-e5fe-4a34-8c16-9e6127d68ca3", 00:12:08.175 "assigned_rate_limits": { 00:12:08.175 "rw_ios_per_sec": 0, 00:12:08.175 "rw_mbytes_per_sec": 0, 00:12:08.175 "r_mbytes_per_sec": 0, 00:12:08.175 "w_mbytes_per_sec": 0 00:12:08.175 }, 00:12:08.175 "claimed": true, 00:12:08.175 "claim_type": "exclusive_write", 00:12:08.175 "zoned": false, 00:12:08.175 "supported_io_types": { 00:12:08.175 "read": true, 00:12:08.175 "write": true, 00:12:08.175 "unmap": true, 00:12:08.175 "flush": true, 00:12:08.175 "reset": true, 00:12:08.175 "nvme_admin": false, 00:12:08.175 "nvme_io": false, 00:12:08.175 "nvme_io_md": false, 00:12:08.175 "write_zeroes": true, 00:12:08.175 "zcopy": true, 00:12:08.175 "get_zone_info": false, 00:12:08.175 "zone_management": false, 00:12:08.175 "zone_append": false, 00:12:08.175 "compare": false, 00:12:08.175 "compare_and_write": false, 00:12:08.175 "abort": true, 00:12:08.175 "seek_hole": false, 00:12:08.175 "seek_data": false, 00:12:08.175 "copy": true, 00:12:08.175 "nvme_iov_md": false 00:12:08.175 }, 00:12:08.175 "memory_domains": [ 00:12:08.175 { 00:12:08.175 "dma_device_id": "system", 00:12:08.175 "dma_device_type": 1 00:12:08.175 }, 00:12:08.175 { 00:12:08.175 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:08.175 "dma_device_type": 2 00:12:08.175 } 00:12:08.175 ], 00:12:08.175 "driver_specific": {} 00:12:08.175 } 00:12:08.175 ] 00:12:08.175 11:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:12:08.175 11:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:12:08.175 11:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:12:08.175 11:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:12:08.175 11:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:12:08.175 11:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:12:08.175 11:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:12:08.175 11:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:12:08.175 11:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:12:08.175 11:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:08.175 11:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:08.175 11:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:08.175 11:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:08.175 11:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:08.175 11:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:08.432 11:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:08.432 "name": "Existed_Raid", 00:12:08.432 "uuid": "c36b4397-3ae3-48c7-925a-dd3a1baa60b4", 00:12:08.432 "strip_size_kb": 64, 00:12:08.432 "state": "online", 00:12:08.432 "raid_level": "concat", 00:12:08.432 "superblock": true, 00:12:08.432 "num_base_bdevs": 2, 00:12:08.432 "num_base_bdevs_discovered": 2, 00:12:08.432 "num_base_bdevs_operational": 2, 00:12:08.432 "base_bdevs_list": [ 00:12:08.432 { 00:12:08.432 "name": "BaseBdev1", 00:12:08.432 "uuid": "1c002ed4-820d-4b70-933c-bc810aee61f5", 00:12:08.432 "is_configured": true, 00:12:08.432 "data_offset": 2048, 00:12:08.432 "data_size": 63488 00:12:08.432 }, 00:12:08.432 { 00:12:08.432 "name": "BaseBdev2", 00:12:08.432 "uuid": "61d88843-e5fe-4a34-8c16-9e6127d68ca3", 00:12:08.432 "is_configured": true, 00:12:08.432 "data_offset": 2048, 00:12:08.432 "data_size": 63488 00:12:08.432 } 00:12:08.432 ] 00:12:08.432 }' 00:12:08.432 11:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:08.432 11:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.998 11:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:12:08.998 11:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:12:08.998 11:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:12:08.998 11:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:12:08.998 11:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:12:08.998 11:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # local name 00:12:08.998 11:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:12:08.998 11:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:12:08.998 [2024-07-15 11:22:52.591875] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:09.256 11:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:12:09.256 "name": "Existed_Raid", 00:12:09.256 "aliases": [ 00:12:09.256 "c36b4397-3ae3-48c7-925a-dd3a1baa60b4" 00:12:09.256 ], 00:12:09.256 "product_name": "Raid Volume", 00:12:09.256 "block_size": 512, 00:12:09.256 "num_blocks": 126976, 00:12:09.256 "uuid": "c36b4397-3ae3-48c7-925a-dd3a1baa60b4", 00:12:09.256 "assigned_rate_limits": { 00:12:09.256 "rw_ios_per_sec": 0, 00:12:09.256 "rw_mbytes_per_sec": 0, 00:12:09.256 "r_mbytes_per_sec": 0, 00:12:09.256 "w_mbytes_per_sec": 0 00:12:09.256 }, 00:12:09.256 "claimed": false, 00:12:09.256 "zoned": false, 00:12:09.256 "supported_io_types": { 00:12:09.256 "read": true, 00:12:09.256 "write": true, 00:12:09.256 "unmap": true, 00:12:09.256 "flush": true, 00:12:09.256 "reset": true, 00:12:09.256 "nvme_admin": false, 00:12:09.256 "nvme_io": false, 00:12:09.256 "nvme_io_md": false, 00:12:09.256 "write_zeroes": true, 00:12:09.256 "zcopy": false, 00:12:09.256 "get_zone_info": false, 00:12:09.256 "zone_management": false, 00:12:09.256 "zone_append": false, 00:12:09.256 "compare": false, 00:12:09.256 "compare_and_write": false, 00:12:09.256 "abort": false, 00:12:09.256 "seek_hole": false, 00:12:09.256 "seek_data": false, 00:12:09.256 "copy": false, 00:12:09.256 "nvme_iov_md": false 00:12:09.256 }, 00:12:09.256 "memory_domains": [ 00:12:09.256 { 00:12:09.256 "dma_device_id": "system", 00:12:09.256 "dma_device_type": 1 00:12:09.256 }, 00:12:09.256 { 00:12:09.256 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:09.256 "dma_device_type": 2 00:12:09.256 }, 00:12:09.256 { 00:12:09.256 "dma_device_id": "system", 00:12:09.256 "dma_device_type": 1 00:12:09.256 }, 00:12:09.256 { 00:12:09.256 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:09.256 "dma_device_type": 2 00:12:09.256 } 00:12:09.256 ], 00:12:09.256 "driver_specific": { 00:12:09.256 "raid": { 00:12:09.256 "uuid": "c36b4397-3ae3-48c7-925a-dd3a1baa60b4", 00:12:09.256 "strip_size_kb": 64, 00:12:09.256 "state": "online", 00:12:09.256 "raid_level": "concat", 00:12:09.256 "superblock": true, 00:12:09.256 "num_base_bdevs": 2, 00:12:09.256 "num_base_bdevs_discovered": 2, 00:12:09.256 "num_base_bdevs_operational": 2, 00:12:09.256 "base_bdevs_list": [ 00:12:09.256 { 00:12:09.256 "name": "BaseBdev1", 00:12:09.256 "uuid": "1c002ed4-820d-4b70-933c-bc810aee61f5", 00:12:09.256 "is_configured": true, 00:12:09.256 "data_offset": 2048, 00:12:09.256 "data_size": 63488 00:12:09.256 }, 00:12:09.256 { 00:12:09.256 "name": "BaseBdev2", 00:12:09.256 "uuid": "61d88843-e5fe-4a34-8c16-9e6127d68ca3", 00:12:09.256 "is_configured": true, 00:12:09.256 "data_offset": 2048, 00:12:09.256 "data_size": 63488 00:12:09.256 } 00:12:09.256 ] 00:12:09.256 } 00:12:09.256 } 00:12:09.256 }' 00:12:09.256 11:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:12:09.256 11:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:12:09.256 BaseBdev2' 00:12:09.256 11:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:12:09.256 11:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:12:09.256 11:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:12:09.515 11:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:12:09.515 "name": "BaseBdev1", 00:12:09.515 "aliases": [ 00:12:09.515 "1c002ed4-820d-4b70-933c-bc810aee61f5" 00:12:09.515 ], 00:12:09.515 "product_name": "Malloc disk", 00:12:09.515 "block_size": 512, 00:12:09.515 "num_blocks": 65536, 00:12:09.515 "uuid": "1c002ed4-820d-4b70-933c-bc810aee61f5", 00:12:09.515 "assigned_rate_limits": { 00:12:09.515 "rw_ios_per_sec": 0, 00:12:09.515 "rw_mbytes_per_sec": 0, 00:12:09.515 "r_mbytes_per_sec": 0, 00:12:09.515 "w_mbytes_per_sec": 0 00:12:09.515 }, 00:12:09.515 "claimed": true, 00:12:09.515 "claim_type": "exclusive_write", 00:12:09.515 "zoned": false, 00:12:09.515 "supported_io_types": { 00:12:09.515 "read": true, 00:12:09.515 "write": true, 00:12:09.515 "unmap": true, 00:12:09.515 "flush": true, 00:12:09.515 "reset": true, 00:12:09.515 "nvme_admin": false, 00:12:09.515 "nvme_io": false, 00:12:09.515 "nvme_io_md": false, 00:12:09.515 "write_zeroes": true, 00:12:09.515 "zcopy": true, 00:12:09.515 "get_zone_info": false, 00:12:09.515 "zone_management": false, 00:12:09.515 "zone_append": false, 00:12:09.515 "compare": false, 00:12:09.515 "compare_and_write": false, 00:12:09.515 "abort": true, 00:12:09.515 "seek_hole": false, 00:12:09.515 "seek_data": false, 00:12:09.515 "copy": true, 00:12:09.515 "nvme_iov_md": false 00:12:09.515 }, 00:12:09.515 "memory_domains": [ 00:12:09.515 { 00:12:09.515 "dma_device_id": "system", 00:12:09.515 "dma_device_type": 1 00:12:09.515 }, 00:12:09.515 { 00:12:09.515 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:09.515 "dma_device_type": 2 00:12:09.515 } 00:12:09.515 ], 00:12:09.515 "driver_specific": {} 00:12:09.515 }' 00:12:09.515 11:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:12:09.515 11:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:12:09.515 11:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:12:09.515 11:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:12:09.515 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:12:09.515 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:12:09.515 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:12:09.773 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:12:09.773 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:12:09.773 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:12:09.773 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:12:09.773 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:12:09.773 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:12:09.773 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:12:09.773 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:12:10.031 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:12:10.031 "name": "BaseBdev2", 00:12:10.031 "aliases": [ 00:12:10.031 "61d88843-e5fe-4a34-8c16-9e6127d68ca3" 00:12:10.032 ], 00:12:10.032 "product_name": "Malloc disk", 00:12:10.032 "block_size": 512, 00:12:10.032 "num_blocks": 65536, 00:12:10.032 "uuid": "61d88843-e5fe-4a34-8c16-9e6127d68ca3", 00:12:10.032 "assigned_rate_limits": { 00:12:10.032 "rw_ios_per_sec": 0, 00:12:10.032 "rw_mbytes_per_sec": 0, 00:12:10.032 "r_mbytes_per_sec": 0, 00:12:10.032 "w_mbytes_per_sec": 0 00:12:10.032 }, 00:12:10.032 "claimed": true, 00:12:10.032 "claim_type": "exclusive_write", 00:12:10.032 "zoned": false, 00:12:10.032 "supported_io_types": { 00:12:10.032 "read": true, 00:12:10.032 "write": true, 00:12:10.032 "unmap": true, 00:12:10.032 "flush": true, 00:12:10.032 "reset": true, 00:12:10.032 "nvme_admin": false, 00:12:10.032 "nvme_io": false, 00:12:10.032 "nvme_io_md": false, 00:12:10.032 "write_zeroes": true, 00:12:10.032 "zcopy": true, 00:12:10.032 "get_zone_info": false, 00:12:10.032 "zone_management": false, 00:12:10.032 "zone_append": false, 00:12:10.032 "compare": false, 00:12:10.032 "compare_and_write": false, 00:12:10.032 "abort": true, 00:12:10.032 "seek_hole": false, 00:12:10.032 "seek_data": false, 00:12:10.032 "copy": true, 00:12:10.032 "nvme_iov_md": false 00:12:10.032 }, 00:12:10.032 "memory_domains": [ 00:12:10.032 { 00:12:10.032 "dma_device_id": "system", 00:12:10.032 "dma_device_type": 1 00:12:10.032 }, 00:12:10.032 { 00:12:10.032 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:10.032 "dma_device_type": 2 00:12:10.032 } 00:12:10.032 ], 00:12:10.032 "driver_specific": {} 00:12:10.032 }' 00:12:10.032 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:12:10.032 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:12:10.032 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:12:10.032 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:12:10.291 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:12:10.291 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:12:10.291 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:12:10.291 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:12:10.291 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:12:10.291 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:12:10.291 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:12:10.548 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:12:10.548 11:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:12:10.548 [2024-07-15 11:22:54.111703] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:10.548 [2024-07-15 11:22:54.111727] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:10.548 [2024-07-15 11:22:54.111768] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:10.548 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@275 -- # local expected_state 00:12:10.548 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # has_redundancy concat 00:12:10.548 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # case $1 in 00:12:10.548 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # return 1 00:12:10.548 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@277 -- # expected_state=offline 00:12:10.548 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:12:10.548 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:12:10.548 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=offline 00:12:10.548 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:12:10.548 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:12:10.548 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:12:10.548 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:10.548 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:10.548 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:10.548 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:10.548 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:10.548 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:10.805 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:10.805 "name": "Existed_Raid", 00:12:10.805 "uuid": "c36b4397-3ae3-48c7-925a-dd3a1baa60b4", 00:12:10.805 "strip_size_kb": 64, 00:12:10.805 "state": "offline", 00:12:10.805 "raid_level": "concat", 00:12:10.805 "superblock": true, 00:12:10.805 "num_base_bdevs": 2, 00:12:10.805 "num_base_bdevs_discovered": 1, 00:12:10.805 "num_base_bdevs_operational": 1, 00:12:10.805 "base_bdevs_list": [ 00:12:10.805 { 00:12:10.805 "name": null, 00:12:10.805 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:10.805 "is_configured": false, 00:12:10.805 "data_offset": 2048, 00:12:10.805 "data_size": 63488 00:12:10.805 }, 00:12:10.805 { 00:12:10.805 "name": "BaseBdev2", 00:12:10.805 "uuid": "61d88843-e5fe-4a34-8c16-9e6127d68ca3", 00:12:10.805 "is_configured": true, 00:12:10.805 "data_offset": 2048, 00:12:10.805 "data_size": 63488 00:12:10.805 } 00:12:10.805 ] 00:12:10.805 }' 00:12:10.805 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:10.805 11:22:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:11.735 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:12:11.735 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:12:11.735 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:11.735 11:22:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:12:11.735 11:22:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:12:11.735 11:22:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:12:11.735 11:22:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:12:12.299 [2024-07-15 11:22:55.692889] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:12.299 [2024-07-15 11:22:55.692950] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1ca0000 name Existed_Raid, state offline 00:12:12.299 11:22:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:12:12.299 11:22:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:12:12.299 11:22:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:12.299 11:22:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:12:12.556 11:22:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:12:12.556 11:22:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:12:12.556 11:22:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # '[' 2 -gt 2 ']' 00:12:12.556 11:22:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@341 -- # killprocess 869942 00:12:12.556 11:22:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@948 -- # '[' -z 869942 ']' 00:12:12.556 11:22:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@952 -- # kill -0 869942 00:12:12.556 11:22:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@953 -- # uname 00:12:12.556 11:22:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:12:12.556 11:22:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 869942 00:12:12.556 11:22:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:12:12.556 11:22:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:12:12.556 11:22:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@966 -- # echo 'killing process with pid 869942' 00:12:12.556 killing process with pid 869942 00:12:12.556 11:22:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@967 -- # kill 869942 00:12:12.556 [2024-07-15 11:22:56.030056] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:12.556 11:22:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # wait 869942 00:12:12.556 [2024-07-15 11:22:56.031025] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:12.815 11:22:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@343 -- # return 0 00:12:12.815 00:12:12.815 real 0m10.675s 00:12:12.815 user 0m19.416s 00:12:12.815 sys 0m2.038s 00:12:12.815 11:22:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1124 -- # xtrace_disable 00:12:12.815 11:22:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:12.815 ************************************ 00:12:12.815 END TEST raid_state_function_test_sb 00:12:12.815 ************************************ 00:12:12.815 11:22:56 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:12:12.815 11:22:56 bdev_raid -- bdev/bdev_raid.sh@869 -- # run_test raid_superblock_test raid_superblock_test concat 2 00:12:12.815 11:22:56 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:12:12.815 11:22:56 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:12:12.815 11:22:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:12.815 ************************************ 00:12:12.815 START TEST raid_superblock_test 00:12:12.815 ************************************ 00:12:12.815 11:22:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1123 -- # raid_superblock_test concat 2 00:12:12.815 11:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@392 -- # local raid_level=concat 00:12:12.815 11:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local num_base_bdevs=2 00:12:12.815 11:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # base_bdevs_malloc=() 00:12:12.815 11:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local base_bdevs_malloc 00:12:12.815 11:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_pt=() 00:12:12.815 11:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_pt 00:12:12.815 11:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt_uuid=() 00:12:12.815 11:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt_uuid 00:12:12.815 11:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local raid_bdev_name=raid_bdev1 00:12:12.815 11:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local strip_size 00:12:12.815 11:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size_create_arg 00:12:12.815 11:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local raid_bdev_uuid 00:12:12.815 11:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev 00:12:12.815 11:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@403 -- # '[' concat '!=' raid1 ']' 00:12:12.815 11:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # strip_size=64 00:12:12.815 11:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size_create_arg='-z 64' 00:12:12.815 11:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # raid_pid=871570 00:12:12.816 11:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # waitforlisten 871570 /var/tmp/spdk-raid.sock 00:12:12.816 11:22:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@829 -- # '[' -z 871570 ']' 00:12:12.816 11:22:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:12:12.816 11:22:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:12:12.816 11:22:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:12:12.816 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:12:12.816 11:22:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:12:12.816 11:22:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:12.816 11:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@410 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -L bdev_raid 00:12:12.816 [2024-07-15 11:22:56.393683] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:12:12.816 [2024-07-15 11:22:56.393745] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid871570 ] 00:12:13.073 [2024-07-15 11:22:56.519556] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:13.073 [2024-07-15 11:22:56.622243] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:12:13.331 [2024-07-15 11:22:56.680903] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:13.331 [2024-07-15 11:22:56.680954] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:14.264 11:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:12:14.264 11:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@862 -- # return 0 00:12:14.264 11:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i = 1 )) 00:12:14.264 11:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:12:14.264 11:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc1 00:12:14.264 11:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt1 00:12:14.264 11:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:12:14.264 11:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:12:14.264 11:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:12:14.264 11:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:12:14.264 11:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc1 00:12:14.264 malloc1 00:12:14.264 11:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:12:14.830 [2024-07-15 11:22:58.323064] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:12:14.830 [2024-07-15 11:22:58.323114] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:14.830 [2024-07-15 11:22:58.323138] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1574570 00:12:14.830 [2024-07-15 11:22:58.323150] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:14.830 [2024-07-15 11:22:58.324891] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:14.830 [2024-07-15 11:22:58.324920] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:12:14.830 pt1 00:12:14.830 11:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:12:14.830 11:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:12:14.830 11:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc2 00:12:14.830 11:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt2 00:12:14.830 11:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:12:14.830 11:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:12:14.830 11:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:12:14.830 11:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:12:14.830 11:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc2 00:12:15.396 malloc2 00:12:15.396 11:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:12:15.654 [2024-07-15 11:22:59.103313] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:12:15.654 [2024-07-15 11:22:59.103360] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:15.654 [2024-07-15 11:22:59.103378] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1575970 00:12:15.654 [2024-07-15 11:22:59.103391] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:15.654 [2024-07-15 11:22:59.105048] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:15.654 [2024-07-15 11:22:59.105076] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:12:15.654 pt2 00:12:15.654 11:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:12:15.654 11:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:12:15.654 11:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@429 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'pt1 pt2' -n raid_bdev1 -s 00:12:16.219 [2024-07-15 11:22:59.604634] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:12:16.219 [2024-07-15 11:22:59.605953] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:12:16.219 [2024-07-15 11:22:59.606098] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1718270 00:12:16.219 [2024-07-15 11:22:59.606111] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:12:16.219 [2024-07-15 11:22:59.606309] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x170dc10 00:12:16.219 [2024-07-15 11:22:59.606454] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1718270 00:12:16.219 [2024-07-15 11:22:59.606464] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x1718270 00:12:16.219 [2024-07-15 11:22:59.606563] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:16.219 11:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:12:16.219 11:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:12:16.219 11:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:12:16.219 11:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:12:16.219 11:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:12:16.219 11:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:12:16.219 11:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:16.219 11:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:16.219 11:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:16.219 11:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:16.219 11:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:16.219 11:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:16.478 11:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:16.478 "name": "raid_bdev1", 00:12:16.478 "uuid": "1c485329-fed8-4d07-b586-e5b1a4fce65b", 00:12:16.478 "strip_size_kb": 64, 00:12:16.478 "state": "online", 00:12:16.478 "raid_level": "concat", 00:12:16.478 "superblock": true, 00:12:16.478 "num_base_bdevs": 2, 00:12:16.478 "num_base_bdevs_discovered": 2, 00:12:16.478 "num_base_bdevs_operational": 2, 00:12:16.478 "base_bdevs_list": [ 00:12:16.478 { 00:12:16.478 "name": "pt1", 00:12:16.478 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:16.478 "is_configured": true, 00:12:16.478 "data_offset": 2048, 00:12:16.478 "data_size": 63488 00:12:16.478 }, 00:12:16.478 { 00:12:16.478 "name": "pt2", 00:12:16.478 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:16.478 "is_configured": true, 00:12:16.478 "data_offset": 2048, 00:12:16.478 "data_size": 63488 00:12:16.478 } 00:12:16.478 ] 00:12:16.478 }' 00:12:16.478 11:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:16.478 11:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:17.045 11:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_properties raid_bdev1 00:12:17.045 11:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:12:17.045 11:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:12:17.045 11:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:12:17.045 11:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:12:17.045 11:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # local name 00:12:17.045 11:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:12:17.045 11:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:12:17.303 [2024-07-15 11:23:00.647606] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:17.303 11:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:12:17.303 "name": "raid_bdev1", 00:12:17.303 "aliases": [ 00:12:17.303 "1c485329-fed8-4d07-b586-e5b1a4fce65b" 00:12:17.303 ], 00:12:17.303 "product_name": "Raid Volume", 00:12:17.303 "block_size": 512, 00:12:17.303 "num_blocks": 126976, 00:12:17.303 "uuid": "1c485329-fed8-4d07-b586-e5b1a4fce65b", 00:12:17.303 "assigned_rate_limits": { 00:12:17.303 "rw_ios_per_sec": 0, 00:12:17.303 "rw_mbytes_per_sec": 0, 00:12:17.303 "r_mbytes_per_sec": 0, 00:12:17.303 "w_mbytes_per_sec": 0 00:12:17.303 }, 00:12:17.303 "claimed": false, 00:12:17.303 "zoned": false, 00:12:17.303 "supported_io_types": { 00:12:17.303 "read": true, 00:12:17.303 "write": true, 00:12:17.303 "unmap": true, 00:12:17.303 "flush": true, 00:12:17.303 "reset": true, 00:12:17.303 "nvme_admin": false, 00:12:17.303 "nvme_io": false, 00:12:17.303 "nvme_io_md": false, 00:12:17.303 "write_zeroes": true, 00:12:17.303 "zcopy": false, 00:12:17.303 "get_zone_info": false, 00:12:17.303 "zone_management": false, 00:12:17.303 "zone_append": false, 00:12:17.303 "compare": false, 00:12:17.303 "compare_and_write": false, 00:12:17.303 "abort": false, 00:12:17.303 "seek_hole": false, 00:12:17.303 "seek_data": false, 00:12:17.303 "copy": false, 00:12:17.303 "nvme_iov_md": false 00:12:17.303 }, 00:12:17.303 "memory_domains": [ 00:12:17.303 { 00:12:17.303 "dma_device_id": "system", 00:12:17.303 "dma_device_type": 1 00:12:17.303 }, 00:12:17.303 { 00:12:17.303 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:17.303 "dma_device_type": 2 00:12:17.303 }, 00:12:17.303 { 00:12:17.303 "dma_device_id": "system", 00:12:17.303 "dma_device_type": 1 00:12:17.303 }, 00:12:17.303 { 00:12:17.303 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:17.303 "dma_device_type": 2 00:12:17.303 } 00:12:17.303 ], 00:12:17.304 "driver_specific": { 00:12:17.304 "raid": { 00:12:17.304 "uuid": "1c485329-fed8-4d07-b586-e5b1a4fce65b", 00:12:17.304 "strip_size_kb": 64, 00:12:17.304 "state": "online", 00:12:17.304 "raid_level": "concat", 00:12:17.304 "superblock": true, 00:12:17.304 "num_base_bdevs": 2, 00:12:17.304 "num_base_bdevs_discovered": 2, 00:12:17.304 "num_base_bdevs_operational": 2, 00:12:17.304 "base_bdevs_list": [ 00:12:17.304 { 00:12:17.304 "name": "pt1", 00:12:17.304 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:17.304 "is_configured": true, 00:12:17.304 "data_offset": 2048, 00:12:17.304 "data_size": 63488 00:12:17.304 }, 00:12:17.304 { 00:12:17.304 "name": "pt2", 00:12:17.304 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:17.304 "is_configured": true, 00:12:17.304 "data_offset": 2048, 00:12:17.304 "data_size": 63488 00:12:17.304 } 00:12:17.304 ] 00:12:17.304 } 00:12:17.304 } 00:12:17.304 }' 00:12:17.304 11:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:12:17.304 11:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:12:17.304 pt2' 00:12:17.304 11:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:12:17.304 11:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:12:17.304 11:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:12:17.561 11:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:12:17.561 "name": "pt1", 00:12:17.561 "aliases": [ 00:12:17.561 "00000000-0000-0000-0000-000000000001" 00:12:17.561 ], 00:12:17.561 "product_name": "passthru", 00:12:17.561 "block_size": 512, 00:12:17.561 "num_blocks": 65536, 00:12:17.561 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:17.561 "assigned_rate_limits": { 00:12:17.561 "rw_ios_per_sec": 0, 00:12:17.561 "rw_mbytes_per_sec": 0, 00:12:17.561 "r_mbytes_per_sec": 0, 00:12:17.561 "w_mbytes_per_sec": 0 00:12:17.561 }, 00:12:17.561 "claimed": true, 00:12:17.561 "claim_type": "exclusive_write", 00:12:17.561 "zoned": false, 00:12:17.561 "supported_io_types": { 00:12:17.561 "read": true, 00:12:17.561 "write": true, 00:12:17.561 "unmap": true, 00:12:17.561 "flush": true, 00:12:17.561 "reset": true, 00:12:17.561 "nvme_admin": false, 00:12:17.561 "nvme_io": false, 00:12:17.561 "nvme_io_md": false, 00:12:17.561 "write_zeroes": true, 00:12:17.561 "zcopy": true, 00:12:17.561 "get_zone_info": false, 00:12:17.561 "zone_management": false, 00:12:17.561 "zone_append": false, 00:12:17.561 "compare": false, 00:12:17.561 "compare_and_write": false, 00:12:17.561 "abort": true, 00:12:17.561 "seek_hole": false, 00:12:17.561 "seek_data": false, 00:12:17.561 "copy": true, 00:12:17.561 "nvme_iov_md": false 00:12:17.561 }, 00:12:17.561 "memory_domains": [ 00:12:17.561 { 00:12:17.562 "dma_device_id": "system", 00:12:17.562 "dma_device_type": 1 00:12:17.562 }, 00:12:17.562 { 00:12:17.562 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:17.562 "dma_device_type": 2 00:12:17.562 } 00:12:17.562 ], 00:12:17.562 "driver_specific": { 00:12:17.562 "passthru": { 00:12:17.562 "name": "pt1", 00:12:17.562 "base_bdev_name": "malloc1" 00:12:17.562 } 00:12:17.562 } 00:12:17.562 }' 00:12:17.562 11:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:12:17.562 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:12:17.562 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:12:17.562 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:12:17.562 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:12:17.562 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:12:17.562 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:12:17.829 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:12:17.829 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:12:17.829 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:12:17.830 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:12:17.830 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:12:17.830 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:12:17.830 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:12:17.830 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:12:18.105 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:12:18.105 "name": "pt2", 00:12:18.105 "aliases": [ 00:12:18.105 "00000000-0000-0000-0000-000000000002" 00:12:18.105 ], 00:12:18.105 "product_name": "passthru", 00:12:18.105 "block_size": 512, 00:12:18.105 "num_blocks": 65536, 00:12:18.105 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:18.105 "assigned_rate_limits": { 00:12:18.105 "rw_ios_per_sec": 0, 00:12:18.105 "rw_mbytes_per_sec": 0, 00:12:18.105 "r_mbytes_per_sec": 0, 00:12:18.105 "w_mbytes_per_sec": 0 00:12:18.105 }, 00:12:18.105 "claimed": true, 00:12:18.105 "claim_type": "exclusive_write", 00:12:18.105 "zoned": false, 00:12:18.105 "supported_io_types": { 00:12:18.105 "read": true, 00:12:18.105 "write": true, 00:12:18.105 "unmap": true, 00:12:18.105 "flush": true, 00:12:18.105 "reset": true, 00:12:18.105 "nvme_admin": false, 00:12:18.105 "nvme_io": false, 00:12:18.105 "nvme_io_md": false, 00:12:18.105 "write_zeroes": true, 00:12:18.105 "zcopy": true, 00:12:18.105 "get_zone_info": false, 00:12:18.105 "zone_management": false, 00:12:18.105 "zone_append": false, 00:12:18.105 "compare": false, 00:12:18.105 "compare_and_write": false, 00:12:18.105 "abort": true, 00:12:18.105 "seek_hole": false, 00:12:18.105 "seek_data": false, 00:12:18.105 "copy": true, 00:12:18.105 "nvme_iov_md": false 00:12:18.105 }, 00:12:18.105 "memory_domains": [ 00:12:18.105 { 00:12:18.105 "dma_device_id": "system", 00:12:18.105 "dma_device_type": 1 00:12:18.105 }, 00:12:18.105 { 00:12:18.105 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:18.105 "dma_device_type": 2 00:12:18.105 } 00:12:18.105 ], 00:12:18.105 "driver_specific": { 00:12:18.105 "passthru": { 00:12:18.105 "name": "pt2", 00:12:18.105 "base_bdev_name": "malloc2" 00:12:18.105 } 00:12:18.105 } 00:12:18.105 }' 00:12:18.105 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:12:18.105 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:12:18.105 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:12:18.105 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:12:18.105 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:12:18.363 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:12:18.363 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:12:18.363 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:12:18.363 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:12:18.363 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:12:18.363 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:12:18.363 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:12:18.363 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:12:18.363 11:23:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # jq -r '.[] | .uuid' 00:12:18.620 [2024-07-15 11:23:02.143567] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:18.620 11:23:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # raid_bdev_uuid=1c485329-fed8-4d07-b586-e5b1a4fce65b 00:12:18.620 11:23:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # '[' -z 1c485329-fed8-4d07-b586-e5b1a4fce65b ']' 00:12:18.620 11:23:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@440 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:12:18.876 [2024-07-15 11:23:02.391987] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:18.876 [2024-07-15 11:23:02.392010] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:18.876 [2024-07-15 11:23:02.392061] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:18.876 [2024-07-15 11:23:02.392105] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:18.876 [2024-07-15 11:23:02.392117] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1718270 name raid_bdev1, state offline 00:12:18.876 11:23:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:18.876 11:23:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # jq -r '.[]' 00:12:19.133 11:23:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # raid_bdev= 00:12:19.133 11:23:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # '[' -n '' ']' 00:12:19.133 11:23:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:12:19.133 11:23:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt1 00:12:19.390 11:23:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:12:19.390 11:23:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:12:19.647 11:23:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs 00:12:19.647 11:23:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:12:19.903 11:23:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # '[' false == true ']' 00:12:19.903 11:23:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@456 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'malloc1 malloc2' -n raid_bdev1 00:12:19.903 11:23:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@648 -- # local es=0 00:12:19.903 11:23:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'malloc1 malloc2' -n raid_bdev1 00:12:19.903 11:23:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:12:19.903 11:23:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:12:19.903 11:23:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:12:19.903 11:23:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:12:19.903 11:23:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:12:19.903 11:23:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:12:19.903 11:23:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:12:19.903 11:23:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py ]] 00:12:19.903 11:23:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'malloc1 malloc2' -n raid_bdev1 00:12:20.159 [2024-07-15 11:23:03.615184] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:12:20.159 [2024-07-15 11:23:03.616569] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:12:20.159 [2024-07-15 11:23:03.616635] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:12:20.159 [2024-07-15 11:23:03.616675] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:12:20.159 [2024-07-15 11:23:03.616694] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:20.159 [2024-07-15 11:23:03.616704] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1717ff0 name raid_bdev1, state configuring 00:12:20.159 request: 00:12:20.159 { 00:12:20.159 "name": "raid_bdev1", 00:12:20.159 "raid_level": "concat", 00:12:20.159 "base_bdevs": [ 00:12:20.159 "malloc1", 00:12:20.159 "malloc2" 00:12:20.159 ], 00:12:20.159 "strip_size_kb": 64, 00:12:20.159 "superblock": false, 00:12:20.159 "method": "bdev_raid_create", 00:12:20.159 "req_id": 1 00:12:20.159 } 00:12:20.159 Got JSON-RPC error response 00:12:20.159 response: 00:12:20.159 { 00:12:20.159 "code": -17, 00:12:20.159 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:12:20.159 } 00:12:20.159 11:23:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@651 -- # es=1 00:12:20.159 11:23:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:12:20.159 11:23:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:12:20.159 11:23:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:12:20.160 11:23:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:20.160 11:23:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # jq -r '.[]' 00:12:20.415 11:23:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # raid_bdev= 00:12:20.415 11:23:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # '[' -n '' ']' 00:12:20.415 11:23:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@464 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:12:20.673 [2024-07-15 11:23:04.108415] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:12:20.673 [2024-07-15 11:23:04.108456] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:20.673 [2024-07-15 11:23:04.108477] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x15747a0 00:12:20.673 [2024-07-15 11:23:04.108490] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:20.673 [2024-07-15 11:23:04.110070] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:20.673 [2024-07-15 11:23:04.110099] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:12:20.673 [2024-07-15 11:23:04.110173] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:12:20.673 [2024-07-15 11:23:04.110199] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:12:20.673 pt1 00:12:20.673 11:23:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@467 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 2 00:12:20.673 11:23:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:12:20.673 11:23:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:12:20.673 11:23:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:12:20.673 11:23:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:12:20.673 11:23:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:12:20.673 11:23:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:20.673 11:23:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:20.673 11:23:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:20.673 11:23:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:20.673 11:23:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:20.673 11:23:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:20.930 11:23:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:20.930 "name": "raid_bdev1", 00:12:20.930 "uuid": "1c485329-fed8-4d07-b586-e5b1a4fce65b", 00:12:20.930 "strip_size_kb": 64, 00:12:20.930 "state": "configuring", 00:12:20.930 "raid_level": "concat", 00:12:20.930 "superblock": true, 00:12:20.930 "num_base_bdevs": 2, 00:12:20.930 "num_base_bdevs_discovered": 1, 00:12:20.930 "num_base_bdevs_operational": 2, 00:12:20.930 "base_bdevs_list": [ 00:12:20.930 { 00:12:20.930 "name": "pt1", 00:12:20.930 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:20.930 "is_configured": true, 00:12:20.930 "data_offset": 2048, 00:12:20.930 "data_size": 63488 00:12:20.930 }, 00:12:20.930 { 00:12:20.930 "name": null, 00:12:20.930 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:20.930 "is_configured": false, 00:12:20.930 "data_offset": 2048, 00:12:20.930 "data_size": 63488 00:12:20.930 } 00:12:20.930 ] 00:12:20.930 }' 00:12:20.930 11:23:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:20.930 11:23:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:21.493 11:23:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@469 -- # '[' 2 -gt 2 ']' 00:12:21.493 11:23:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i = 1 )) 00:12:21.493 11:23:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:12:21.493 11:23:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:12:21.750 [2024-07-15 11:23:05.115091] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:12:21.750 [2024-07-15 11:23:05.115138] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:21.750 [2024-07-15 11:23:05.115156] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x170e820 00:12:21.750 [2024-07-15 11:23:05.115169] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:21.750 [2024-07-15 11:23:05.115504] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:21.750 [2024-07-15 11:23:05.115522] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:12:21.750 [2024-07-15 11:23:05.115581] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:12:21.750 [2024-07-15 11:23:05.115600] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:12:21.750 [2024-07-15 11:23:05.115693] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x156aec0 00:12:21.750 [2024-07-15 11:23:05.115704] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:12:21.750 [2024-07-15 11:23:05.115870] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x156bf00 00:12:21.750 [2024-07-15 11:23:05.116006] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x156aec0 00:12:21.750 [2024-07-15 11:23:05.116017] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x156aec0 00:12:21.751 [2024-07-15 11:23:05.116113] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:21.751 pt2 00:12:21.751 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:12:21.751 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:12:21.751 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@482 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:12:21.751 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:12:21.751 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:12:21.751 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:12:21.751 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:12:21.751 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:12:21.751 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:21.751 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:21.751 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:21.751 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:21.751 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:21.751 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:21.751 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:21.751 "name": "raid_bdev1", 00:12:21.751 "uuid": "1c485329-fed8-4d07-b586-e5b1a4fce65b", 00:12:21.751 "strip_size_kb": 64, 00:12:21.751 "state": "online", 00:12:21.751 "raid_level": "concat", 00:12:21.751 "superblock": true, 00:12:21.751 "num_base_bdevs": 2, 00:12:21.751 "num_base_bdevs_discovered": 2, 00:12:21.751 "num_base_bdevs_operational": 2, 00:12:21.751 "base_bdevs_list": [ 00:12:21.751 { 00:12:21.751 "name": "pt1", 00:12:21.751 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:21.751 "is_configured": true, 00:12:21.751 "data_offset": 2048, 00:12:21.751 "data_size": 63488 00:12:21.751 }, 00:12:21.751 { 00:12:21.751 "name": "pt2", 00:12:21.751 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:21.751 "is_configured": true, 00:12:21.751 "data_offset": 2048, 00:12:21.751 "data_size": 63488 00:12:21.751 } 00:12:21.751 ] 00:12:21.751 }' 00:12:21.751 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:21.751 11:23:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:22.317 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_properties raid_bdev1 00:12:22.317 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:12:22.317 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:12:22.317 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:12:22.317 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:12:22.317 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # local name 00:12:22.575 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:12:22.575 11:23:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:12:22.575 [2024-07-15 11:23:06.134045] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:22.575 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:12:22.575 "name": "raid_bdev1", 00:12:22.575 "aliases": [ 00:12:22.575 "1c485329-fed8-4d07-b586-e5b1a4fce65b" 00:12:22.575 ], 00:12:22.575 "product_name": "Raid Volume", 00:12:22.575 "block_size": 512, 00:12:22.575 "num_blocks": 126976, 00:12:22.575 "uuid": "1c485329-fed8-4d07-b586-e5b1a4fce65b", 00:12:22.575 "assigned_rate_limits": { 00:12:22.575 "rw_ios_per_sec": 0, 00:12:22.575 "rw_mbytes_per_sec": 0, 00:12:22.575 "r_mbytes_per_sec": 0, 00:12:22.575 "w_mbytes_per_sec": 0 00:12:22.575 }, 00:12:22.575 "claimed": false, 00:12:22.575 "zoned": false, 00:12:22.575 "supported_io_types": { 00:12:22.575 "read": true, 00:12:22.575 "write": true, 00:12:22.575 "unmap": true, 00:12:22.575 "flush": true, 00:12:22.575 "reset": true, 00:12:22.575 "nvme_admin": false, 00:12:22.575 "nvme_io": false, 00:12:22.575 "nvme_io_md": false, 00:12:22.575 "write_zeroes": true, 00:12:22.575 "zcopy": false, 00:12:22.575 "get_zone_info": false, 00:12:22.575 "zone_management": false, 00:12:22.575 "zone_append": false, 00:12:22.575 "compare": false, 00:12:22.575 "compare_and_write": false, 00:12:22.575 "abort": false, 00:12:22.575 "seek_hole": false, 00:12:22.575 "seek_data": false, 00:12:22.575 "copy": false, 00:12:22.575 "nvme_iov_md": false 00:12:22.575 }, 00:12:22.575 "memory_domains": [ 00:12:22.575 { 00:12:22.575 "dma_device_id": "system", 00:12:22.575 "dma_device_type": 1 00:12:22.575 }, 00:12:22.575 { 00:12:22.575 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:22.575 "dma_device_type": 2 00:12:22.575 }, 00:12:22.575 { 00:12:22.575 "dma_device_id": "system", 00:12:22.575 "dma_device_type": 1 00:12:22.575 }, 00:12:22.575 { 00:12:22.575 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:22.575 "dma_device_type": 2 00:12:22.575 } 00:12:22.575 ], 00:12:22.575 "driver_specific": { 00:12:22.575 "raid": { 00:12:22.575 "uuid": "1c485329-fed8-4d07-b586-e5b1a4fce65b", 00:12:22.575 "strip_size_kb": 64, 00:12:22.575 "state": "online", 00:12:22.575 "raid_level": "concat", 00:12:22.575 "superblock": true, 00:12:22.575 "num_base_bdevs": 2, 00:12:22.575 "num_base_bdevs_discovered": 2, 00:12:22.575 "num_base_bdevs_operational": 2, 00:12:22.575 "base_bdevs_list": [ 00:12:22.575 { 00:12:22.575 "name": "pt1", 00:12:22.575 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:22.575 "is_configured": true, 00:12:22.575 "data_offset": 2048, 00:12:22.575 "data_size": 63488 00:12:22.575 }, 00:12:22.575 { 00:12:22.575 "name": "pt2", 00:12:22.575 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:22.575 "is_configured": true, 00:12:22.575 "data_offset": 2048, 00:12:22.575 "data_size": 63488 00:12:22.575 } 00:12:22.575 ] 00:12:22.575 } 00:12:22.575 } 00:12:22.575 }' 00:12:22.575 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:12:22.834 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:12:22.834 pt2' 00:12:22.834 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:12:22.834 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:12:22.834 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:12:23.092 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:12:23.092 "name": "pt1", 00:12:23.092 "aliases": [ 00:12:23.092 "00000000-0000-0000-0000-000000000001" 00:12:23.092 ], 00:12:23.092 "product_name": "passthru", 00:12:23.092 "block_size": 512, 00:12:23.092 "num_blocks": 65536, 00:12:23.092 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:23.092 "assigned_rate_limits": { 00:12:23.092 "rw_ios_per_sec": 0, 00:12:23.092 "rw_mbytes_per_sec": 0, 00:12:23.092 "r_mbytes_per_sec": 0, 00:12:23.092 "w_mbytes_per_sec": 0 00:12:23.092 }, 00:12:23.092 "claimed": true, 00:12:23.092 "claim_type": "exclusive_write", 00:12:23.092 "zoned": false, 00:12:23.092 "supported_io_types": { 00:12:23.092 "read": true, 00:12:23.092 "write": true, 00:12:23.092 "unmap": true, 00:12:23.092 "flush": true, 00:12:23.092 "reset": true, 00:12:23.092 "nvme_admin": false, 00:12:23.092 "nvme_io": false, 00:12:23.092 "nvme_io_md": false, 00:12:23.092 "write_zeroes": true, 00:12:23.092 "zcopy": true, 00:12:23.092 "get_zone_info": false, 00:12:23.092 "zone_management": false, 00:12:23.092 "zone_append": false, 00:12:23.092 "compare": false, 00:12:23.092 "compare_and_write": false, 00:12:23.092 "abort": true, 00:12:23.092 "seek_hole": false, 00:12:23.092 "seek_data": false, 00:12:23.092 "copy": true, 00:12:23.092 "nvme_iov_md": false 00:12:23.092 }, 00:12:23.092 "memory_domains": [ 00:12:23.092 { 00:12:23.092 "dma_device_id": "system", 00:12:23.092 "dma_device_type": 1 00:12:23.092 }, 00:12:23.092 { 00:12:23.092 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:23.092 "dma_device_type": 2 00:12:23.092 } 00:12:23.092 ], 00:12:23.092 "driver_specific": { 00:12:23.092 "passthru": { 00:12:23.092 "name": "pt1", 00:12:23.092 "base_bdev_name": "malloc1" 00:12:23.092 } 00:12:23.092 } 00:12:23.092 }' 00:12:23.092 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:12:23.092 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:12:23.092 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:12:23.092 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:12:23.092 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:12:23.092 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:12:23.092 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:12:23.092 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:12:23.092 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:12:23.352 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:12:23.352 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:12:23.352 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:12:23.352 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:12:23.352 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:12:23.352 11:23:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:12:23.611 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:12:23.611 "name": "pt2", 00:12:23.611 "aliases": [ 00:12:23.611 "00000000-0000-0000-0000-000000000002" 00:12:23.611 ], 00:12:23.611 "product_name": "passthru", 00:12:23.611 "block_size": 512, 00:12:23.611 "num_blocks": 65536, 00:12:23.611 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:23.611 "assigned_rate_limits": { 00:12:23.611 "rw_ios_per_sec": 0, 00:12:23.611 "rw_mbytes_per_sec": 0, 00:12:23.611 "r_mbytes_per_sec": 0, 00:12:23.611 "w_mbytes_per_sec": 0 00:12:23.611 }, 00:12:23.611 "claimed": true, 00:12:23.611 "claim_type": "exclusive_write", 00:12:23.611 "zoned": false, 00:12:23.611 "supported_io_types": { 00:12:23.611 "read": true, 00:12:23.611 "write": true, 00:12:23.611 "unmap": true, 00:12:23.611 "flush": true, 00:12:23.611 "reset": true, 00:12:23.611 "nvme_admin": false, 00:12:23.611 "nvme_io": false, 00:12:23.611 "nvme_io_md": false, 00:12:23.611 "write_zeroes": true, 00:12:23.611 "zcopy": true, 00:12:23.611 "get_zone_info": false, 00:12:23.611 "zone_management": false, 00:12:23.611 "zone_append": false, 00:12:23.611 "compare": false, 00:12:23.611 "compare_and_write": false, 00:12:23.611 "abort": true, 00:12:23.611 "seek_hole": false, 00:12:23.611 "seek_data": false, 00:12:23.611 "copy": true, 00:12:23.611 "nvme_iov_md": false 00:12:23.611 }, 00:12:23.611 "memory_domains": [ 00:12:23.611 { 00:12:23.611 "dma_device_id": "system", 00:12:23.611 "dma_device_type": 1 00:12:23.611 }, 00:12:23.611 { 00:12:23.611 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:23.611 "dma_device_type": 2 00:12:23.611 } 00:12:23.611 ], 00:12:23.611 "driver_specific": { 00:12:23.611 "passthru": { 00:12:23.611 "name": "pt2", 00:12:23.611 "base_bdev_name": "malloc2" 00:12:23.611 } 00:12:23.611 } 00:12:23.611 }' 00:12:23.611 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:12:23.611 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:12:23.611 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:12:23.611 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:12:23.611 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:12:23.869 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:12:23.869 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:12:23.869 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:12:23.869 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:12:23.869 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:12:23.869 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:12:23.869 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:12:23.869 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:12:23.869 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # jq -r '.[] | .uuid' 00:12:24.128 [2024-07-15 11:23:07.621993] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:24.128 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # '[' 1c485329-fed8-4d07-b586-e5b1a4fce65b '!=' 1c485329-fed8-4d07-b586-e5b1a4fce65b ']' 00:12:24.128 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@490 -- # has_redundancy concat 00:12:24.128 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:12:24.128 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@215 -- # return 1 00:12:24.128 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@562 -- # killprocess 871570 00:12:24.128 11:23:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@948 -- # '[' -z 871570 ']' 00:12:24.128 11:23:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@952 -- # kill -0 871570 00:12:24.128 11:23:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@953 -- # uname 00:12:24.128 11:23:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:12:24.128 11:23:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 871570 00:12:24.128 11:23:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:12:24.128 11:23:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:12:24.128 11:23:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 871570' 00:12:24.128 killing process with pid 871570 00:12:24.128 11:23:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@967 -- # kill 871570 00:12:24.128 [2024-07-15 11:23:07.685514] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:24.128 [2024-07-15 11:23:07.685570] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:24.128 [2024-07-15 11:23:07.685611] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:24.128 [2024-07-15 11:23:07.685623] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x156aec0 name raid_bdev1, state offline 00:12:24.128 11:23:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # wait 871570 00:12:24.128 [2024-07-15 11:23:07.702604] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:24.387 11:23:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@564 -- # return 0 00:12:24.387 00:12:24.387 real 0m11.583s 00:12:24.387 user 0m20.719s 00:12:24.387 sys 0m2.126s 00:12:24.387 11:23:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:12:24.387 11:23:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:24.387 ************************************ 00:12:24.387 END TEST raid_superblock_test 00:12:24.387 ************************************ 00:12:24.387 11:23:07 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:12:24.387 11:23:07 bdev_raid -- bdev/bdev_raid.sh@870 -- # run_test raid_read_error_test raid_io_error_test concat 2 read 00:12:24.387 11:23:07 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:12:24.387 11:23:07 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:12:24.387 11:23:07 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:24.645 ************************************ 00:12:24.645 START TEST raid_read_error_test 00:12:24.645 ************************************ 00:12:24.645 11:23:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1123 -- # raid_io_error_test concat 2 read 00:12:24.646 11:23:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@788 -- # local raid_level=concat 00:12:24.646 11:23:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@789 -- # local num_base_bdevs=2 00:12:24.646 11:23:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local error_io_type=read 00:12:24.646 11:23:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i = 1 )) 00:12:24.646 11:23:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:12:24.646 11:23:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev1 00:12:24.646 11:23:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:12:24.646 11:23:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:12:24.646 11:23:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev2 00:12:24.646 11:23:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:12:24.646 11:23:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local base_bdevs 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local raid_bdev_name=raid_bdev1 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local strip_size 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local create_arg 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local bdevperf_log 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local fail_per_s 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # '[' concat '!=' raid1 ']' 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@799 -- # strip_size=64 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # create_arg+=' -z 64' 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@805 -- # mktemp -p /raidtest 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@805 -- # bdevperf_log=/raidtest/tmp.UhPBkLZdBS 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@808 -- # raid_pid=873212 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # waitforlisten 873212 /var/tmp/spdk-raid.sock 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@829 -- # '[' -z 873212 ']' 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:12:24.646 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:12:24.646 11:23:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:24.646 [2024-07-15 11:23:08.072028] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:12:24.646 [2024-07-15 11:23:08.072094] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid873212 ] 00:12:24.646 [2024-07-15 11:23:08.203426] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:24.904 [2024-07-15 11:23:08.311575] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:12:24.904 [2024-07-15 11:23:08.378379] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:24.904 [2024-07-15 11:23:08.378423] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:25.470 11:23:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:12:25.470 11:23:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@862 -- # return 0 00:12:25.470 11:23:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:12:25.470 11:23:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:25.727 BaseBdev1_malloc 00:12:25.727 11:23:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev1_malloc 00:12:25.985 true 00:12:25.985 11:23:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:12:26.243 [2024-07-15 11:23:09.710146] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:12:26.243 [2024-07-15 11:23:09.710191] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:26.243 [2024-07-15 11:23:09.710212] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xceb0d0 00:12:26.243 [2024-07-15 11:23:09.710225] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:26.243 [2024-07-15 11:23:09.712100] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:26.243 [2024-07-15 11:23:09.712130] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:26.243 BaseBdev1 00:12:26.243 11:23:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:12:26.243 11:23:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:26.500 BaseBdev2_malloc 00:12:26.500 11:23:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev2_malloc 00:12:26.757 true 00:12:26.757 11:23:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:12:27.015 [2024-07-15 11:23:10.441874] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:12:27.015 [2024-07-15 11:23:10.441931] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:27.015 [2024-07-15 11:23:10.441951] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xcef910 00:12:27.015 [2024-07-15 11:23:10.441964] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:27.015 [2024-07-15 11:23:10.443462] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:27.015 [2024-07-15 11:23:10.443490] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:27.015 BaseBdev2 00:12:27.015 11:23:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@819 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'BaseBdev1 BaseBdev2' -n raid_bdev1 -s 00:12:27.273 [2024-07-15 11:23:10.690569] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:27.273 [2024-07-15 11:23:10.691816] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:27.273 [2024-07-15 11:23:10.692008] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xcf1320 00:12:27.273 [2024-07-15 11:23:10.692021] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:12:27.273 [2024-07-15 11:23:10.692206] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xcf2290 00:12:27.273 [2024-07-15 11:23:10.692351] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xcf1320 00:12:27.273 [2024-07-15 11:23:10.692361] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0xcf1320 00:12:27.273 [2024-07-15 11:23:10.692463] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:27.273 11:23:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@820 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:12:27.273 11:23:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:12:27.273 11:23:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:12:27.273 11:23:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:12:27.273 11:23:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:12:27.273 11:23:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:12:27.273 11:23:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:27.273 11:23:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:27.273 11:23:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:27.273 11:23:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:27.273 11:23:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:27.273 11:23:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:27.531 11:23:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:27.531 "name": "raid_bdev1", 00:12:27.531 "uuid": "b4ca7d7b-7deb-45c9-b662-8ad6ed11f2b7", 00:12:27.531 "strip_size_kb": 64, 00:12:27.531 "state": "online", 00:12:27.531 "raid_level": "concat", 00:12:27.531 "superblock": true, 00:12:27.531 "num_base_bdevs": 2, 00:12:27.531 "num_base_bdevs_discovered": 2, 00:12:27.531 "num_base_bdevs_operational": 2, 00:12:27.531 "base_bdevs_list": [ 00:12:27.531 { 00:12:27.531 "name": "BaseBdev1", 00:12:27.531 "uuid": "9f981dab-c984-5a0f-a7be-114593f005ff", 00:12:27.531 "is_configured": true, 00:12:27.531 "data_offset": 2048, 00:12:27.531 "data_size": 63488 00:12:27.531 }, 00:12:27.531 { 00:12:27.531 "name": "BaseBdev2", 00:12:27.531 "uuid": "04fedf9a-11f9-53f4-90d5-96dba4603f89", 00:12:27.531 "is_configured": true, 00:12:27.531 "data_offset": 2048, 00:12:27.531 "data_size": 63488 00:12:27.531 } 00:12:27.531 ] 00:12:27.531 }' 00:12:27.531 11:23:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:27.531 11:23:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:28.096 11:23:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@824 -- # sleep 1 00:12:28.096 11:23:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@823 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:12:28.096 [2024-07-15 11:23:11.669439] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xcec9b0 00:12:29.031 11:23:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@827 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:12:29.289 11:23:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # local expected_num_base_bdevs 00:12:29.289 11:23:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@830 -- # [[ concat = \r\a\i\d\1 ]] 00:12:29.289 11:23:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=2 00:12:29.289 11:23:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:12:29.289 11:23:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:12:29.289 11:23:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:12:29.289 11:23:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:12:29.289 11:23:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:12:29.289 11:23:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:12:29.289 11:23:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:29.289 11:23:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:29.289 11:23:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:29.289 11:23:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:29.289 11:23:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:29.289 11:23:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:29.546 11:23:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:29.546 "name": "raid_bdev1", 00:12:29.546 "uuid": "b4ca7d7b-7deb-45c9-b662-8ad6ed11f2b7", 00:12:29.546 "strip_size_kb": 64, 00:12:29.546 "state": "online", 00:12:29.546 "raid_level": "concat", 00:12:29.546 "superblock": true, 00:12:29.546 "num_base_bdevs": 2, 00:12:29.546 "num_base_bdevs_discovered": 2, 00:12:29.546 "num_base_bdevs_operational": 2, 00:12:29.546 "base_bdevs_list": [ 00:12:29.546 { 00:12:29.546 "name": "BaseBdev1", 00:12:29.546 "uuid": "9f981dab-c984-5a0f-a7be-114593f005ff", 00:12:29.546 "is_configured": true, 00:12:29.546 "data_offset": 2048, 00:12:29.546 "data_size": 63488 00:12:29.546 }, 00:12:29.546 { 00:12:29.546 "name": "BaseBdev2", 00:12:29.546 "uuid": "04fedf9a-11f9-53f4-90d5-96dba4603f89", 00:12:29.546 "is_configured": true, 00:12:29.546 "data_offset": 2048, 00:12:29.546 "data_size": 63488 00:12:29.546 } 00:12:29.546 ] 00:12:29.546 }' 00:12:29.546 11:23:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:29.546 11:23:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:30.111 11:23:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:12:30.369 [2024-07-15 11:23:13.813518] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:30.369 [2024-07-15 11:23:13.813556] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:30.369 [2024-07-15 11:23:13.816726] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:30.369 [2024-07-15 11:23:13.816759] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:30.369 [2024-07-15 11:23:13.816788] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:30.369 [2024-07-15 11:23:13.816800] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xcf1320 name raid_bdev1, state offline 00:12:30.369 0 00:12:30.369 11:23:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # killprocess 873212 00:12:30.369 11:23:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@948 -- # '[' -z 873212 ']' 00:12:30.369 11:23:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@952 -- # kill -0 873212 00:12:30.369 11:23:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@953 -- # uname 00:12:30.369 11:23:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:12:30.369 11:23:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 873212 00:12:30.369 11:23:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:12:30.369 11:23:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:12:30.369 11:23:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 873212' 00:12:30.369 killing process with pid 873212 00:12:30.369 11:23:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@967 -- # kill 873212 00:12:30.369 [2024-07-15 11:23:13.885414] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:30.369 11:23:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # wait 873212 00:12:30.369 [2024-07-15 11:23:13.897631] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:30.627 11:23:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # grep -v Job /raidtest/tmp.UhPBkLZdBS 00:12:30.627 11:23:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # grep raid_bdev1 00:12:30.627 11:23:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # awk '{print $6}' 00:12:30.627 11:23:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # fail_per_s=0.47 00:12:30.627 11:23:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@844 -- # has_redundancy concat 00:12:30.627 11:23:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:12:30.627 11:23:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@215 -- # return 1 00:12:30.627 11:23:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.47 != \0\.\0\0 ]] 00:12:30.627 00:12:30.627 real 0m6.145s 00:12:30.627 user 0m9.583s 00:12:30.627 sys 0m1.075s 00:12:30.627 11:23:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:12:30.627 11:23:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:30.627 ************************************ 00:12:30.627 END TEST raid_read_error_test 00:12:30.627 ************************************ 00:12:30.627 11:23:14 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:12:30.627 11:23:14 bdev_raid -- bdev/bdev_raid.sh@871 -- # run_test raid_write_error_test raid_io_error_test concat 2 write 00:12:30.627 11:23:14 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:12:30.627 11:23:14 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:12:30.627 11:23:14 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:30.627 ************************************ 00:12:30.627 START TEST raid_write_error_test 00:12:30.627 ************************************ 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1123 -- # raid_io_error_test concat 2 write 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@788 -- # local raid_level=concat 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@789 -- # local num_base_bdevs=2 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local error_io_type=write 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i = 1 )) 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev1 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev2 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local base_bdevs 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local raid_bdev_name=raid_bdev1 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local strip_size 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local create_arg 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local bdevperf_log 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local fail_per_s 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # '[' concat '!=' raid1 ']' 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@799 -- # strip_size=64 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # create_arg+=' -z 64' 00:12:30.627 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@805 -- # mktemp -p /raidtest 00:12:30.886 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@805 -- # bdevperf_log=/raidtest/tmp.9MSebf8RB9 00:12:30.886 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@808 -- # raid_pid=874181 00:12:30.886 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # waitforlisten 874181 /var/tmp/spdk-raid.sock 00:12:30.886 11:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:12:30.886 11:23:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@829 -- # '[' -z 874181 ']' 00:12:30.886 11:23:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:12:30.886 11:23:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:12:30.886 11:23:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:12:30.886 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:12:30.886 11:23:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:12:30.886 11:23:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:30.886 [2024-07-15 11:23:14.284126] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:12:30.886 [2024-07-15 11:23:14.284190] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid874181 ] 00:12:30.886 [2024-07-15 11:23:14.411449] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:31.144 [2024-07-15 11:23:14.518572] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:12:31.144 [2024-07-15 11:23:14.587835] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:31.144 [2024-07-15 11:23:14.587872] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:31.709 11:23:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:12:31.709 11:23:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@862 -- # return 0 00:12:31.709 11:23:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:12:31.709 11:23:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:31.967 BaseBdev1_malloc 00:12:31.967 11:23:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev1_malloc 00:12:32.225 true 00:12:32.225 11:23:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:12:32.482 [2024-07-15 11:23:15.916143] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:12:32.482 [2024-07-15 11:23:15.916187] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:32.482 [2024-07-15 11:23:15.916210] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xda40d0 00:12:32.482 [2024-07-15 11:23:15.916223] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:32.482 [2024-07-15 11:23:15.918110] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:32.482 [2024-07-15 11:23:15.918145] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:32.482 BaseBdev1 00:12:32.482 11:23:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:12:32.482 11:23:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:32.739 BaseBdev2_malloc 00:12:32.739 11:23:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev2_malloc 00:12:32.997 true 00:12:32.997 11:23:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:12:33.253 [2024-07-15 11:23:16.643880] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:12:33.253 [2024-07-15 11:23:16.643923] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:33.253 [2024-07-15 11:23:16.643951] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xda8910 00:12:33.253 [2024-07-15 11:23:16.643964] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:33.253 [2024-07-15 11:23:16.645571] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:33.253 [2024-07-15 11:23:16.645598] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:33.253 BaseBdev2 00:12:33.253 11:23:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@819 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'BaseBdev1 BaseBdev2' -n raid_bdev1 -s 00:12:33.511 [2024-07-15 11:23:16.888561] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:33.511 [2024-07-15 11:23:16.889814] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:33.511 [2024-07-15 11:23:16.890008] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xdaa320 00:12:33.511 [2024-07-15 11:23:16.890022] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:12:33.511 [2024-07-15 11:23:16.890206] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xdab290 00:12:33.511 [2024-07-15 11:23:16.890347] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xdaa320 00:12:33.511 [2024-07-15 11:23:16.890357] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0xdaa320 00:12:33.511 [2024-07-15 11:23:16.890457] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:33.511 11:23:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@820 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:12:33.511 11:23:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:12:33.511 11:23:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:12:33.511 11:23:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:12:33.511 11:23:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:12:33.511 11:23:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:12:33.511 11:23:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:33.511 11:23:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:33.511 11:23:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:33.511 11:23:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:33.511 11:23:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:33.511 11:23:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:33.813 11:23:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:33.813 "name": "raid_bdev1", 00:12:33.813 "uuid": "4dab0dfd-59e1-4696-954f-4779104081e3", 00:12:33.813 "strip_size_kb": 64, 00:12:33.813 "state": "online", 00:12:33.813 "raid_level": "concat", 00:12:33.813 "superblock": true, 00:12:33.813 "num_base_bdevs": 2, 00:12:33.813 "num_base_bdevs_discovered": 2, 00:12:33.813 "num_base_bdevs_operational": 2, 00:12:33.813 "base_bdevs_list": [ 00:12:33.813 { 00:12:33.813 "name": "BaseBdev1", 00:12:33.813 "uuid": "28084683-d936-514c-aa3b-157b73fa2bc8", 00:12:33.813 "is_configured": true, 00:12:33.813 "data_offset": 2048, 00:12:33.813 "data_size": 63488 00:12:33.813 }, 00:12:33.813 { 00:12:33.813 "name": "BaseBdev2", 00:12:33.813 "uuid": "3d821208-4492-5b24-94dc-d7a1a90b4a24", 00:12:33.813 "is_configured": true, 00:12:33.813 "data_offset": 2048, 00:12:33.813 "data_size": 63488 00:12:33.813 } 00:12:33.813 ] 00:12:33.813 }' 00:12:33.813 11:23:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:33.813 11:23:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:34.395 11:23:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@824 -- # sleep 1 00:12:34.395 11:23:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@823 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:12:34.395 [2024-07-15 11:23:17.871452] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xda59b0 00:12:35.328 11:23:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@827 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:12:35.586 11:23:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # local expected_num_base_bdevs 00:12:35.586 11:23:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@830 -- # [[ concat = \r\a\i\d\1 ]] 00:12:35.586 11:23:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=2 00:12:35.586 11:23:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:12:35.586 11:23:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:12:35.586 11:23:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:12:35.586 11:23:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:12:35.586 11:23:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:12:35.586 11:23:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:12:35.586 11:23:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:35.586 11:23:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:35.586 11:23:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:35.586 11:23:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:35.586 11:23:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:35.586 11:23:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:35.843 11:23:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:35.843 "name": "raid_bdev1", 00:12:35.843 "uuid": "4dab0dfd-59e1-4696-954f-4779104081e3", 00:12:35.843 "strip_size_kb": 64, 00:12:35.843 "state": "online", 00:12:35.843 "raid_level": "concat", 00:12:35.843 "superblock": true, 00:12:35.843 "num_base_bdevs": 2, 00:12:35.843 "num_base_bdevs_discovered": 2, 00:12:35.843 "num_base_bdevs_operational": 2, 00:12:35.843 "base_bdevs_list": [ 00:12:35.843 { 00:12:35.844 "name": "BaseBdev1", 00:12:35.844 "uuid": "28084683-d936-514c-aa3b-157b73fa2bc8", 00:12:35.844 "is_configured": true, 00:12:35.844 "data_offset": 2048, 00:12:35.844 "data_size": 63488 00:12:35.844 }, 00:12:35.844 { 00:12:35.844 "name": "BaseBdev2", 00:12:35.844 "uuid": "3d821208-4492-5b24-94dc-d7a1a90b4a24", 00:12:35.844 "is_configured": true, 00:12:35.844 "data_offset": 2048, 00:12:35.844 "data_size": 63488 00:12:35.844 } 00:12:35.844 ] 00:12:35.844 }' 00:12:35.844 11:23:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:35.844 11:23:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:36.409 11:23:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:12:36.667 [2024-07-15 11:23:20.044667] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:36.667 [2024-07-15 11:23:20.044704] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:36.667 [2024-07-15 11:23:20.047865] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:36.667 [2024-07-15 11:23:20.047896] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:36.667 [2024-07-15 11:23:20.047929] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:36.667 [2024-07-15 11:23:20.047941] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xdaa320 name raid_bdev1, state offline 00:12:36.667 0 00:12:36.667 11:23:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # killprocess 874181 00:12:36.667 11:23:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@948 -- # '[' -z 874181 ']' 00:12:36.667 11:23:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@952 -- # kill -0 874181 00:12:36.667 11:23:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@953 -- # uname 00:12:36.667 11:23:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:12:36.667 11:23:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 874181 00:12:36.667 11:23:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:12:36.667 11:23:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:12:36.667 11:23:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 874181' 00:12:36.667 killing process with pid 874181 00:12:36.667 11:23:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@967 -- # kill 874181 00:12:36.667 [2024-07-15 11:23:20.113130] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:36.667 11:23:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # wait 874181 00:12:36.667 [2024-07-15 11:23:20.123560] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:36.925 11:23:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # grep -v Job /raidtest/tmp.9MSebf8RB9 00:12:36.925 11:23:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # grep raid_bdev1 00:12:36.925 11:23:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # awk '{print $6}' 00:12:36.925 11:23:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # fail_per_s=0.46 00:12:36.925 11:23:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@844 -- # has_redundancy concat 00:12:36.925 11:23:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:12:36.925 11:23:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@215 -- # return 1 00:12:36.925 11:23:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.46 != \0\.\0\0 ]] 00:12:36.925 00:12:36.925 real 0m6.148s 00:12:36.925 user 0m9.629s 00:12:36.925 sys 0m1.080s 00:12:36.925 11:23:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:12:36.925 11:23:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:36.925 ************************************ 00:12:36.925 END TEST raid_write_error_test 00:12:36.925 ************************************ 00:12:36.925 11:23:20 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:12:36.925 11:23:20 bdev_raid -- bdev/bdev_raid.sh@866 -- # for level in raid0 concat raid1 00:12:36.925 11:23:20 bdev_raid -- bdev/bdev_raid.sh@867 -- # run_test raid_state_function_test raid_state_function_test raid1 2 false 00:12:36.925 11:23:20 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:12:36.925 11:23:20 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:12:36.925 11:23:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:36.925 ************************************ 00:12:36.925 START TEST raid_state_function_test 00:12:36.925 ************************************ 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1123 -- # raid_state_function_test raid1 2 false 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@220 -- # local raid_level=raid1 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=2 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # local superblock=false 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # local strip_size 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # '[' raid1 '!=' raid1 ']' 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@234 -- # strip_size=0 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # '[' false = true ']' 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@240 -- # superblock_create_arg= 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # raid_pid=875048 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 875048' 00:12:36.925 Process raid pid: 875048 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@246 -- # waitforlisten 875048 /var/tmp/spdk-raid.sock 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@829 -- # '[' -z 875048 ']' 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:12:36.925 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:12:36.925 11:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:36.925 [2024-07-15 11:23:20.517567] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:12:36.925 [2024-07-15 11:23:20.517639] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:37.183 [2024-07-15 11:23:20.650064] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:37.183 [2024-07-15 11:23:20.751194] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:12:37.441 [2024-07-15 11:23:20.815818] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:37.441 [2024-07-15 11:23:20.815854] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:38.005 11:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:12:38.005 11:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@862 -- # return 0 00:12:38.005 11:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:12:38.264 [2024-07-15 11:23:21.603126] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:12:38.264 [2024-07-15 11:23:21.603168] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:12:38.264 [2024-07-15 11:23:21.603179] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:12:38.264 [2024-07-15 11:23:21.603191] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:12:38.264 11:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:12:38.264 11:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:12:38.264 11:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:12:38.264 11:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:12:38.264 11:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:12:38.264 11:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:12:38.264 11:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:38.264 11:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:38.264 11:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:38.264 11:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:38.264 11:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:38.264 11:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:38.264 11:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:38.264 "name": "Existed_Raid", 00:12:38.264 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:38.264 "strip_size_kb": 0, 00:12:38.264 "state": "configuring", 00:12:38.264 "raid_level": "raid1", 00:12:38.264 "superblock": false, 00:12:38.264 "num_base_bdevs": 2, 00:12:38.264 "num_base_bdevs_discovered": 0, 00:12:38.264 "num_base_bdevs_operational": 2, 00:12:38.264 "base_bdevs_list": [ 00:12:38.264 { 00:12:38.264 "name": "BaseBdev1", 00:12:38.264 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:38.264 "is_configured": false, 00:12:38.264 "data_offset": 0, 00:12:38.264 "data_size": 0 00:12:38.264 }, 00:12:38.264 { 00:12:38.264 "name": "BaseBdev2", 00:12:38.264 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:38.264 "is_configured": false, 00:12:38.264 "data_offset": 0, 00:12:38.264 "data_size": 0 00:12:38.264 } 00:12:38.264 ] 00:12:38.264 }' 00:12:38.264 11:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:38.264 11:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:38.830 11:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:12:39.088 [2024-07-15 11:23:22.529486] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:12:39.088 [2024-07-15 11:23:22.529516] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1717a80 name Existed_Raid, state configuring 00:12:39.088 11:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:12:39.346 [2024-07-15 11:23:22.774132] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:12:39.346 [2024-07-15 11:23:22.774162] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:12:39.346 [2024-07-15 11:23:22.774173] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:12:39.346 [2024-07-15 11:23:22.774184] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:12:39.346 11:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:12:39.604 [2024-07-15 11:23:23.024678] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:39.604 BaseBdev1 00:12:39.604 11:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:12:39.604 11:23:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:12:39.604 11:23:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:12:39.604 11:23:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:12:39.604 11:23:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:12:39.604 11:23:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:12:39.604 11:23:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:12:39.861 11:23:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:12:39.861 [ 00:12:39.861 { 00:12:39.861 "name": "BaseBdev1", 00:12:39.861 "aliases": [ 00:12:39.861 "eef29721-bd44-402d-ab87-d711a8621925" 00:12:39.861 ], 00:12:39.861 "product_name": "Malloc disk", 00:12:39.861 "block_size": 512, 00:12:39.861 "num_blocks": 65536, 00:12:39.861 "uuid": "eef29721-bd44-402d-ab87-d711a8621925", 00:12:39.861 "assigned_rate_limits": { 00:12:39.861 "rw_ios_per_sec": 0, 00:12:39.861 "rw_mbytes_per_sec": 0, 00:12:39.861 "r_mbytes_per_sec": 0, 00:12:39.861 "w_mbytes_per_sec": 0 00:12:39.861 }, 00:12:39.861 "claimed": true, 00:12:39.861 "claim_type": "exclusive_write", 00:12:39.861 "zoned": false, 00:12:39.861 "supported_io_types": { 00:12:39.861 "read": true, 00:12:39.861 "write": true, 00:12:39.861 "unmap": true, 00:12:39.861 "flush": true, 00:12:39.861 "reset": true, 00:12:39.861 "nvme_admin": false, 00:12:39.861 "nvme_io": false, 00:12:39.861 "nvme_io_md": false, 00:12:39.861 "write_zeroes": true, 00:12:39.861 "zcopy": true, 00:12:39.861 "get_zone_info": false, 00:12:39.861 "zone_management": false, 00:12:39.861 "zone_append": false, 00:12:39.861 "compare": false, 00:12:39.861 "compare_and_write": false, 00:12:39.861 "abort": true, 00:12:39.861 "seek_hole": false, 00:12:39.861 "seek_data": false, 00:12:39.861 "copy": true, 00:12:39.861 "nvme_iov_md": false 00:12:39.861 }, 00:12:39.861 "memory_domains": [ 00:12:39.861 { 00:12:39.861 "dma_device_id": "system", 00:12:39.861 "dma_device_type": 1 00:12:39.861 }, 00:12:39.861 { 00:12:39.861 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:39.861 "dma_device_type": 2 00:12:39.861 } 00:12:39.861 ], 00:12:39.861 "driver_specific": {} 00:12:39.861 } 00:12:39.861 ] 00:12:40.119 11:23:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:12:40.119 11:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:12:40.119 11:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:12:40.119 11:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:12:40.119 11:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:12:40.119 11:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:12:40.119 11:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:12:40.119 11:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:40.119 11:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:40.119 11:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:40.119 11:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:40.119 11:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:40.119 11:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:40.119 11:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:40.119 "name": "Existed_Raid", 00:12:40.119 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:40.119 "strip_size_kb": 0, 00:12:40.119 "state": "configuring", 00:12:40.119 "raid_level": "raid1", 00:12:40.119 "superblock": false, 00:12:40.119 "num_base_bdevs": 2, 00:12:40.119 "num_base_bdevs_discovered": 1, 00:12:40.119 "num_base_bdevs_operational": 2, 00:12:40.119 "base_bdevs_list": [ 00:12:40.119 { 00:12:40.119 "name": "BaseBdev1", 00:12:40.119 "uuid": "eef29721-bd44-402d-ab87-d711a8621925", 00:12:40.119 "is_configured": true, 00:12:40.119 "data_offset": 0, 00:12:40.119 "data_size": 65536 00:12:40.119 }, 00:12:40.119 { 00:12:40.119 "name": "BaseBdev2", 00:12:40.119 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:40.119 "is_configured": false, 00:12:40.119 "data_offset": 0, 00:12:40.119 "data_size": 0 00:12:40.119 } 00:12:40.119 ] 00:12:40.119 }' 00:12:40.119 11:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:40.119 11:23:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:40.685 11:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:12:40.943 [2024-07-15 11:23:24.476533] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:12:40.943 [2024-07-15 11:23:24.476572] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1717350 name Existed_Raid, state configuring 00:12:40.943 11:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:12:41.202 [2024-07-15 11:23:24.733257] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:41.202 [2024-07-15 11:23:24.734802] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:12:41.202 [2024-07-15 11:23:24.734839] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:12:41.202 11:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:12:41.202 11:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:12:41.202 11:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:12:41.202 11:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:12:41.202 11:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:12:41.202 11:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:12:41.202 11:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:12:41.202 11:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:12:41.202 11:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:41.202 11:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:41.202 11:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:41.202 11:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:41.202 11:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:41.202 11:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:41.460 11:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:41.460 "name": "Existed_Raid", 00:12:41.460 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:41.460 "strip_size_kb": 0, 00:12:41.460 "state": "configuring", 00:12:41.460 "raid_level": "raid1", 00:12:41.460 "superblock": false, 00:12:41.460 "num_base_bdevs": 2, 00:12:41.460 "num_base_bdevs_discovered": 1, 00:12:41.460 "num_base_bdevs_operational": 2, 00:12:41.460 "base_bdevs_list": [ 00:12:41.460 { 00:12:41.460 "name": "BaseBdev1", 00:12:41.460 "uuid": "eef29721-bd44-402d-ab87-d711a8621925", 00:12:41.460 "is_configured": true, 00:12:41.460 "data_offset": 0, 00:12:41.460 "data_size": 65536 00:12:41.460 }, 00:12:41.460 { 00:12:41.460 "name": "BaseBdev2", 00:12:41.460 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:41.460 "is_configured": false, 00:12:41.460 "data_offset": 0, 00:12:41.460 "data_size": 0 00:12:41.460 } 00:12:41.460 ] 00:12:41.460 }' 00:12:41.460 11:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:41.460 11:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:42.025 11:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:12:42.283 [2024-07-15 11:23:25.811545] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:42.283 [2024-07-15 11:23:25.811586] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1718000 00:12:42.283 [2024-07-15 11:23:25.811595] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:12:42.283 [2024-07-15 11:23:25.811788] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x16320c0 00:12:42.283 [2024-07-15 11:23:25.811904] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1718000 00:12:42.283 [2024-07-15 11:23:25.811914] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x1718000 00:12:42.283 [2024-07-15 11:23:25.812094] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:42.283 BaseBdev2 00:12:42.283 11:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:12:42.283 11:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:12:42.283 11:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:12:42.283 11:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:12:42.283 11:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:12:42.283 11:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:12:42.283 11:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:12:42.541 11:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:12:42.799 [ 00:12:42.799 { 00:12:42.799 "name": "BaseBdev2", 00:12:42.799 "aliases": [ 00:12:42.799 "9e74ea1d-26b5-4c41-9416-2cca97cfc2de" 00:12:42.799 ], 00:12:42.799 "product_name": "Malloc disk", 00:12:42.799 "block_size": 512, 00:12:42.799 "num_blocks": 65536, 00:12:42.799 "uuid": "9e74ea1d-26b5-4c41-9416-2cca97cfc2de", 00:12:42.799 "assigned_rate_limits": { 00:12:42.799 "rw_ios_per_sec": 0, 00:12:42.799 "rw_mbytes_per_sec": 0, 00:12:42.799 "r_mbytes_per_sec": 0, 00:12:42.799 "w_mbytes_per_sec": 0 00:12:42.799 }, 00:12:42.799 "claimed": true, 00:12:42.799 "claim_type": "exclusive_write", 00:12:42.799 "zoned": false, 00:12:42.799 "supported_io_types": { 00:12:42.799 "read": true, 00:12:42.799 "write": true, 00:12:42.799 "unmap": true, 00:12:42.799 "flush": true, 00:12:42.799 "reset": true, 00:12:42.799 "nvme_admin": false, 00:12:42.799 "nvme_io": false, 00:12:42.799 "nvme_io_md": false, 00:12:42.799 "write_zeroes": true, 00:12:42.799 "zcopy": true, 00:12:42.799 "get_zone_info": false, 00:12:42.799 "zone_management": false, 00:12:42.799 "zone_append": false, 00:12:42.799 "compare": false, 00:12:42.799 "compare_and_write": false, 00:12:42.799 "abort": true, 00:12:42.799 "seek_hole": false, 00:12:42.799 "seek_data": false, 00:12:42.799 "copy": true, 00:12:42.799 "nvme_iov_md": false 00:12:42.799 }, 00:12:42.799 "memory_domains": [ 00:12:42.799 { 00:12:42.799 "dma_device_id": "system", 00:12:42.799 "dma_device_type": 1 00:12:42.799 }, 00:12:42.799 { 00:12:42.799 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:42.799 "dma_device_type": 2 00:12:42.799 } 00:12:42.799 ], 00:12:42.799 "driver_specific": {} 00:12:42.799 } 00:12:42.799 ] 00:12:42.799 11:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:12:42.799 11:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:12:42.799 11:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:12:42.799 11:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:12:42.799 11:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:12:42.799 11:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:12:42.799 11:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:12:42.799 11:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:12:42.799 11:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:12:42.799 11:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:42.799 11:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:42.799 11:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:42.799 11:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:42.799 11:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:42.799 11:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:43.057 11:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:43.057 "name": "Existed_Raid", 00:12:43.057 "uuid": "f1d9f262-c068-4a04-af17-c73d4546230a", 00:12:43.057 "strip_size_kb": 0, 00:12:43.057 "state": "online", 00:12:43.057 "raid_level": "raid1", 00:12:43.058 "superblock": false, 00:12:43.058 "num_base_bdevs": 2, 00:12:43.058 "num_base_bdevs_discovered": 2, 00:12:43.058 "num_base_bdevs_operational": 2, 00:12:43.058 "base_bdevs_list": [ 00:12:43.058 { 00:12:43.058 "name": "BaseBdev1", 00:12:43.058 "uuid": "eef29721-bd44-402d-ab87-d711a8621925", 00:12:43.058 "is_configured": true, 00:12:43.058 "data_offset": 0, 00:12:43.058 "data_size": 65536 00:12:43.058 }, 00:12:43.058 { 00:12:43.058 "name": "BaseBdev2", 00:12:43.058 "uuid": "9e74ea1d-26b5-4c41-9416-2cca97cfc2de", 00:12:43.058 "is_configured": true, 00:12:43.058 "data_offset": 0, 00:12:43.058 "data_size": 65536 00:12:43.058 } 00:12:43.058 ] 00:12:43.058 }' 00:12:43.058 11:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:43.058 11:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:43.623 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:12:43.623 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:12:43.624 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:12:43.624 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:12:43.624 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:12:43.624 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # local name 00:12:43.624 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:12:43.624 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:12:43.883 [2024-07-15 11:23:27.243646] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:43.883 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:12:43.883 "name": "Existed_Raid", 00:12:43.883 "aliases": [ 00:12:43.883 "f1d9f262-c068-4a04-af17-c73d4546230a" 00:12:43.883 ], 00:12:43.883 "product_name": "Raid Volume", 00:12:43.883 "block_size": 512, 00:12:43.883 "num_blocks": 65536, 00:12:43.883 "uuid": "f1d9f262-c068-4a04-af17-c73d4546230a", 00:12:43.883 "assigned_rate_limits": { 00:12:43.883 "rw_ios_per_sec": 0, 00:12:43.883 "rw_mbytes_per_sec": 0, 00:12:43.883 "r_mbytes_per_sec": 0, 00:12:43.883 "w_mbytes_per_sec": 0 00:12:43.883 }, 00:12:43.883 "claimed": false, 00:12:43.883 "zoned": false, 00:12:43.883 "supported_io_types": { 00:12:43.883 "read": true, 00:12:43.883 "write": true, 00:12:43.883 "unmap": false, 00:12:43.883 "flush": false, 00:12:43.883 "reset": true, 00:12:43.883 "nvme_admin": false, 00:12:43.883 "nvme_io": false, 00:12:43.883 "nvme_io_md": false, 00:12:43.883 "write_zeroes": true, 00:12:43.883 "zcopy": false, 00:12:43.883 "get_zone_info": false, 00:12:43.883 "zone_management": false, 00:12:43.883 "zone_append": false, 00:12:43.883 "compare": false, 00:12:43.883 "compare_and_write": false, 00:12:43.883 "abort": false, 00:12:43.883 "seek_hole": false, 00:12:43.883 "seek_data": false, 00:12:43.883 "copy": false, 00:12:43.883 "nvme_iov_md": false 00:12:43.883 }, 00:12:43.883 "memory_domains": [ 00:12:43.883 { 00:12:43.883 "dma_device_id": "system", 00:12:43.883 "dma_device_type": 1 00:12:43.883 }, 00:12:43.883 { 00:12:43.883 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:43.883 "dma_device_type": 2 00:12:43.883 }, 00:12:43.883 { 00:12:43.883 "dma_device_id": "system", 00:12:43.883 "dma_device_type": 1 00:12:43.883 }, 00:12:43.883 { 00:12:43.883 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:43.883 "dma_device_type": 2 00:12:43.883 } 00:12:43.883 ], 00:12:43.883 "driver_specific": { 00:12:43.883 "raid": { 00:12:43.883 "uuid": "f1d9f262-c068-4a04-af17-c73d4546230a", 00:12:43.883 "strip_size_kb": 0, 00:12:43.883 "state": "online", 00:12:43.883 "raid_level": "raid1", 00:12:43.883 "superblock": false, 00:12:43.883 "num_base_bdevs": 2, 00:12:43.883 "num_base_bdevs_discovered": 2, 00:12:43.883 "num_base_bdevs_operational": 2, 00:12:43.883 "base_bdevs_list": [ 00:12:43.883 { 00:12:43.883 "name": "BaseBdev1", 00:12:43.883 "uuid": "eef29721-bd44-402d-ab87-d711a8621925", 00:12:43.883 "is_configured": true, 00:12:43.883 "data_offset": 0, 00:12:43.883 "data_size": 65536 00:12:43.883 }, 00:12:43.883 { 00:12:43.883 "name": "BaseBdev2", 00:12:43.883 "uuid": "9e74ea1d-26b5-4c41-9416-2cca97cfc2de", 00:12:43.883 "is_configured": true, 00:12:43.883 "data_offset": 0, 00:12:43.883 "data_size": 65536 00:12:43.883 } 00:12:43.883 ] 00:12:43.883 } 00:12:43.883 } 00:12:43.883 }' 00:12:43.883 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:12:43.883 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:12:43.883 BaseBdev2' 00:12:43.883 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:12:43.883 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:12:43.883 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:12:44.142 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:12:44.142 "name": "BaseBdev1", 00:12:44.142 "aliases": [ 00:12:44.142 "eef29721-bd44-402d-ab87-d711a8621925" 00:12:44.142 ], 00:12:44.142 "product_name": "Malloc disk", 00:12:44.142 "block_size": 512, 00:12:44.142 "num_blocks": 65536, 00:12:44.142 "uuid": "eef29721-bd44-402d-ab87-d711a8621925", 00:12:44.142 "assigned_rate_limits": { 00:12:44.142 "rw_ios_per_sec": 0, 00:12:44.142 "rw_mbytes_per_sec": 0, 00:12:44.142 "r_mbytes_per_sec": 0, 00:12:44.142 "w_mbytes_per_sec": 0 00:12:44.142 }, 00:12:44.142 "claimed": true, 00:12:44.142 "claim_type": "exclusive_write", 00:12:44.142 "zoned": false, 00:12:44.142 "supported_io_types": { 00:12:44.142 "read": true, 00:12:44.142 "write": true, 00:12:44.142 "unmap": true, 00:12:44.142 "flush": true, 00:12:44.142 "reset": true, 00:12:44.142 "nvme_admin": false, 00:12:44.142 "nvme_io": false, 00:12:44.142 "nvme_io_md": false, 00:12:44.142 "write_zeroes": true, 00:12:44.142 "zcopy": true, 00:12:44.142 "get_zone_info": false, 00:12:44.142 "zone_management": false, 00:12:44.142 "zone_append": false, 00:12:44.142 "compare": false, 00:12:44.142 "compare_and_write": false, 00:12:44.142 "abort": true, 00:12:44.142 "seek_hole": false, 00:12:44.142 "seek_data": false, 00:12:44.142 "copy": true, 00:12:44.142 "nvme_iov_md": false 00:12:44.142 }, 00:12:44.142 "memory_domains": [ 00:12:44.142 { 00:12:44.142 "dma_device_id": "system", 00:12:44.142 "dma_device_type": 1 00:12:44.142 }, 00:12:44.142 { 00:12:44.142 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:44.142 "dma_device_type": 2 00:12:44.142 } 00:12:44.142 ], 00:12:44.142 "driver_specific": {} 00:12:44.142 }' 00:12:44.142 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:12:44.142 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:12:44.142 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:12:44.142 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:12:44.142 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:12:44.142 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:12:44.142 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:12:44.400 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:12:44.400 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:12:44.400 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:12:44.400 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:12:44.400 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:12:44.400 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:12:44.400 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:12:44.400 11:23:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:12:44.659 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:12:44.659 "name": "BaseBdev2", 00:12:44.659 "aliases": [ 00:12:44.659 "9e74ea1d-26b5-4c41-9416-2cca97cfc2de" 00:12:44.659 ], 00:12:44.659 "product_name": "Malloc disk", 00:12:44.659 "block_size": 512, 00:12:44.659 "num_blocks": 65536, 00:12:44.659 "uuid": "9e74ea1d-26b5-4c41-9416-2cca97cfc2de", 00:12:44.659 "assigned_rate_limits": { 00:12:44.659 "rw_ios_per_sec": 0, 00:12:44.659 "rw_mbytes_per_sec": 0, 00:12:44.659 "r_mbytes_per_sec": 0, 00:12:44.659 "w_mbytes_per_sec": 0 00:12:44.659 }, 00:12:44.659 "claimed": true, 00:12:44.659 "claim_type": "exclusive_write", 00:12:44.659 "zoned": false, 00:12:44.659 "supported_io_types": { 00:12:44.659 "read": true, 00:12:44.659 "write": true, 00:12:44.659 "unmap": true, 00:12:44.659 "flush": true, 00:12:44.659 "reset": true, 00:12:44.659 "nvme_admin": false, 00:12:44.659 "nvme_io": false, 00:12:44.659 "nvme_io_md": false, 00:12:44.659 "write_zeroes": true, 00:12:44.659 "zcopy": true, 00:12:44.659 "get_zone_info": false, 00:12:44.659 "zone_management": false, 00:12:44.659 "zone_append": false, 00:12:44.659 "compare": false, 00:12:44.659 "compare_and_write": false, 00:12:44.659 "abort": true, 00:12:44.659 "seek_hole": false, 00:12:44.659 "seek_data": false, 00:12:44.659 "copy": true, 00:12:44.659 "nvme_iov_md": false 00:12:44.659 }, 00:12:44.659 "memory_domains": [ 00:12:44.659 { 00:12:44.659 "dma_device_id": "system", 00:12:44.659 "dma_device_type": 1 00:12:44.659 }, 00:12:44.659 { 00:12:44.659 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:44.659 "dma_device_type": 2 00:12:44.659 } 00:12:44.659 ], 00:12:44.659 "driver_specific": {} 00:12:44.659 }' 00:12:44.659 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:12:44.659 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:12:44.659 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:12:44.659 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:12:44.659 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:12:44.659 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:12:44.659 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:12:44.917 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:12:44.917 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:12:44.917 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:12:44.917 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:12:44.917 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:12:44.917 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:12:45.174 [2024-07-15 11:23:28.639133] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:45.174 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@275 -- # local expected_state 00:12:45.174 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # has_redundancy raid1 00:12:45.174 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:12:45.174 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@214 -- # return 0 00:12:45.174 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # expected_state=online 00:12:45.174 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:12:45.174 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:12:45.174 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:12:45.174 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:12:45.174 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:12:45.174 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:12:45.174 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:45.174 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:45.174 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:45.174 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:45.174 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:45.174 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:45.432 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:45.432 "name": "Existed_Raid", 00:12:45.432 "uuid": "f1d9f262-c068-4a04-af17-c73d4546230a", 00:12:45.432 "strip_size_kb": 0, 00:12:45.432 "state": "online", 00:12:45.432 "raid_level": "raid1", 00:12:45.432 "superblock": false, 00:12:45.432 "num_base_bdevs": 2, 00:12:45.432 "num_base_bdevs_discovered": 1, 00:12:45.432 "num_base_bdevs_operational": 1, 00:12:45.432 "base_bdevs_list": [ 00:12:45.432 { 00:12:45.432 "name": null, 00:12:45.432 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:45.432 "is_configured": false, 00:12:45.432 "data_offset": 0, 00:12:45.432 "data_size": 65536 00:12:45.432 }, 00:12:45.432 { 00:12:45.432 "name": "BaseBdev2", 00:12:45.432 "uuid": "9e74ea1d-26b5-4c41-9416-2cca97cfc2de", 00:12:45.432 "is_configured": true, 00:12:45.432 "data_offset": 0, 00:12:45.432 "data_size": 65536 00:12:45.432 } 00:12:45.432 ] 00:12:45.432 }' 00:12:45.432 11:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:45.432 11:23:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:45.998 11:23:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:12:45.998 11:23:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:12:45.998 11:23:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:45.998 11:23:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:12:46.257 11:23:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:12:46.257 11:23:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:12:46.257 11:23:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:12:46.514 [2024-07-15 11:23:29.940485] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:46.514 [2024-07-15 11:23:29.940571] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:46.514 [2024-07-15 11:23:29.953247] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:46.514 [2024-07-15 11:23:29.953287] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:46.514 [2024-07-15 11:23:29.953300] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1718000 name Existed_Raid, state offline 00:12:46.514 11:23:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:12:46.514 11:23:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:12:46.514 11:23:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:46.514 11:23:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:12:46.772 11:23:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:12:46.772 11:23:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:12:46.772 11:23:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # '[' 2 -gt 2 ']' 00:12:46.772 11:23:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@341 -- # killprocess 875048 00:12:46.772 11:23:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@948 -- # '[' -z 875048 ']' 00:12:46.772 11:23:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@952 -- # kill -0 875048 00:12:46.772 11:23:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@953 -- # uname 00:12:46.772 11:23:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:12:46.772 11:23:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 875048 00:12:46.772 11:23:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:12:46.772 11:23:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:12:46.772 11:23:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 875048' 00:12:46.772 killing process with pid 875048 00:12:46.772 11:23:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@967 -- # kill 875048 00:12:46.772 [2024-07-15 11:23:30.268721] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:46.772 11:23:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # wait 875048 00:12:46.772 [2024-07-15 11:23:30.269731] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@343 -- # return 0 00:12:47.030 00:12:47.030 real 0m10.042s 00:12:47.030 user 0m17.748s 00:12:47.030 sys 0m1.944s 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:12:47.030 ************************************ 00:12:47.030 END TEST raid_state_function_test 00:12:47.030 ************************************ 00:12:47.030 11:23:30 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:12:47.030 11:23:30 bdev_raid -- bdev/bdev_raid.sh@868 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 2 true 00:12:47.030 11:23:30 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:12:47.030 11:23:30 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:12:47.030 11:23:30 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:47.030 ************************************ 00:12:47.030 START TEST raid_state_function_test_sb 00:12:47.030 ************************************ 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1123 -- # raid_state_function_test raid1 2 true 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@220 -- # local raid_level=raid1 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=2 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # local superblock=true 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # local strip_size 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # '[' raid1 '!=' raid1 ']' 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@234 -- # strip_size=0 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # '[' true = true ']' 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@238 -- # superblock_create_arg=-s 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # raid_pid=876619 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 876619' 00:12:47.030 Process raid pid: 876619 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@246 -- # waitforlisten 876619 /var/tmp/spdk-raid.sock 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@829 -- # '[' -z 876619 ']' 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@834 -- # local max_retries=100 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:12:47.030 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # xtrace_disable 00:12:47.030 11:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:47.288 [2024-07-15 11:23:30.637941] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:12:47.288 [2024-07-15 11:23:30.638010] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:47.288 [2024-07-15 11:23:30.766175] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:47.288 [2024-07-15 11:23:30.868025] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:12:47.546 [2024-07-15 11:23:30.932837] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:47.546 [2024-07-15 11:23:30.932869] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:48.111 11:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:12:48.111 11:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@862 -- # return 0 00:12:48.111 11:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:12:48.368 [2024-07-15 11:23:31.792351] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:12:48.368 [2024-07-15 11:23:31.792393] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:12:48.368 [2024-07-15 11:23:31.792405] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:12:48.368 [2024-07-15 11:23:31.792416] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:12:48.368 11:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:12:48.368 11:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:12:48.368 11:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:12:48.368 11:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:12:48.368 11:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:12:48.368 11:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:12:48.368 11:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:48.368 11:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:48.368 11:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:48.368 11:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:48.368 11:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:48.368 11:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:48.624 11:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:48.624 "name": "Existed_Raid", 00:12:48.624 "uuid": "9058f2a3-1e58-4333-acba-d379d40eb27a", 00:12:48.624 "strip_size_kb": 0, 00:12:48.624 "state": "configuring", 00:12:48.624 "raid_level": "raid1", 00:12:48.624 "superblock": true, 00:12:48.624 "num_base_bdevs": 2, 00:12:48.624 "num_base_bdevs_discovered": 0, 00:12:48.624 "num_base_bdevs_operational": 2, 00:12:48.624 "base_bdevs_list": [ 00:12:48.624 { 00:12:48.624 "name": "BaseBdev1", 00:12:48.624 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:48.624 "is_configured": false, 00:12:48.624 "data_offset": 0, 00:12:48.624 "data_size": 0 00:12:48.624 }, 00:12:48.624 { 00:12:48.624 "name": "BaseBdev2", 00:12:48.624 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:48.624 "is_configured": false, 00:12:48.624 "data_offset": 0, 00:12:48.624 "data_size": 0 00:12:48.624 } 00:12:48.624 ] 00:12:48.624 }' 00:12:48.624 11:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:48.624 11:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.187 11:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:12:49.443 [2024-07-15 11:23:32.863027] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:12:49.443 [2024-07-15 11:23:32.863058] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x261fa80 name Existed_Raid, state configuring 00:12:49.443 11:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:12:49.699 [2024-07-15 11:23:33.039515] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:12:49.699 [2024-07-15 11:23:33.039541] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:12:49.699 [2024-07-15 11:23:33.039551] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:12:49.699 [2024-07-15 11:23:33.039562] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:12:49.699 11:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:12:49.699 [2024-07-15 11:23:33.221921] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:49.699 BaseBdev1 00:12:49.699 11:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:12:49.699 11:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:12:49.699 11:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:12:49.699 11:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:12:49.699 11:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:12:49.699 11:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:12:49.699 11:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:12:49.956 11:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:12:50.213 [ 00:12:50.213 { 00:12:50.213 "name": "BaseBdev1", 00:12:50.213 "aliases": [ 00:12:50.213 "d6eea670-af4c-411a-bc6d-6019a3f65c5b" 00:12:50.213 ], 00:12:50.213 "product_name": "Malloc disk", 00:12:50.213 "block_size": 512, 00:12:50.213 "num_blocks": 65536, 00:12:50.213 "uuid": "d6eea670-af4c-411a-bc6d-6019a3f65c5b", 00:12:50.213 "assigned_rate_limits": { 00:12:50.213 "rw_ios_per_sec": 0, 00:12:50.213 "rw_mbytes_per_sec": 0, 00:12:50.213 "r_mbytes_per_sec": 0, 00:12:50.213 "w_mbytes_per_sec": 0 00:12:50.213 }, 00:12:50.213 "claimed": true, 00:12:50.213 "claim_type": "exclusive_write", 00:12:50.213 "zoned": false, 00:12:50.213 "supported_io_types": { 00:12:50.213 "read": true, 00:12:50.213 "write": true, 00:12:50.213 "unmap": true, 00:12:50.213 "flush": true, 00:12:50.213 "reset": true, 00:12:50.213 "nvme_admin": false, 00:12:50.213 "nvme_io": false, 00:12:50.213 "nvme_io_md": false, 00:12:50.213 "write_zeroes": true, 00:12:50.213 "zcopy": true, 00:12:50.213 "get_zone_info": false, 00:12:50.213 "zone_management": false, 00:12:50.213 "zone_append": false, 00:12:50.213 "compare": false, 00:12:50.213 "compare_and_write": false, 00:12:50.213 "abort": true, 00:12:50.213 "seek_hole": false, 00:12:50.213 "seek_data": false, 00:12:50.213 "copy": true, 00:12:50.213 "nvme_iov_md": false 00:12:50.213 }, 00:12:50.213 "memory_domains": [ 00:12:50.213 { 00:12:50.213 "dma_device_id": "system", 00:12:50.213 "dma_device_type": 1 00:12:50.213 }, 00:12:50.213 { 00:12:50.213 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:50.213 "dma_device_type": 2 00:12:50.213 } 00:12:50.213 ], 00:12:50.213 "driver_specific": {} 00:12:50.213 } 00:12:50.213 ] 00:12:50.213 11:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:12:50.213 11:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:12:50.213 11:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:12:50.213 11:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:12:50.213 11:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:12:50.213 11:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:12:50.213 11:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:12:50.213 11:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:50.213 11:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:50.213 11:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:50.213 11:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:50.213 11:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:50.213 11:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:50.213 11:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:50.213 "name": "Existed_Raid", 00:12:50.213 "uuid": "1bf54050-9a43-4ac1-a04c-ced2760162b5", 00:12:50.213 "strip_size_kb": 0, 00:12:50.213 "state": "configuring", 00:12:50.213 "raid_level": "raid1", 00:12:50.213 "superblock": true, 00:12:50.213 "num_base_bdevs": 2, 00:12:50.213 "num_base_bdevs_discovered": 1, 00:12:50.213 "num_base_bdevs_operational": 2, 00:12:50.213 "base_bdevs_list": [ 00:12:50.213 { 00:12:50.213 "name": "BaseBdev1", 00:12:50.213 "uuid": "d6eea670-af4c-411a-bc6d-6019a3f65c5b", 00:12:50.213 "is_configured": true, 00:12:50.213 "data_offset": 2048, 00:12:50.213 "data_size": 63488 00:12:50.213 }, 00:12:50.213 { 00:12:50.213 "name": "BaseBdev2", 00:12:50.213 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:50.213 "is_configured": false, 00:12:50.213 "data_offset": 0, 00:12:50.213 "data_size": 0 00:12:50.213 } 00:12:50.213 ] 00:12:50.213 }' 00:12:50.213 11:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:50.213 11:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:50.794 11:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:12:51.067 [2024-07-15 11:23:34.505334] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:12:51.067 [2024-07-15 11:23:34.505367] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x261f350 name Existed_Raid, state configuring 00:12:51.067 11:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:12:51.325 [2024-07-15 11:23:34.681840] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:51.325 [2024-07-15 11:23:34.683328] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:12:51.325 [2024-07-15 11:23:34.683366] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:12:51.325 11:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:12:51.325 11:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:12:51.325 11:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:12:51.325 11:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:12:51.325 11:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:12:51.325 11:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:12:51.325 11:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:12:51.325 11:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:12:51.325 11:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:51.325 11:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:51.325 11:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:51.325 11:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:51.325 11:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:51.325 11:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:51.325 11:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:51.325 "name": "Existed_Raid", 00:12:51.325 "uuid": "2f96c4fa-c6a7-4931-88c1-176155439294", 00:12:51.325 "strip_size_kb": 0, 00:12:51.325 "state": "configuring", 00:12:51.325 "raid_level": "raid1", 00:12:51.325 "superblock": true, 00:12:51.325 "num_base_bdevs": 2, 00:12:51.325 "num_base_bdevs_discovered": 1, 00:12:51.325 "num_base_bdevs_operational": 2, 00:12:51.325 "base_bdevs_list": [ 00:12:51.325 { 00:12:51.325 "name": "BaseBdev1", 00:12:51.325 "uuid": "d6eea670-af4c-411a-bc6d-6019a3f65c5b", 00:12:51.325 "is_configured": true, 00:12:51.325 "data_offset": 2048, 00:12:51.325 "data_size": 63488 00:12:51.325 }, 00:12:51.325 { 00:12:51.325 "name": "BaseBdev2", 00:12:51.325 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:51.325 "is_configured": false, 00:12:51.325 "data_offset": 0, 00:12:51.325 "data_size": 0 00:12:51.325 } 00:12:51.325 ] 00:12:51.325 }' 00:12:51.325 11:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:51.325 11:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:51.891 11:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:12:52.457 [2024-07-15 11:23:35.893584] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:52.457 [2024-07-15 11:23:35.893745] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x2620000 00:12:52.457 [2024-07-15 11:23:35.893759] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:52.457 [2024-07-15 11:23:35.893943] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x253a0c0 00:12:52.457 [2024-07-15 11:23:35.894071] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x2620000 00:12:52.457 [2024-07-15 11:23:35.894082] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x2620000 00:12:52.457 [2024-07-15 11:23:35.894172] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:52.457 BaseBdev2 00:12:52.457 11:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:12:52.457 11:23:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:12:52.457 11:23:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:12:52.457 11:23:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:12:52.457 11:23:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:12:52.457 11:23:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:12:52.457 11:23:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:12:52.715 11:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:12:52.973 [ 00:12:52.973 { 00:12:52.973 "name": "BaseBdev2", 00:12:52.973 "aliases": [ 00:12:52.973 "d9c473f2-2910-4059-9351-358f02dfcc35" 00:12:52.973 ], 00:12:52.973 "product_name": "Malloc disk", 00:12:52.973 "block_size": 512, 00:12:52.973 "num_blocks": 65536, 00:12:52.973 "uuid": "d9c473f2-2910-4059-9351-358f02dfcc35", 00:12:52.973 "assigned_rate_limits": { 00:12:52.973 "rw_ios_per_sec": 0, 00:12:52.973 "rw_mbytes_per_sec": 0, 00:12:52.973 "r_mbytes_per_sec": 0, 00:12:52.973 "w_mbytes_per_sec": 0 00:12:52.973 }, 00:12:52.973 "claimed": true, 00:12:52.973 "claim_type": "exclusive_write", 00:12:52.973 "zoned": false, 00:12:52.973 "supported_io_types": { 00:12:52.973 "read": true, 00:12:52.973 "write": true, 00:12:52.973 "unmap": true, 00:12:52.973 "flush": true, 00:12:52.973 "reset": true, 00:12:52.973 "nvme_admin": false, 00:12:52.973 "nvme_io": false, 00:12:52.973 "nvme_io_md": false, 00:12:52.973 "write_zeroes": true, 00:12:52.973 "zcopy": true, 00:12:52.973 "get_zone_info": false, 00:12:52.973 "zone_management": false, 00:12:52.973 "zone_append": false, 00:12:52.973 "compare": false, 00:12:52.973 "compare_and_write": false, 00:12:52.973 "abort": true, 00:12:52.973 "seek_hole": false, 00:12:52.973 "seek_data": false, 00:12:52.973 "copy": true, 00:12:52.973 "nvme_iov_md": false 00:12:52.973 }, 00:12:52.973 "memory_domains": [ 00:12:52.973 { 00:12:52.973 "dma_device_id": "system", 00:12:52.973 "dma_device_type": 1 00:12:52.973 }, 00:12:52.973 { 00:12:52.973 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:52.973 "dma_device_type": 2 00:12:52.973 } 00:12:52.973 ], 00:12:52.973 "driver_specific": {} 00:12:52.973 } 00:12:52.973 ] 00:12:52.973 11:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:12:52.973 11:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:12:52.973 11:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:12:52.973 11:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:12:52.973 11:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:12:52.973 11:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:12:52.973 11:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:12:52.973 11:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:12:52.973 11:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:12:52.973 11:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:52.973 11:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:52.973 11:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:52.973 11:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:52.973 11:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:52.973 11:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:53.231 11:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:53.231 "name": "Existed_Raid", 00:12:53.231 "uuid": "2f96c4fa-c6a7-4931-88c1-176155439294", 00:12:53.231 "strip_size_kb": 0, 00:12:53.231 "state": "online", 00:12:53.231 "raid_level": "raid1", 00:12:53.231 "superblock": true, 00:12:53.231 "num_base_bdevs": 2, 00:12:53.231 "num_base_bdevs_discovered": 2, 00:12:53.231 "num_base_bdevs_operational": 2, 00:12:53.231 "base_bdevs_list": [ 00:12:53.231 { 00:12:53.231 "name": "BaseBdev1", 00:12:53.231 "uuid": "d6eea670-af4c-411a-bc6d-6019a3f65c5b", 00:12:53.231 "is_configured": true, 00:12:53.231 "data_offset": 2048, 00:12:53.231 "data_size": 63488 00:12:53.231 }, 00:12:53.231 { 00:12:53.231 "name": "BaseBdev2", 00:12:53.231 "uuid": "d9c473f2-2910-4059-9351-358f02dfcc35", 00:12:53.231 "is_configured": true, 00:12:53.231 "data_offset": 2048, 00:12:53.231 "data_size": 63488 00:12:53.231 } 00:12:53.231 ] 00:12:53.231 }' 00:12:53.231 11:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:53.231 11:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.796 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:12:53.796 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:12:53.796 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:12:53.796 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:12:53.796 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:12:53.796 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # local name 00:12:53.796 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:12:53.796 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:12:53.796 [2024-07-15 11:23:37.361742] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:53.796 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:12:53.796 "name": "Existed_Raid", 00:12:53.796 "aliases": [ 00:12:53.796 "2f96c4fa-c6a7-4931-88c1-176155439294" 00:12:53.796 ], 00:12:53.796 "product_name": "Raid Volume", 00:12:53.796 "block_size": 512, 00:12:53.796 "num_blocks": 63488, 00:12:53.796 "uuid": "2f96c4fa-c6a7-4931-88c1-176155439294", 00:12:53.796 "assigned_rate_limits": { 00:12:53.796 "rw_ios_per_sec": 0, 00:12:53.796 "rw_mbytes_per_sec": 0, 00:12:53.796 "r_mbytes_per_sec": 0, 00:12:53.796 "w_mbytes_per_sec": 0 00:12:53.796 }, 00:12:53.796 "claimed": false, 00:12:53.796 "zoned": false, 00:12:53.796 "supported_io_types": { 00:12:53.796 "read": true, 00:12:53.796 "write": true, 00:12:53.796 "unmap": false, 00:12:53.796 "flush": false, 00:12:53.796 "reset": true, 00:12:53.796 "nvme_admin": false, 00:12:53.796 "nvme_io": false, 00:12:53.796 "nvme_io_md": false, 00:12:53.796 "write_zeroes": true, 00:12:53.796 "zcopy": false, 00:12:53.796 "get_zone_info": false, 00:12:53.796 "zone_management": false, 00:12:53.796 "zone_append": false, 00:12:53.796 "compare": false, 00:12:53.796 "compare_and_write": false, 00:12:53.796 "abort": false, 00:12:53.796 "seek_hole": false, 00:12:53.796 "seek_data": false, 00:12:53.796 "copy": false, 00:12:53.796 "nvme_iov_md": false 00:12:53.796 }, 00:12:53.796 "memory_domains": [ 00:12:53.796 { 00:12:53.796 "dma_device_id": "system", 00:12:53.796 "dma_device_type": 1 00:12:53.796 }, 00:12:53.796 { 00:12:53.796 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:53.796 "dma_device_type": 2 00:12:53.796 }, 00:12:53.796 { 00:12:53.796 "dma_device_id": "system", 00:12:53.796 "dma_device_type": 1 00:12:53.796 }, 00:12:53.796 { 00:12:53.796 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:53.796 "dma_device_type": 2 00:12:53.796 } 00:12:53.796 ], 00:12:53.796 "driver_specific": { 00:12:53.796 "raid": { 00:12:53.796 "uuid": "2f96c4fa-c6a7-4931-88c1-176155439294", 00:12:53.796 "strip_size_kb": 0, 00:12:53.796 "state": "online", 00:12:53.796 "raid_level": "raid1", 00:12:53.796 "superblock": true, 00:12:53.796 "num_base_bdevs": 2, 00:12:53.796 "num_base_bdevs_discovered": 2, 00:12:53.796 "num_base_bdevs_operational": 2, 00:12:53.796 "base_bdevs_list": [ 00:12:53.796 { 00:12:53.796 "name": "BaseBdev1", 00:12:53.796 "uuid": "d6eea670-af4c-411a-bc6d-6019a3f65c5b", 00:12:53.796 "is_configured": true, 00:12:53.796 "data_offset": 2048, 00:12:53.796 "data_size": 63488 00:12:53.796 }, 00:12:53.796 { 00:12:53.796 "name": "BaseBdev2", 00:12:53.796 "uuid": "d9c473f2-2910-4059-9351-358f02dfcc35", 00:12:53.796 "is_configured": true, 00:12:53.796 "data_offset": 2048, 00:12:53.796 "data_size": 63488 00:12:53.796 } 00:12:53.796 ] 00:12:53.796 } 00:12:53.796 } 00:12:53.796 }' 00:12:53.796 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:12:54.054 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:12:54.054 BaseBdev2' 00:12:54.054 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:12:54.054 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:12:54.054 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:12:54.311 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:12:54.311 "name": "BaseBdev1", 00:12:54.311 "aliases": [ 00:12:54.311 "d6eea670-af4c-411a-bc6d-6019a3f65c5b" 00:12:54.311 ], 00:12:54.311 "product_name": "Malloc disk", 00:12:54.311 "block_size": 512, 00:12:54.311 "num_blocks": 65536, 00:12:54.311 "uuid": "d6eea670-af4c-411a-bc6d-6019a3f65c5b", 00:12:54.311 "assigned_rate_limits": { 00:12:54.311 "rw_ios_per_sec": 0, 00:12:54.311 "rw_mbytes_per_sec": 0, 00:12:54.311 "r_mbytes_per_sec": 0, 00:12:54.311 "w_mbytes_per_sec": 0 00:12:54.311 }, 00:12:54.311 "claimed": true, 00:12:54.311 "claim_type": "exclusive_write", 00:12:54.311 "zoned": false, 00:12:54.311 "supported_io_types": { 00:12:54.311 "read": true, 00:12:54.311 "write": true, 00:12:54.311 "unmap": true, 00:12:54.311 "flush": true, 00:12:54.311 "reset": true, 00:12:54.311 "nvme_admin": false, 00:12:54.311 "nvme_io": false, 00:12:54.311 "nvme_io_md": false, 00:12:54.311 "write_zeroes": true, 00:12:54.311 "zcopy": true, 00:12:54.311 "get_zone_info": false, 00:12:54.311 "zone_management": false, 00:12:54.311 "zone_append": false, 00:12:54.311 "compare": false, 00:12:54.311 "compare_and_write": false, 00:12:54.311 "abort": true, 00:12:54.311 "seek_hole": false, 00:12:54.311 "seek_data": false, 00:12:54.311 "copy": true, 00:12:54.311 "nvme_iov_md": false 00:12:54.311 }, 00:12:54.311 "memory_domains": [ 00:12:54.311 { 00:12:54.311 "dma_device_id": "system", 00:12:54.311 "dma_device_type": 1 00:12:54.311 }, 00:12:54.311 { 00:12:54.311 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:54.311 "dma_device_type": 2 00:12:54.312 } 00:12:54.312 ], 00:12:54.312 "driver_specific": {} 00:12:54.312 }' 00:12:54.312 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:12:54.312 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:12:54.312 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:12:54.312 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:12:54.312 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:12:54.312 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:12:54.312 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:12:54.312 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:12:54.569 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:12:54.569 11:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:12:54.569 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:12:54.569 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:12:54.569 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:12:54.569 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:12:54.569 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:12:54.826 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:12:54.826 "name": "BaseBdev2", 00:12:54.826 "aliases": [ 00:12:54.826 "d9c473f2-2910-4059-9351-358f02dfcc35" 00:12:54.826 ], 00:12:54.826 "product_name": "Malloc disk", 00:12:54.826 "block_size": 512, 00:12:54.826 "num_blocks": 65536, 00:12:54.826 "uuid": "d9c473f2-2910-4059-9351-358f02dfcc35", 00:12:54.826 "assigned_rate_limits": { 00:12:54.826 "rw_ios_per_sec": 0, 00:12:54.826 "rw_mbytes_per_sec": 0, 00:12:54.826 "r_mbytes_per_sec": 0, 00:12:54.826 "w_mbytes_per_sec": 0 00:12:54.826 }, 00:12:54.826 "claimed": true, 00:12:54.826 "claim_type": "exclusive_write", 00:12:54.826 "zoned": false, 00:12:54.826 "supported_io_types": { 00:12:54.826 "read": true, 00:12:54.826 "write": true, 00:12:54.826 "unmap": true, 00:12:54.826 "flush": true, 00:12:54.826 "reset": true, 00:12:54.826 "nvme_admin": false, 00:12:54.826 "nvme_io": false, 00:12:54.826 "nvme_io_md": false, 00:12:54.826 "write_zeroes": true, 00:12:54.826 "zcopy": true, 00:12:54.826 "get_zone_info": false, 00:12:54.826 "zone_management": false, 00:12:54.826 "zone_append": false, 00:12:54.826 "compare": false, 00:12:54.826 "compare_and_write": false, 00:12:54.826 "abort": true, 00:12:54.826 "seek_hole": false, 00:12:54.826 "seek_data": false, 00:12:54.826 "copy": true, 00:12:54.826 "nvme_iov_md": false 00:12:54.826 }, 00:12:54.826 "memory_domains": [ 00:12:54.826 { 00:12:54.826 "dma_device_id": "system", 00:12:54.826 "dma_device_type": 1 00:12:54.826 }, 00:12:54.826 { 00:12:54.826 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:54.826 "dma_device_type": 2 00:12:54.826 } 00:12:54.826 ], 00:12:54.826 "driver_specific": {} 00:12:54.826 }' 00:12:54.826 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:12:54.826 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:12:54.826 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:12:54.826 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:12:54.826 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:12:54.826 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:12:54.826 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:12:54.826 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:12:55.084 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:12:55.084 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:12:55.084 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:12:55.084 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:12:55.084 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:12:55.342 [2024-07-15 11:23:38.709119] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:55.342 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@275 -- # local expected_state 00:12:55.342 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # has_redundancy raid1 00:12:55.342 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # case $1 in 00:12:55.342 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@214 -- # return 0 00:12:55.342 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # expected_state=online 00:12:55.342 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:12:55.342 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:12:55.342 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:12:55.342 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:12:55.342 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:12:55.342 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:12:55.342 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:55.342 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:55.342 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:55.342 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:55.342 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:55.342 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:55.600 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:55.600 "name": "Existed_Raid", 00:12:55.600 "uuid": "2f96c4fa-c6a7-4931-88c1-176155439294", 00:12:55.600 "strip_size_kb": 0, 00:12:55.600 "state": "online", 00:12:55.600 "raid_level": "raid1", 00:12:55.600 "superblock": true, 00:12:55.600 "num_base_bdevs": 2, 00:12:55.600 "num_base_bdevs_discovered": 1, 00:12:55.600 "num_base_bdevs_operational": 1, 00:12:55.600 "base_bdevs_list": [ 00:12:55.600 { 00:12:55.600 "name": null, 00:12:55.600 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:55.600 "is_configured": false, 00:12:55.600 "data_offset": 2048, 00:12:55.600 "data_size": 63488 00:12:55.600 }, 00:12:55.600 { 00:12:55.600 "name": "BaseBdev2", 00:12:55.600 "uuid": "d9c473f2-2910-4059-9351-358f02dfcc35", 00:12:55.600 "is_configured": true, 00:12:55.600 "data_offset": 2048, 00:12:55.600 "data_size": 63488 00:12:55.600 } 00:12:55.600 ] 00:12:55.600 }' 00:12:55.600 11:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:55.600 11:23:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:56.164 11:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:12:56.164 11:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:12:56.164 11:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:56.164 11:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:12:56.421 11:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:12:56.421 11:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:12:56.421 11:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:12:56.679 [2024-07-15 11:23:40.074281] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:56.679 [2024-07-15 11:23:40.074366] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:56.679 [2024-07-15 11:23:40.085271] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:56.679 [2024-07-15 11:23:40.085303] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:56.679 [2024-07-15 11:23:40.085316] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x2620000 name Existed_Raid, state offline 00:12:56.679 11:23:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:12:56.679 11:23:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:12:56.679 11:23:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:56.679 11:23:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:12:56.937 11:23:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:12:56.937 11:23:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:12:56.937 11:23:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # '[' 2 -gt 2 ']' 00:12:56.937 11:23:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@341 -- # killprocess 876619 00:12:56.937 11:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@948 -- # '[' -z 876619 ']' 00:12:56.937 11:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@952 -- # kill -0 876619 00:12:56.937 11:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@953 -- # uname 00:12:56.937 11:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:12:56.937 11:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 876619 00:12:56.937 11:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:12:56.937 11:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:12:56.937 11:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@966 -- # echo 'killing process with pid 876619' 00:12:56.937 killing process with pid 876619 00:12:56.937 11:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@967 -- # kill 876619 00:12:56.937 [2024-07-15 11:23:40.423275] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:56.937 11:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # wait 876619 00:12:56.937 [2024-07-15 11:23:40.424193] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:57.195 11:23:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@343 -- # return 0 00:12:57.195 00:12:57.195 real 0m10.080s 00:12:57.195 user 0m17.930s 00:12:57.195 sys 0m1.872s 00:12:57.195 11:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1124 -- # xtrace_disable 00:12:57.195 11:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:57.195 ************************************ 00:12:57.195 END TEST raid_state_function_test_sb 00:12:57.195 ************************************ 00:12:57.195 11:23:40 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:12:57.195 11:23:40 bdev_raid -- bdev/bdev_raid.sh@869 -- # run_test raid_superblock_test raid_superblock_test raid1 2 00:12:57.195 11:23:40 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:12:57.195 11:23:40 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:12:57.195 11:23:40 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:57.195 ************************************ 00:12:57.195 START TEST raid_superblock_test 00:12:57.195 ************************************ 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1123 -- # raid_superblock_test raid1 2 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@392 -- # local raid_level=raid1 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local num_base_bdevs=2 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # base_bdevs_malloc=() 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local base_bdevs_malloc 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_pt=() 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_pt 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt_uuid=() 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt_uuid 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local raid_bdev_name=raid_bdev1 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local strip_size 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size_create_arg 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local raid_bdev_uuid 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@403 -- # '[' raid1 '!=' raid1 ']' 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@407 -- # strip_size=0 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@410 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -L bdev_raid 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # raid_pid=878170 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # waitforlisten 878170 /var/tmp/spdk-raid.sock 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@829 -- # '[' -z 878170 ']' 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:12:57.195 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:12:57.195 11:23:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:57.195 [2024-07-15 11:23:40.780058] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:12:57.195 [2024-07-15 11:23:40.780120] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid878170 ] 00:12:57.453 [2024-07-15 11:23:40.908365] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:57.453 [2024-07-15 11:23:41.016261] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:12:57.712 [2024-07-15 11:23:41.078935] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:57.712 [2024-07-15 11:23:41.078968] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:57.712 11:23:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:12:57.712 11:23:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@862 -- # return 0 00:12:57.712 11:23:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i = 1 )) 00:12:57.712 11:23:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:12:57.712 11:23:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc1 00:12:57.712 11:23:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt1 00:12:57.712 11:23:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:12:57.712 11:23:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:12:57.712 11:23:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:12:57.712 11:23:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:12:57.712 11:23:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc1 00:12:57.970 malloc1 00:12:57.970 11:23:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:12:58.230 [2024-07-15 11:23:41.646596] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:12:58.230 [2024-07-15 11:23:41.646643] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:58.230 [2024-07-15 11:23:41.646667] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xf2b570 00:12:58.230 [2024-07-15 11:23:41.646679] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:58.230 [2024-07-15 11:23:41.648416] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:58.230 [2024-07-15 11:23:41.648446] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:12:58.230 pt1 00:12:58.230 11:23:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:12:58.230 11:23:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:12:58.230 11:23:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc2 00:12:58.230 11:23:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt2 00:12:58.230 11:23:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:12:58.230 11:23:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:12:58.230 11:23:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:12:58.230 11:23:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:12:58.230 11:23:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc2 00:12:58.490 malloc2 00:12:58.490 11:23:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:12:58.748 [2024-07-15 11:23:42.113403] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:12:58.748 [2024-07-15 11:23:42.113449] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:58.748 [2024-07-15 11:23:42.113470] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xf2c970 00:12:58.748 [2024-07-15 11:23:42.113483] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:58.749 [2024-07-15 11:23:42.115132] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:58.749 [2024-07-15 11:23:42.115161] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:12:58.749 pt2 00:12:58.749 11:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:12:58.749 11:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:12:58.749 11:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@429 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'pt1 pt2' -n raid_bdev1 -s 00:12:59.007 [2024-07-15 11:23:42.346031] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:12:59.007 [2024-07-15 11:23:42.347344] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:12:59.008 [2024-07-15 11:23:42.347490] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x10cf270 00:12:59.008 [2024-07-15 11:23:42.347502] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:59.008 [2024-07-15 11:23:42.347702] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xf230e0 00:12:59.008 [2024-07-15 11:23:42.347848] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x10cf270 00:12:59.008 [2024-07-15 11:23:42.347858] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x10cf270 00:12:59.008 [2024-07-15 11:23:42.347968] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:59.008 11:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:59.008 11:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:12:59.008 11:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:12:59.008 11:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:12:59.008 11:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:12:59.008 11:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:12:59.008 11:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:12:59.008 11:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:12:59.008 11:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:12:59.008 11:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:12:59.008 11:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:12:59.008 11:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:59.008 11:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:12:59.008 "name": "raid_bdev1", 00:12:59.008 "uuid": "64691137-5845-4643-9feb-b61cf4017884", 00:12:59.008 "strip_size_kb": 0, 00:12:59.008 "state": "online", 00:12:59.008 "raid_level": "raid1", 00:12:59.008 "superblock": true, 00:12:59.008 "num_base_bdevs": 2, 00:12:59.008 "num_base_bdevs_discovered": 2, 00:12:59.008 "num_base_bdevs_operational": 2, 00:12:59.008 "base_bdevs_list": [ 00:12:59.008 { 00:12:59.008 "name": "pt1", 00:12:59.008 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:59.008 "is_configured": true, 00:12:59.008 "data_offset": 2048, 00:12:59.008 "data_size": 63488 00:12:59.008 }, 00:12:59.008 { 00:12:59.008 "name": "pt2", 00:12:59.008 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:59.008 "is_configured": true, 00:12:59.008 "data_offset": 2048, 00:12:59.008 "data_size": 63488 00:12:59.008 } 00:12:59.008 ] 00:12:59.008 }' 00:12:59.008 11:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:12:59.008 11:23:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:59.943 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_properties raid_bdev1 00:12:59.944 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:12:59.944 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:12:59.944 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:12:59.944 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:12:59.944 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # local name 00:12:59.944 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:12:59.944 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:12:59.944 [2024-07-15 11:23:43.348897] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:59.944 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:12:59.944 "name": "raid_bdev1", 00:12:59.944 "aliases": [ 00:12:59.944 "64691137-5845-4643-9feb-b61cf4017884" 00:12:59.944 ], 00:12:59.944 "product_name": "Raid Volume", 00:12:59.944 "block_size": 512, 00:12:59.944 "num_blocks": 63488, 00:12:59.944 "uuid": "64691137-5845-4643-9feb-b61cf4017884", 00:12:59.944 "assigned_rate_limits": { 00:12:59.944 "rw_ios_per_sec": 0, 00:12:59.944 "rw_mbytes_per_sec": 0, 00:12:59.944 "r_mbytes_per_sec": 0, 00:12:59.944 "w_mbytes_per_sec": 0 00:12:59.944 }, 00:12:59.944 "claimed": false, 00:12:59.944 "zoned": false, 00:12:59.944 "supported_io_types": { 00:12:59.944 "read": true, 00:12:59.944 "write": true, 00:12:59.944 "unmap": false, 00:12:59.944 "flush": false, 00:12:59.944 "reset": true, 00:12:59.944 "nvme_admin": false, 00:12:59.944 "nvme_io": false, 00:12:59.944 "nvme_io_md": false, 00:12:59.944 "write_zeroes": true, 00:12:59.944 "zcopy": false, 00:12:59.944 "get_zone_info": false, 00:12:59.944 "zone_management": false, 00:12:59.944 "zone_append": false, 00:12:59.944 "compare": false, 00:12:59.944 "compare_and_write": false, 00:12:59.944 "abort": false, 00:12:59.944 "seek_hole": false, 00:12:59.944 "seek_data": false, 00:12:59.944 "copy": false, 00:12:59.944 "nvme_iov_md": false 00:12:59.944 }, 00:12:59.944 "memory_domains": [ 00:12:59.944 { 00:12:59.944 "dma_device_id": "system", 00:12:59.944 "dma_device_type": 1 00:12:59.944 }, 00:12:59.944 { 00:12:59.944 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:59.944 "dma_device_type": 2 00:12:59.944 }, 00:12:59.944 { 00:12:59.944 "dma_device_id": "system", 00:12:59.944 "dma_device_type": 1 00:12:59.944 }, 00:12:59.944 { 00:12:59.944 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:59.944 "dma_device_type": 2 00:12:59.944 } 00:12:59.944 ], 00:12:59.944 "driver_specific": { 00:12:59.944 "raid": { 00:12:59.944 "uuid": "64691137-5845-4643-9feb-b61cf4017884", 00:12:59.944 "strip_size_kb": 0, 00:12:59.944 "state": "online", 00:12:59.944 "raid_level": "raid1", 00:12:59.944 "superblock": true, 00:12:59.944 "num_base_bdevs": 2, 00:12:59.944 "num_base_bdevs_discovered": 2, 00:12:59.944 "num_base_bdevs_operational": 2, 00:12:59.944 "base_bdevs_list": [ 00:12:59.944 { 00:12:59.944 "name": "pt1", 00:12:59.944 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:59.944 "is_configured": true, 00:12:59.944 "data_offset": 2048, 00:12:59.944 "data_size": 63488 00:12:59.944 }, 00:12:59.944 { 00:12:59.944 "name": "pt2", 00:12:59.944 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:59.944 "is_configured": true, 00:12:59.944 "data_offset": 2048, 00:12:59.944 "data_size": 63488 00:12:59.944 } 00:12:59.944 ] 00:12:59.944 } 00:12:59.944 } 00:12:59.944 }' 00:12:59.944 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:12:59.944 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:12:59.944 pt2' 00:12:59.944 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:12:59.944 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:12:59.944 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:13:00.202 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:13:00.202 "name": "pt1", 00:13:00.202 "aliases": [ 00:13:00.202 "00000000-0000-0000-0000-000000000001" 00:13:00.202 ], 00:13:00.202 "product_name": "passthru", 00:13:00.202 "block_size": 512, 00:13:00.202 "num_blocks": 65536, 00:13:00.202 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:00.202 "assigned_rate_limits": { 00:13:00.202 "rw_ios_per_sec": 0, 00:13:00.202 "rw_mbytes_per_sec": 0, 00:13:00.202 "r_mbytes_per_sec": 0, 00:13:00.202 "w_mbytes_per_sec": 0 00:13:00.202 }, 00:13:00.202 "claimed": true, 00:13:00.202 "claim_type": "exclusive_write", 00:13:00.202 "zoned": false, 00:13:00.202 "supported_io_types": { 00:13:00.202 "read": true, 00:13:00.202 "write": true, 00:13:00.202 "unmap": true, 00:13:00.202 "flush": true, 00:13:00.202 "reset": true, 00:13:00.202 "nvme_admin": false, 00:13:00.202 "nvme_io": false, 00:13:00.202 "nvme_io_md": false, 00:13:00.202 "write_zeroes": true, 00:13:00.202 "zcopy": true, 00:13:00.202 "get_zone_info": false, 00:13:00.202 "zone_management": false, 00:13:00.202 "zone_append": false, 00:13:00.202 "compare": false, 00:13:00.202 "compare_and_write": false, 00:13:00.202 "abort": true, 00:13:00.202 "seek_hole": false, 00:13:00.202 "seek_data": false, 00:13:00.202 "copy": true, 00:13:00.202 "nvme_iov_md": false 00:13:00.202 }, 00:13:00.202 "memory_domains": [ 00:13:00.202 { 00:13:00.202 "dma_device_id": "system", 00:13:00.202 "dma_device_type": 1 00:13:00.202 }, 00:13:00.202 { 00:13:00.202 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:00.202 "dma_device_type": 2 00:13:00.202 } 00:13:00.202 ], 00:13:00.202 "driver_specific": { 00:13:00.202 "passthru": { 00:13:00.202 "name": "pt1", 00:13:00.202 "base_bdev_name": "malloc1" 00:13:00.202 } 00:13:00.202 } 00:13:00.202 }' 00:13:00.202 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:13:00.202 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:13:00.202 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:13:00.202 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:13:00.202 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:13:00.202 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:13:00.202 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:13:00.460 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:13:00.460 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:13:00.460 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:13:00.460 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:13:00.460 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:13:00.460 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:13:00.460 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:13:00.460 11:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:13:00.718 11:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:13:00.718 "name": "pt2", 00:13:00.718 "aliases": [ 00:13:00.718 "00000000-0000-0000-0000-000000000002" 00:13:00.718 ], 00:13:00.718 "product_name": "passthru", 00:13:00.718 "block_size": 512, 00:13:00.718 "num_blocks": 65536, 00:13:00.718 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:00.718 "assigned_rate_limits": { 00:13:00.718 "rw_ios_per_sec": 0, 00:13:00.718 "rw_mbytes_per_sec": 0, 00:13:00.718 "r_mbytes_per_sec": 0, 00:13:00.718 "w_mbytes_per_sec": 0 00:13:00.718 }, 00:13:00.718 "claimed": true, 00:13:00.718 "claim_type": "exclusive_write", 00:13:00.718 "zoned": false, 00:13:00.718 "supported_io_types": { 00:13:00.718 "read": true, 00:13:00.718 "write": true, 00:13:00.718 "unmap": true, 00:13:00.718 "flush": true, 00:13:00.718 "reset": true, 00:13:00.718 "nvme_admin": false, 00:13:00.718 "nvme_io": false, 00:13:00.718 "nvme_io_md": false, 00:13:00.718 "write_zeroes": true, 00:13:00.718 "zcopy": true, 00:13:00.718 "get_zone_info": false, 00:13:00.718 "zone_management": false, 00:13:00.718 "zone_append": false, 00:13:00.718 "compare": false, 00:13:00.718 "compare_and_write": false, 00:13:00.718 "abort": true, 00:13:00.718 "seek_hole": false, 00:13:00.718 "seek_data": false, 00:13:00.718 "copy": true, 00:13:00.718 "nvme_iov_md": false 00:13:00.718 }, 00:13:00.718 "memory_domains": [ 00:13:00.718 { 00:13:00.718 "dma_device_id": "system", 00:13:00.718 "dma_device_type": 1 00:13:00.718 }, 00:13:00.718 { 00:13:00.718 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:00.718 "dma_device_type": 2 00:13:00.718 } 00:13:00.718 ], 00:13:00.718 "driver_specific": { 00:13:00.718 "passthru": { 00:13:00.718 "name": "pt2", 00:13:00.718 "base_bdev_name": "malloc2" 00:13:00.719 } 00:13:00.719 } 00:13:00.719 }' 00:13:00.719 11:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:13:00.719 11:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:13:00.719 11:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:13:00.719 11:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:13:00.977 11:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:13:00.977 11:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:13:00.977 11:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:13:00.977 11:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:13:00.977 11:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:13:00.977 11:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:13:01.234 11:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:13:01.234 11:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:13:01.234 11:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:13:01.234 11:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # jq -r '.[] | .uuid' 00:13:01.492 [2024-07-15 11:23:44.836853] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:01.492 11:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # raid_bdev_uuid=64691137-5845-4643-9feb-b61cf4017884 00:13:01.492 11:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # '[' -z 64691137-5845-4643-9feb-b61cf4017884 ']' 00:13:01.492 11:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@440 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:13:01.492 [2024-07-15 11:23:45.081246] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:01.492 [2024-07-15 11:23:45.081266] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:01.492 [2024-07-15 11:23:45.081319] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:01.492 [2024-07-15 11:23:45.081375] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:01.492 [2024-07-15 11:23:45.081386] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x10cf270 name raid_bdev1, state offline 00:13:01.750 11:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:01.750 11:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # jq -r '.[]' 00:13:02.316 11:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # raid_bdev= 00:13:02.317 11:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # '[' -n '' ']' 00:13:02.317 11:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:13:02.317 11:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt1 00:13:02.317 11:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:13:02.317 11:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:13:02.575 11:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs 00:13:02.575 11:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:13:03.140 11:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # '[' false == true ']' 00:13:03.140 11:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@456 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'malloc1 malloc2' -n raid_bdev1 00:13:03.140 11:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@648 -- # local es=0 00:13:03.140 11:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'malloc1 malloc2' -n raid_bdev1 00:13:03.140 11:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:13:03.141 11:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:13:03.141 11:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:13:03.141 11:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:13:03.141 11:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:13:03.141 11:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:13:03.141 11:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:13:03.141 11:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py ]] 00:13:03.141 11:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'malloc1 malloc2' -n raid_bdev1 00:13:03.141 [2024-07-15 11:23:46.701442] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:13:03.141 [2024-07-15 11:23:46.702777] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:13:03.141 [2024-07-15 11:23:46.702830] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:13:03.141 [2024-07-15 11:23:46.702869] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:13:03.141 [2024-07-15 11:23:46.702889] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:03.141 [2024-07-15 11:23:46.702899] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x10ceff0 name raid_bdev1, state configuring 00:13:03.141 request: 00:13:03.141 { 00:13:03.141 "name": "raid_bdev1", 00:13:03.141 "raid_level": "raid1", 00:13:03.141 "base_bdevs": [ 00:13:03.141 "malloc1", 00:13:03.141 "malloc2" 00:13:03.141 ], 00:13:03.141 "superblock": false, 00:13:03.141 "method": "bdev_raid_create", 00:13:03.141 "req_id": 1 00:13:03.141 } 00:13:03.141 Got JSON-RPC error response 00:13:03.141 response: 00:13:03.141 { 00:13:03.141 "code": -17, 00:13:03.141 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:13:03.141 } 00:13:03.141 11:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@651 -- # es=1 00:13:03.141 11:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:13:03.141 11:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:13:03.141 11:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:13:03.141 11:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:03.141 11:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # jq -r '.[]' 00:13:03.399 11:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # raid_bdev= 00:13:03.399 11:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # '[' -n '' ']' 00:13:03.399 11:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@464 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:13:03.656 [2024-07-15 11:23:47.122511] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:13:03.656 [2024-07-15 11:23:47.122552] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:03.656 [2024-07-15 11:23:47.122574] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xf2b7a0 00:13:03.656 [2024-07-15 11:23:47.122587] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:03.656 [2024-07-15 11:23:47.124109] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:03.656 [2024-07-15 11:23:47.124136] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:13:03.656 [2024-07-15 11:23:47.124198] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:13:03.656 [2024-07-15 11:23:47.124222] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:13:03.656 pt1 00:13:03.656 11:23:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@467 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:13:03.656 11:23:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:13:03.656 11:23:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:13:03.656 11:23:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:13:03.656 11:23:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:13:03.656 11:23:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:13:03.656 11:23:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:03.656 11:23:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:03.656 11:23:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:03.656 11:23:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:03.656 11:23:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:03.656 11:23:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:03.914 11:23:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:03.914 "name": "raid_bdev1", 00:13:03.914 "uuid": "64691137-5845-4643-9feb-b61cf4017884", 00:13:03.914 "strip_size_kb": 0, 00:13:03.914 "state": "configuring", 00:13:03.914 "raid_level": "raid1", 00:13:03.914 "superblock": true, 00:13:03.914 "num_base_bdevs": 2, 00:13:03.914 "num_base_bdevs_discovered": 1, 00:13:03.914 "num_base_bdevs_operational": 2, 00:13:03.914 "base_bdevs_list": [ 00:13:03.914 { 00:13:03.914 "name": "pt1", 00:13:03.914 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:03.914 "is_configured": true, 00:13:03.914 "data_offset": 2048, 00:13:03.914 "data_size": 63488 00:13:03.914 }, 00:13:03.914 { 00:13:03.914 "name": null, 00:13:03.914 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:03.914 "is_configured": false, 00:13:03.914 "data_offset": 2048, 00:13:03.914 "data_size": 63488 00:13:03.914 } 00:13:03.914 ] 00:13:03.914 }' 00:13:03.914 11:23:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:03.914 11:23:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:04.847 11:23:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@469 -- # '[' 2 -gt 2 ']' 00:13:04.847 11:23:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i = 1 )) 00:13:04.847 11:23:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:13:04.847 11:23:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:05.105 [2024-07-15 11:23:48.466083] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:05.105 [2024-07-15 11:23:48.466130] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:05.105 [2024-07-15 11:23:48.466149] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x10c36f0 00:13:05.105 [2024-07-15 11:23:48.466162] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:05.105 [2024-07-15 11:23:48.466504] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:05.105 [2024-07-15 11:23:48.466522] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:05.105 [2024-07-15 11:23:48.466582] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:13:05.105 [2024-07-15 11:23:48.466601] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:05.105 [2024-07-15 11:23:48.466697] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x10c4590 00:13:05.105 [2024-07-15 11:23:48.466707] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:05.105 [2024-07-15 11:23:48.466873] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xf25540 00:13:05.105 [2024-07-15 11:23:48.467009] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x10c4590 00:13:05.105 [2024-07-15 11:23:48.467019] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x10c4590 00:13:05.105 [2024-07-15 11:23:48.467115] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:05.105 pt2 00:13:05.105 11:23:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:13:05.105 11:23:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:13:05.105 11:23:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@482 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:05.105 11:23:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:13:05.105 11:23:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:13:05.105 11:23:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:13:05.105 11:23:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:13:05.105 11:23:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:13:05.105 11:23:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:05.105 11:23:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:05.105 11:23:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:05.105 11:23:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:05.105 11:23:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:05.105 11:23:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:05.362 11:23:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:05.362 "name": "raid_bdev1", 00:13:05.362 "uuid": "64691137-5845-4643-9feb-b61cf4017884", 00:13:05.362 "strip_size_kb": 0, 00:13:05.362 "state": "online", 00:13:05.362 "raid_level": "raid1", 00:13:05.362 "superblock": true, 00:13:05.362 "num_base_bdevs": 2, 00:13:05.362 "num_base_bdevs_discovered": 2, 00:13:05.362 "num_base_bdevs_operational": 2, 00:13:05.362 "base_bdevs_list": [ 00:13:05.362 { 00:13:05.362 "name": "pt1", 00:13:05.362 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:05.362 "is_configured": true, 00:13:05.362 "data_offset": 2048, 00:13:05.362 "data_size": 63488 00:13:05.362 }, 00:13:05.362 { 00:13:05.362 "name": "pt2", 00:13:05.362 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:05.362 "is_configured": true, 00:13:05.362 "data_offset": 2048, 00:13:05.362 "data_size": 63488 00:13:05.362 } 00:13:05.362 ] 00:13:05.362 }' 00:13:05.362 11:23:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:05.362 11:23:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:05.926 11:23:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_properties raid_bdev1 00:13:05.926 11:23:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:13:05.926 11:23:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:13:05.926 11:23:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:13:05.926 11:23:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:13:05.926 11:23:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # local name 00:13:05.926 11:23:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:13:05.926 11:23:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:13:06.183 [2024-07-15 11:23:49.549221] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:06.183 11:23:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:13:06.183 "name": "raid_bdev1", 00:13:06.183 "aliases": [ 00:13:06.183 "64691137-5845-4643-9feb-b61cf4017884" 00:13:06.183 ], 00:13:06.183 "product_name": "Raid Volume", 00:13:06.183 "block_size": 512, 00:13:06.183 "num_blocks": 63488, 00:13:06.183 "uuid": "64691137-5845-4643-9feb-b61cf4017884", 00:13:06.183 "assigned_rate_limits": { 00:13:06.183 "rw_ios_per_sec": 0, 00:13:06.183 "rw_mbytes_per_sec": 0, 00:13:06.183 "r_mbytes_per_sec": 0, 00:13:06.183 "w_mbytes_per_sec": 0 00:13:06.183 }, 00:13:06.183 "claimed": false, 00:13:06.183 "zoned": false, 00:13:06.183 "supported_io_types": { 00:13:06.183 "read": true, 00:13:06.183 "write": true, 00:13:06.183 "unmap": false, 00:13:06.183 "flush": false, 00:13:06.183 "reset": true, 00:13:06.183 "nvme_admin": false, 00:13:06.183 "nvme_io": false, 00:13:06.183 "nvme_io_md": false, 00:13:06.183 "write_zeroes": true, 00:13:06.183 "zcopy": false, 00:13:06.183 "get_zone_info": false, 00:13:06.183 "zone_management": false, 00:13:06.183 "zone_append": false, 00:13:06.183 "compare": false, 00:13:06.183 "compare_and_write": false, 00:13:06.183 "abort": false, 00:13:06.183 "seek_hole": false, 00:13:06.183 "seek_data": false, 00:13:06.183 "copy": false, 00:13:06.183 "nvme_iov_md": false 00:13:06.183 }, 00:13:06.183 "memory_domains": [ 00:13:06.183 { 00:13:06.183 "dma_device_id": "system", 00:13:06.183 "dma_device_type": 1 00:13:06.183 }, 00:13:06.183 { 00:13:06.183 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:06.183 "dma_device_type": 2 00:13:06.183 }, 00:13:06.183 { 00:13:06.183 "dma_device_id": "system", 00:13:06.183 "dma_device_type": 1 00:13:06.183 }, 00:13:06.183 { 00:13:06.183 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:06.183 "dma_device_type": 2 00:13:06.183 } 00:13:06.183 ], 00:13:06.183 "driver_specific": { 00:13:06.183 "raid": { 00:13:06.183 "uuid": "64691137-5845-4643-9feb-b61cf4017884", 00:13:06.183 "strip_size_kb": 0, 00:13:06.183 "state": "online", 00:13:06.183 "raid_level": "raid1", 00:13:06.183 "superblock": true, 00:13:06.183 "num_base_bdevs": 2, 00:13:06.183 "num_base_bdevs_discovered": 2, 00:13:06.183 "num_base_bdevs_operational": 2, 00:13:06.183 "base_bdevs_list": [ 00:13:06.183 { 00:13:06.183 "name": "pt1", 00:13:06.183 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:06.183 "is_configured": true, 00:13:06.183 "data_offset": 2048, 00:13:06.183 "data_size": 63488 00:13:06.183 }, 00:13:06.183 { 00:13:06.183 "name": "pt2", 00:13:06.183 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:06.183 "is_configured": true, 00:13:06.183 "data_offset": 2048, 00:13:06.183 "data_size": 63488 00:13:06.183 } 00:13:06.183 ] 00:13:06.183 } 00:13:06.183 } 00:13:06.183 }' 00:13:06.183 11:23:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:06.183 11:23:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:13:06.183 pt2' 00:13:06.183 11:23:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:13:06.183 11:23:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:13:06.183 11:23:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:13:06.442 11:23:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:13:06.442 "name": "pt1", 00:13:06.442 "aliases": [ 00:13:06.442 "00000000-0000-0000-0000-000000000001" 00:13:06.442 ], 00:13:06.442 "product_name": "passthru", 00:13:06.442 "block_size": 512, 00:13:06.442 "num_blocks": 65536, 00:13:06.442 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:06.442 "assigned_rate_limits": { 00:13:06.442 "rw_ios_per_sec": 0, 00:13:06.442 "rw_mbytes_per_sec": 0, 00:13:06.442 "r_mbytes_per_sec": 0, 00:13:06.442 "w_mbytes_per_sec": 0 00:13:06.442 }, 00:13:06.442 "claimed": true, 00:13:06.442 "claim_type": "exclusive_write", 00:13:06.442 "zoned": false, 00:13:06.442 "supported_io_types": { 00:13:06.442 "read": true, 00:13:06.442 "write": true, 00:13:06.442 "unmap": true, 00:13:06.442 "flush": true, 00:13:06.442 "reset": true, 00:13:06.442 "nvme_admin": false, 00:13:06.442 "nvme_io": false, 00:13:06.442 "nvme_io_md": false, 00:13:06.442 "write_zeroes": true, 00:13:06.442 "zcopy": true, 00:13:06.442 "get_zone_info": false, 00:13:06.442 "zone_management": false, 00:13:06.442 "zone_append": false, 00:13:06.442 "compare": false, 00:13:06.442 "compare_and_write": false, 00:13:06.442 "abort": true, 00:13:06.442 "seek_hole": false, 00:13:06.442 "seek_data": false, 00:13:06.442 "copy": true, 00:13:06.442 "nvme_iov_md": false 00:13:06.442 }, 00:13:06.442 "memory_domains": [ 00:13:06.442 { 00:13:06.442 "dma_device_id": "system", 00:13:06.442 "dma_device_type": 1 00:13:06.442 }, 00:13:06.442 { 00:13:06.442 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:06.442 "dma_device_type": 2 00:13:06.442 } 00:13:06.442 ], 00:13:06.442 "driver_specific": { 00:13:06.442 "passthru": { 00:13:06.442 "name": "pt1", 00:13:06.442 "base_bdev_name": "malloc1" 00:13:06.442 } 00:13:06.442 } 00:13:06.442 }' 00:13:06.442 11:23:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:13:06.442 11:23:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:13:06.442 11:23:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:13:06.442 11:23:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:13:06.442 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:13:06.442 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:13:06.442 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:13:06.752 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:13:06.752 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:13:06.752 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:13:06.752 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:13:06.752 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:13:06.752 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:13:06.752 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:13:06.752 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:13:07.038 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:13:07.038 "name": "pt2", 00:13:07.038 "aliases": [ 00:13:07.038 "00000000-0000-0000-0000-000000000002" 00:13:07.038 ], 00:13:07.038 "product_name": "passthru", 00:13:07.038 "block_size": 512, 00:13:07.038 "num_blocks": 65536, 00:13:07.038 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:07.038 "assigned_rate_limits": { 00:13:07.038 "rw_ios_per_sec": 0, 00:13:07.038 "rw_mbytes_per_sec": 0, 00:13:07.038 "r_mbytes_per_sec": 0, 00:13:07.038 "w_mbytes_per_sec": 0 00:13:07.038 }, 00:13:07.038 "claimed": true, 00:13:07.038 "claim_type": "exclusive_write", 00:13:07.038 "zoned": false, 00:13:07.038 "supported_io_types": { 00:13:07.038 "read": true, 00:13:07.038 "write": true, 00:13:07.038 "unmap": true, 00:13:07.038 "flush": true, 00:13:07.038 "reset": true, 00:13:07.038 "nvme_admin": false, 00:13:07.038 "nvme_io": false, 00:13:07.038 "nvme_io_md": false, 00:13:07.038 "write_zeroes": true, 00:13:07.038 "zcopy": true, 00:13:07.038 "get_zone_info": false, 00:13:07.038 "zone_management": false, 00:13:07.038 "zone_append": false, 00:13:07.038 "compare": false, 00:13:07.038 "compare_and_write": false, 00:13:07.038 "abort": true, 00:13:07.038 "seek_hole": false, 00:13:07.038 "seek_data": false, 00:13:07.038 "copy": true, 00:13:07.038 "nvme_iov_md": false 00:13:07.038 }, 00:13:07.038 "memory_domains": [ 00:13:07.038 { 00:13:07.038 "dma_device_id": "system", 00:13:07.038 "dma_device_type": 1 00:13:07.038 }, 00:13:07.038 { 00:13:07.038 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:07.038 "dma_device_type": 2 00:13:07.038 } 00:13:07.038 ], 00:13:07.038 "driver_specific": { 00:13:07.038 "passthru": { 00:13:07.038 "name": "pt2", 00:13:07.038 "base_bdev_name": "malloc2" 00:13:07.038 } 00:13:07.038 } 00:13:07.038 }' 00:13:07.038 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:13:07.038 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:13:07.038 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:13:07.038 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:13:07.038 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:13:07.295 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:13:07.295 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:13:07.295 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:13:07.295 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:13:07.295 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:13:07.295 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:13:07.295 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:13:07.295 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:13:07.295 11:23:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # jq -r '.[] | .uuid' 00:13:07.552 [2024-07-15 11:23:51.057233] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:07.552 11:23:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # '[' 64691137-5845-4643-9feb-b61cf4017884 '!=' 64691137-5845-4643-9feb-b61cf4017884 ']' 00:13:07.552 11:23:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@490 -- # has_redundancy raid1 00:13:07.552 11:23:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:13:07.552 11:23:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@214 -- # return 0 00:13:07.552 11:23:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@492 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt1 00:13:07.810 [2024-07-15 11:23:51.305661] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:13:07.810 11:23:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@495 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:07.810 11:23:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:13:07.810 11:23:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:13:07.810 11:23:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:13:07.810 11:23:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:13:07.810 11:23:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:13:07.810 11:23:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:07.810 11:23:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:07.810 11:23:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:07.810 11:23:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:07.810 11:23:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:07.810 11:23:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:08.068 11:23:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:08.068 "name": "raid_bdev1", 00:13:08.068 "uuid": "64691137-5845-4643-9feb-b61cf4017884", 00:13:08.068 "strip_size_kb": 0, 00:13:08.068 "state": "online", 00:13:08.068 "raid_level": "raid1", 00:13:08.068 "superblock": true, 00:13:08.068 "num_base_bdevs": 2, 00:13:08.068 "num_base_bdevs_discovered": 1, 00:13:08.068 "num_base_bdevs_operational": 1, 00:13:08.068 "base_bdevs_list": [ 00:13:08.068 { 00:13:08.068 "name": null, 00:13:08.068 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:08.068 "is_configured": false, 00:13:08.068 "data_offset": 2048, 00:13:08.068 "data_size": 63488 00:13:08.068 }, 00:13:08.068 { 00:13:08.068 "name": "pt2", 00:13:08.068 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:08.068 "is_configured": true, 00:13:08.068 "data_offset": 2048, 00:13:08.069 "data_size": 63488 00:13:08.069 } 00:13:08.069 ] 00:13:08.069 }' 00:13:08.069 11:23:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:08.069 11:23:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.635 11:23:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@498 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:13:08.893 [2024-07-15 11:23:52.396536] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:08.893 [2024-07-15 11:23:52.396562] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:08.893 [2024-07-15 11:23:52.396611] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:08.893 [2024-07-15 11:23:52.396651] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:08.893 [2024-07-15 11:23:52.396662] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x10c4590 name raid_bdev1, state offline 00:13:08.893 11:23:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:08.893 11:23:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # jq -r '.[]' 00:13:09.151 11:23:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # raid_bdev= 00:13:09.151 11:23:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # '[' -n '' ']' 00:13:09.151 11:23:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@505 -- # (( i = 1 )) 00:13:09.151 11:23:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@505 -- # (( i < num_base_bdevs )) 00:13:09.151 11:23:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:13:09.409 11:23:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@505 -- # (( i++ )) 00:13:09.409 11:23:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@505 -- # (( i < num_base_bdevs )) 00:13:09.409 11:23:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@510 -- # (( i = 1 )) 00:13:09.409 11:23:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@510 -- # (( i < num_base_bdevs - 1 )) 00:13:09.409 11:23:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@518 -- # i=1 00:13:09.409 11:23:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:09.667 [2024-07-15 11:23:53.138477] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:09.667 [2024-07-15 11:23:53.138520] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:09.667 [2024-07-15 11:23:53.138539] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xf2c160 00:13:09.667 [2024-07-15 11:23:53.138551] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:09.667 [2024-07-15 11:23:53.140177] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:09.667 [2024-07-15 11:23:53.140203] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:09.667 [2024-07-15 11:23:53.140267] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:13:09.667 [2024-07-15 11:23:53.140298] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:09.667 [2024-07-15 11:23:53.140382] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xf22380 00:13:09.667 [2024-07-15 11:23:53.140392] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:09.667 [2024-07-15 11:23:53.140563] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xf23a80 00:13:09.667 [2024-07-15 11:23:53.140682] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xf22380 00:13:09.667 [2024-07-15 11:23:53.140692] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0xf22380 00:13:09.667 [2024-07-15 11:23:53.140785] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:09.667 pt2 00:13:09.668 11:23:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@522 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:09.668 11:23:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:13:09.668 11:23:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:13:09.668 11:23:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:13:09.668 11:23:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:13:09.668 11:23:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:13:09.668 11:23:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:09.668 11:23:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:09.668 11:23:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:09.668 11:23:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:09.668 11:23:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:09.668 11:23:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:09.926 11:23:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:09.926 "name": "raid_bdev1", 00:13:09.926 "uuid": "64691137-5845-4643-9feb-b61cf4017884", 00:13:09.926 "strip_size_kb": 0, 00:13:09.926 "state": "online", 00:13:09.926 "raid_level": "raid1", 00:13:09.926 "superblock": true, 00:13:09.926 "num_base_bdevs": 2, 00:13:09.926 "num_base_bdevs_discovered": 1, 00:13:09.926 "num_base_bdevs_operational": 1, 00:13:09.926 "base_bdevs_list": [ 00:13:09.926 { 00:13:09.926 "name": null, 00:13:09.926 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:09.926 "is_configured": false, 00:13:09.926 "data_offset": 2048, 00:13:09.926 "data_size": 63488 00:13:09.926 }, 00:13:09.926 { 00:13:09.926 "name": "pt2", 00:13:09.926 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:09.926 "is_configured": true, 00:13:09.926 "data_offset": 2048, 00:13:09.926 "data_size": 63488 00:13:09.926 } 00:13:09.926 ] 00:13:09.926 }' 00:13:09.926 11:23:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:09.926 11:23:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:10.491 11:23:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@525 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:13:10.750 [2024-07-15 11:23:54.213312] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:10.750 [2024-07-15 11:23:54.213337] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:10.750 [2024-07-15 11:23:54.213387] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:10.750 [2024-07-15 11:23:54.213429] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:10.750 [2024-07-15 11:23:54.213440] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xf22380 name raid_bdev1, state offline 00:13:10.750 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:10.750 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # jq -r '.[]' 00:13:11.008 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # raid_bdev= 00:13:11.008 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # '[' -n '' ']' 00:13:11.008 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@531 -- # '[' 2 -gt 2 ']' 00:13:11.008 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@539 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:13:11.267 [2024-07-15 11:23:54.630394] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:13:11.267 [2024-07-15 11:23:54.630436] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:11.267 [2024-07-15 11:23:54.630457] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x10ce520 00:13:11.267 [2024-07-15 11:23:54.630469] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:11.267 [2024-07-15 11:23:54.632084] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:11.267 [2024-07-15 11:23:54.632113] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:13:11.267 [2024-07-15 11:23:54.632178] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:13:11.267 [2024-07-15 11:23:54.632203] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:13:11.267 [2024-07-15 11:23:54.632300] bdev_raid.c:3547:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:13:11.267 [2024-07-15 11:23:54.632313] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:11.267 [2024-07-15 11:23:54.632327] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xf233f0 name raid_bdev1, state configuring 00:13:11.267 [2024-07-15 11:23:54.632349] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:11.267 [2024-07-15 11:23:54.632409] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xf252b0 00:13:11.267 [2024-07-15 11:23:54.632419] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:11.267 [2024-07-15 11:23:54.632582] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xf22350 00:13:11.267 [2024-07-15 11:23:54.632702] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xf252b0 00:13:11.267 [2024-07-15 11:23:54.632712] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0xf252b0 00:13:11.267 [2024-07-15 11:23:54.632808] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:11.267 pt1 00:13:11.267 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@541 -- # '[' 2 -gt 2 ']' 00:13:11.267 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@553 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:11.267 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:13:11.267 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:13:11.267 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:13:11.267 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:13:11.267 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:13:11.267 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:11.267 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:11.267 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:11.267 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:11.267 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:11.267 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:11.525 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:11.525 "name": "raid_bdev1", 00:13:11.525 "uuid": "64691137-5845-4643-9feb-b61cf4017884", 00:13:11.525 "strip_size_kb": 0, 00:13:11.525 "state": "online", 00:13:11.525 "raid_level": "raid1", 00:13:11.525 "superblock": true, 00:13:11.525 "num_base_bdevs": 2, 00:13:11.525 "num_base_bdevs_discovered": 1, 00:13:11.525 "num_base_bdevs_operational": 1, 00:13:11.525 "base_bdevs_list": [ 00:13:11.525 { 00:13:11.525 "name": null, 00:13:11.525 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.525 "is_configured": false, 00:13:11.525 "data_offset": 2048, 00:13:11.525 "data_size": 63488 00:13:11.525 }, 00:13:11.525 { 00:13:11.525 "name": "pt2", 00:13:11.525 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:11.525 "is_configured": true, 00:13:11.525 "data_offset": 2048, 00:13:11.525 "data_size": 63488 00:13:11.525 } 00:13:11.525 ] 00:13:11.525 }' 00:13:11.525 11:23:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:11.525 11:23:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.092 11:23:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:13:12.092 11:23:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs online 00:13:12.349 11:23:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # [[ false == \f\a\l\s\e ]] 00:13:12.350 11:23:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@557 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:13:12.350 11:23:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@557 -- # jq -r '.[] | .uuid' 00:13:12.608 [2024-07-15 11:23:55.982180] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:12.608 11:23:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@557 -- # '[' 64691137-5845-4643-9feb-b61cf4017884 '!=' 64691137-5845-4643-9feb-b61cf4017884 ']' 00:13:12.608 11:23:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@562 -- # killprocess 878170 00:13:12.608 11:23:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@948 -- # '[' -z 878170 ']' 00:13:12.608 11:23:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@952 -- # kill -0 878170 00:13:12.608 11:23:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@953 -- # uname 00:13:12.608 11:23:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:13:12.608 11:23:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 878170 00:13:12.609 11:23:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:13:12.609 11:23:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:13:12.609 11:23:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 878170' 00:13:12.609 killing process with pid 878170 00:13:12.609 11:23:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@967 -- # kill 878170 00:13:12.609 [2024-07-15 11:23:56.057105] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:12.609 [2024-07-15 11:23:56.057157] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:12.609 [2024-07-15 11:23:56.057199] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:12.609 [2024-07-15 11:23:56.057211] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xf252b0 name raid_bdev1, state offline 00:13:12.609 11:23:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # wait 878170 00:13:12.609 [2024-07-15 11:23:56.074826] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:12.868 11:23:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@564 -- # return 0 00:13:12.868 00:13:12.868 real 0m15.567s 00:13:12.868 user 0m28.771s 00:13:12.868 sys 0m2.840s 00:13:12.868 11:23:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:13:12.868 11:23:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.868 ************************************ 00:13:12.868 END TEST raid_superblock_test 00:13:12.868 ************************************ 00:13:12.868 11:23:56 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:13:12.868 11:23:56 bdev_raid -- bdev/bdev_raid.sh@870 -- # run_test raid_read_error_test raid_io_error_test raid1 2 read 00:13:12.868 11:23:56 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:13:12.868 11:23:56 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:13:12.868 11:23:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:12.868 ************************************ 00:13:12.868 START TEST raid_read_error_test 00:13:12.868 ************************************ 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1123 -- # raid_io_error_test raid1 2 read 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@788 -- # local raid_level=raid1 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@789 -- # local num_base_bdevs=2 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local error_io_type=read 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i = 1 )) 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev1 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev2 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local base_bdevs 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local raid_bdev_name=raid_bdev1 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local strip_size 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local create_arg 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local bdevperf_log 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local fail_per_s 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # '[' raid1 '!=' raid1 ']' 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # strip_size=0 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@805 -- # mktemp -p /raidtest 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@805 -- # bdevperf_log=/raidtest/tmp.Fycp65e6Cw 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@808 -- # raid_pid=880509 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # waitforlisten 880509 /var/tmp/spdk-raid.sock 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@829 -- # '[' -z 880509 ']' 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:13:12.868 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:13:12.868 11:23:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.868 [2024-07-15 11:23:56.451711] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:13:12.868 [2024-07-15 11:23:56.451775] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid880509 ] 00:13:13.126 [2024-07-15 11:23:56.580294] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:13.126 [2024-07-15 11:23:56.677951] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:13:13.384 [2024-07-15 11:23:56.748911] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:13.384 [2024-07-15 11:23:56.748975] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:13.978 11:23:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:13:13.978 11:23:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@862 -- # return 0 00:13:13.978 11:23:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:13:13.978 11:23:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:13.978 BaseBdev1_malloc 00:13:14.236 11:23:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev1_malloc 00:13:14.236 true 00:13:14.494 11:23:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:13:14.494 [2024-07-15 11:23:58.056639] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:13:14.494 [2024-07-15 11:23:58.056684] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:14.494 [2024-07-15 11:23:58.056707] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x10650d0 00:13:14.494 [2024-07-15 11:23:58.056719] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:14.494 [2024-07-15 11:23:58.058607] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:14.494 [2024-07-15 11:23:58.058636] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:14.494 BaseBdev1 00:13:14.494 11:23:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:13:14.495 11:23:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:14.753 BaseBdev2_malloc 00:13:14.753 11:23:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev2_malloc 00:13:15.011 true 00:13:15.011 11:23:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:13:15.269 [2024-07-15 11:23:58.784430] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:13:15.269 [2024-07-15 11:23:58.784472] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:15.269 [2024-07-15 11:23:58.784491] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1069910 00:13:15.269 [2024-07-15 11:23:58.784504] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:15.269 [2024-07-15 11:23:58.785980] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:15.269 [2024-07-15 11:23:58.786007] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:15.269 BaseBdev2 00:13:15.270 11:23:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@819 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2' -n raid_bdev1 -s 00:13:15.528 [2024-07-15 11:23:59.041131] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:15.528 [2024-07-15 11:23:59.042490] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:15.528 [2024-07-15 11:23:59.042684] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x106b320 00:13:15.528 [2024-07-15 11:23:59.042698] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:15.528 [2024-07-15 11:23:59.042896] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xed2d00 00:13:15.528 [2024-07-15 11:23:59.043059] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x106b320 00:13:15.528 [2024-07-15 11:23:59.043069] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x106b320 00:13:15.528 [2024-07-15 11:23:59.043185] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:15.528 11:23:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@820 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:15.528 11:23:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:13:15.528 11:23:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:13:15.528 11:23:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:13:15.528 11:23:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:13:15.528 11:23:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:13:15.528 11:23:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:15.528 11:23:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:15.528 11:23:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:15.528 11:23:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:15.528 11:23:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:15.528 11:23:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:15.786 11:23:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:15.786 "name": "raid_bdev1", 00:13:15.786 "uuid": "e6d70f55-3458-42bd-8022-22f51ed2fd20", 00:13:15.786 "strip_size_kb": 0, 00:13:15.786 "state": "online", 00:13:15.786 "raid_level": "raid1", 00:13:15.786 "superblock": true, 00:13:15.786 "num_base_bdevs": 2, 00:13:15.786 "num_base_bdevs_discovered": 2, 00:13:15.786 "num_base_bdevs_operational": 2, 00:13:15.786 "base_bdevs_list": [ 00:13:15.786 { 00:13:15.786 "name": "BaseBdev1", 00:13:15.786 "uuid": "a578c10f-5b4e-55d2-b3e0-93697efe18e2", 00:13:15.786 "is_configured": true, 00:13:15.786 "data_offset": 2048, 00:13:15.786 "data_size": 63488 00:13:15.786 }, 00:13:15.786 { 00:13:15.786 "name": "BaseBdev2", 00:13:15.786 "uuid": "ebbfc8d5-be1a-5cca-9072-796428d37b17", 00:13:15.786 "is_configured": true, 00:13:15.786 "data_offset": 2048, 00:13:15.786 "data_size": 63488 00:13:15.786 } 00:13:15.786 ] 00:13:15.786 }' 00:13:15.787 11:23:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:15.787 11:23:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:13:16.353 11:23:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@824 -- # sleep 1 00:13:16.353 11:23:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@823 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:13:16.611 [2024-07-15 11:24:00.020020] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1066c70 00:13:17.546 11:24:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@827 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:13:17.804 11:24:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # local expected_num_base_bdevs 00:13:17.804 11:24:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@830 -- # [[ raid1 = \r\a\i\d\1 ]] 00:13:17.804 11:24:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@830 -- # [[ read = \w\r\i\t\e ]] 00:13:17.804 11:24:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=2 00:13:17.804 11:24:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:17.804 11:24:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:13:17.804 11:24:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:13:17.804 11:24:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:13:17.804 11:24:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:13:17.804 11:24:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:13:17.804 11:24:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:17.804 11:24:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:17.804 11:24:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:17.804 11:24:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:17.804 11:24:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:17.804 11:24:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:18.062 11:24:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:18.062 "name": "raid_bdev1", 00:13:18.062 "uuid": "e6d70f55-3458-42bd-8022-22f51ed2fd20", 00:13:18.062 "strip_size_kb": 0, 00:13:18.062 "state": "online", 00:13:18.063 "raid_level": "raid1", 00:13:18.063 "superblock": true, 00:13:18.063 "num_base_bdevs": 2, 00:13:18.063 "num_base_bdevs_discovered": 2, 00:13:18.063 "num_base_bdevs_operational": 2, 00:13:18.063 "base_bdevs_list": [ 00:13:18.063 { 00:13:18.063 "name": "BaseBdev1", 00:13:18.063 "uuid": "a578c10f-5b4e-55d2-b3e0-93697efe18e2", 00:13:18.063 "is_configured": true, 00:13:18.063 "data_offset": 2048, 00:13:18.063 "data_size": 63488 00:13:18.063 }, 00:13:18.063 { 00:13:18.063 "name": "BaseBdev2", 00:13:18.063 "uuid": "ebbfc8d5-be1a-5cca-9072-796428d37b17", 00:13:18.063 "is_configured": true, 00:13:18.063 "data_offset": 2048, 00:13:18.063 "data_size": 63488 00:13:18.063 } 00:13:18.063 ] 00:13:18.063 }' 00:13:18.063 11:24:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:18.063 11:24:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:13:18.628 11:24:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:13:18.885 [2024-07-15 11:24:02.227470] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:18.885 [2024-07-15 11:24:02.227502] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:18.885 [2024-07-15 11:24:02.230739] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:18.885 [2024-07-15 11:24:02.230769] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:18.885 [2024-07-15 11:24:02.230850] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:18.885 [2024-07-15 11:24:02.230862] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x106b320 name raid_bdev1, state offline 00:13:18.885 0 00:13:18.885 11:24:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # killprocess 880509 00:13:18.885 11:24:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@948 -- # '[' -z 880509 ']' 00:13:18.885 11:24:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@952 -- # kill -0 880509 00:13:18.885 11:24:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@953 -- # uname 00:13:18.885 11:24:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:13:18.885 11:24:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 880509 00:13:18.885 11:24:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:13:18.885 11:24:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:13:18.885 11:24:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 880509' 00:13:18.885 killing process with pid 880509 00:13:18.885 11:24:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@967 -- # kill 880509 00:13:18.885 [2024-07-15 11:24:02.295960] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:18.885 11:24:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # wait 880509 00:13:18.885 [2024-07-15 11:24:02.306590] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:19.143 11:24:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # grep raid_bdev1 00:13:19.143 11:24:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # grep -v Job /raidtest/tmp.Fycp65e6Cw 00:13:19.143 11:24:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # awk '{print $6}' 00:13:19.143 11:24:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # fail_per_s=0.00 00:13:19.143 11:24:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@844 -- # has_redundancy raid1 00:13:19.143 11:24:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:13:19.143 11:24:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@214 -- # return 0 00:13:19.143 11:24:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # [[ 0.00 = \0\.\0\0 ]] 00:13:19.143 00:13:19.143 real 0m6.167s 00:13:19.143 user 0m9.582s 00:13:19.143 sys 0m1.129s 00:13:19.143 11:24:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:13:19.143 11:24:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:13:19.143 ************************************ 00:13:19.143 END TEST raid_read_error_test 00:13:19.143 ************************************ 00:13:19.143 11:24:02 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:13:19.143 11:24:02 bdev_raid -- bdev/bdev_raid.sh@871 -- # run_test raid_write_error_test raid_io_error_test raid1 2 write 00:13:19.143 11:24:02 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:13:19.143 11:24:02 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:13:19.143 11:24:02 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:19.143 ************************************ 00:13:19.143 START TEST raid_write_error_test 00:13:19.143 ************************************ 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1123 -- # raid_io_error_test raid1 2 write 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@788 -- # local raid_level=raid1 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@789 -- # local num_base_bdevs=2 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local error_io_type=write 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i = 1 )) 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev1 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev2 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local base_bdevs 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local raid_bdev_name=raid_bdev1 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local strip_size 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local create_arg 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local bdevperf_log 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local fail_per_s 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # '[' raid1 '!=' raid1 ']' 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # strip_size=0 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@805 -- # mktemp -p /raidtest 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@805 -- # bdevperf_log=/raidtest/tmp.Je5DFjbH1U 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@808 -- # raid_pid=881601 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # waitforlisten 881601 /var/tmp/spdk-raid.sock 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@829 -- # '[' -z 881601 ']' 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:13:19.143 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:13:19.143 11:24:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:13:19.143 [2024-07-15 11:24:02.701750] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:13:19.144 [2024-07-15 11:24:02.701817] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid881601 ] 00:13:19.401 [2024-07-15 11:24:02.831108] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:19.401 [2024-07-15 11:24:02.939643] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:13:19.659 [2024-07-15 11:24:03.001324] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:19.659 [2024-07-15 11:24:03.001351] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:19.659 11:24:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:13:19.659 11:24:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@862 -- # return 0 00:13:19.659 11:24:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:13:19.659 11:24:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:19.915 BaseBdev1_malloc 00:13:19.915 11:24:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev1_malloc 00:13:20.180 true 00:13:20.180 11:24:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:13:20.438 [2024-07-15 11:24:03.869710] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:13:20.438 [2024-07-15 11:24:03.869752] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:20.438 [2024-07-15 11:24:03.869773] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xb190d0 00:13:20.438 [2024-07-15 11:24:03.869786] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:20.438 [2024-07-15 11:24:03.871672] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:20.438 [2024-07-15 11:24:03.871700] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:20.438 BaseBdev1 00:13:20.438 11:24:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:13:20.438 11:24:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:20.696 BaseBdev2_malloc 00:13:20.696 11:24:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev2_malloc 00:13:20.953 true 00:13:20.953 11:24:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:13:21.210 [2024-07-15 11:24:04.585438] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:13:21.210 [2024-07-15 11:24:04.585481] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:21.210 [2024-07-15 11:24:04.585502] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xb1d910 00:13:21.210 [2024-07-15 11:24:04.585514] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:21.210 [2024-07-15 11:24:04.587140] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:21.210 [2024-07-15 11:24:04.587167] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:21.210 BaseBdev2 00:13:21.210 11:24:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@819 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2' -n raid_bdev1 -s 00:13:21.467 [2024-07-15 11:24:04.830109] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:21.467 [2024-07-15 11:24:04.831473] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:21.467 [2024-07-15 11:24:04.831668] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xb1f320 00:13:21.467 [2024-07-15 11:24:04.831682] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:21.467 [2024-07-15 11:24:04.831878] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x986d00 00:13:21.467 [2024-07-15 11:24:04.832044] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xb1f320 00:13:21.467 [2024-07-15 11:24:04.832056] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0xb1f320 00:13:21.467 [2024-07-15 11:24:04.832172] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:21.467 11:24:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@820 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:21.467 11:24:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:13:21.467 11:24:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:13:21.467 11:24:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:13:21.467 11:24:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:13:21.467 11:24:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:13:21.467 11:24:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:21.467 11:24:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:21.467 11:24:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:21.467 11:24:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:21.467 11:24:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:21.467 11:24:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:21.725 11:24:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:21.725 "name": "raid_bdev1", 00:13:21.725 "uuid": "1124b6fe-e0c6-402c-a356-c03cd3a3571c", 00:13:21.725 "strip_size_kb": 0, 00:13:21.725 "state": "online", 00:13:21.725 "raid_level": "raid1", 00:13:21.725 "superblock": true, 00:13:21.725 "num_base_bdevs": 2, 00:13:21.725 "num_base_bdevs_discovered": 2, 00:13:21.725 "num_base_bdevs_operational": 2, 00:13:21.725 "base_bdevs_list": [ 00:13:21.725 { 00:13:21.725 "name": "BaseBdev1", 00:13:21.725 "uuid": "c4b7f15a-172f-5ce6-a5b0-5dd892db0ea3", 00:13:21.725 "is_configured": true, 00:13:21.725 "data_offset": 2048, 00:13:21.725 "data_size": 63488 00:13:21.725 }, 00:13:21.725 { 00:13:21.725 "name": "BaseBdev2", 00:13:21.725 "uuid": "613b5c71-a7d6-509f-82ef-80ed43b57391", 00:13:21.725 "is_configured": true, 00:13:21.725 "data_offset": 2048, 00:13:21.725 "data_size": 63488 00:13:21.725 } 00:13:21.725 ] 00:13:21.725 }' 00:13:21.725 11:24:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:21.725 11:24:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:13:22.289 11:24:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@824 -- # sleep 1 00:13:22.289 11:24:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@823 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:13:22.289 [2024-07-15 11:24:05.800999] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xb1ac70 00:13:23.221 11:24:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@827 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:13:23.478 [2024-07-15 11:24:06.916572] bdev_raid.c:2221:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:13:23.478 [2024-07-15 11:24:06.916638] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:23.478 [2024-07-15 11:24:06.916812] bdev_raid.c:1919:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0xb1ac70 00:13:23.478 11:24:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # local expected_num_base_bdevs 00:13:23.478 11:24:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@830 -- # [[ raid1 = \r\a\i\d\1 ]] 00:13:23.478 11:24:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@830 -- # [[ write = \w\r\i\t\e ]] 00:13:23.478 11:24:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # expected_num_base_bdevs=1 00:13:23.478 11:24:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:23.478 11:24:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:13:23.478 11:24:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:13:23.479 11:24:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:13:23.479 11:24:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:13:23.479 11:24:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:13:23.479 11:24:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:23.479 11:24:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:23.479 11:24:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:23.479 11:24:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:23.479 11:24:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:23.479 11:24:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:23.738 11:24:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:23.738 "name": "raid_bdev1", 00:13:23.738 "uuid": "1124b6fe-e0c6-402c-a356-c03cd3a3571c", 00:13:23.738 "strip_size_kb": 0, 00:13:23.738 "state": "online", 00:13:23.738 "raid_level": "raid1", 00:13:23.738 "superblock": true, 00:13:23.738 "num_base_bdevs": 2, 00:13:23.738 "num_base_bdevs_discovered": 1, 00:13:23.738 "num_base_bdevs_operational": 1, 00:13:23.738 "base_bdevs_list": [ 00:13:23.738 { 00:13:23.738 "name": null, 00:13:23.738 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:23.738 "is_configured": false, 00:13:23.738 "data_offset": 2048, 00:13:23.738 "data_size": 63488 00:13:23.738 }, 00:13:23.738 { 00:13:23.738 "name": "BaseBdev2", 00:13:23.738 "uuid": "613b5c71-a7d6-509f-82ef-80ed43b57391", 00:13:23.738 "is_configured": true, 00:13:23.738 "data_offset": 2048, 00:13:23.738 "data_size": 63488 00:13:23.738 } 00:13:23.738 ] 00:13:23.738 }' 00:13:23.738 11:24:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:23.738 11:24:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.364 11:24:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:13:24.623 [2024-07-15 11:24:08.012575] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:24.623 [2024-07-15 11:24:08.012608] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:24.623 [2024-07-15 11:24:08.015739] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:24.623 [2024-07-15 11:24:08.015765] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:24.623 [2024-07-15 11:24:08.015816] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:24.623 [2024-07-15 11:24:08.015827] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xb1f320 name raid_bdev1, state offline 00:13:24.623 0 00:13:24.623 11:24:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # killprocess 881601 00:13:24.623 11:24:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@948 -- # '[' -z 881601 ']' 00:13:24.623 11:24:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@952 -- # kill -0 881601 00:13:24.623 11:24:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@953 -- # uname 00:13:24.623 11:24:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:13:24.623 11:24:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 881601 00:13:24.623 11:24:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:13:24.623 11:24:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:13:24.623 11:24:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 881601' 00:13:24.623 killing process with pid 881601 00:13:24.623 11:24:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@967 -- # kill 881601 00:13:24.623 [2024-07-15 11:24:08.084289] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:24.623 11:24:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # wait 881601 00:13:24.623 [2024-07-15 11:24:08.094462] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:24.882 11:24:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # grep -v Job /raidtest/tmp.Je5DFjbH1U 00:13:24.882 11:24:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # grep raid_bdev1 00:13:24.882 11:24:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # awk '{print $6}' 00:13:24.882 11:24:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # fail_per_s=0.00 00:13:24.882 11:24:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@844 -- # has_redundancy raid1 00:13:24.882 11:24:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:13:24.882 11:24:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@214 -- # return 0 00:13:24.882 11:24:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # [[ 0.00 = \0\.\0\0 ]] 00:13:24.882 00:13:24.882 real 0m5.694s 00:13:24.882 user 0m9.170s 00:13:24.882 sys 0m1.103s 00:13:24.882 11:24:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:13:24.882 11:24:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.882 ************************************ 00:13:24.882 END TEST raid_write_error_test 00:13:24.882 ************************************ 00:13:24.882 11:24:08 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:13:24.882 11:24:08 bdev_raid -- bdev/bdev_raid.sh@865 -- # for n in {2..4} 00:13:24.882 11:24:08 bdev_raid -- bdev/bdev_raid.sh@866 -- # for level in raid0 concat raid1 00:13:24.882 11:24:08 bdev_raid -- bdev/bdev_raid.sh@867 -- # run_test raid_state_function_test raid_state_function_test raid0 3 false 00:13:24.882 11:24:08 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:13:24.882 11:24:08 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:13:24.882 11:24:08 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:24.882 ************************************ 00:13:24.882 START TEST raid_state_function_test 00:13:24.882 ************************************ 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1123 -- # raid_state_function_test raid0 3 false 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@220 -- # local raid_level=raid0 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=3 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # local superblock=false 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev3 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # local strip_size 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # '[' raid0 '!=' raid1 ']' 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # strip_size=64 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@232 -- # strip_size_create_arg='-z 64' 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # '[' false = true ']' 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@240 -- # superblock_create_arg= 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # raid_pid=882792 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 882792' 00:13:24.882 Process raid pid: 882792 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@246 -- # waitforlisten 882792 /var/tmp/spdk-raid.sock 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@829 -- # '[' -z 882792 ']' 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:13:24.882 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:13:24.882 11:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.882 [2024-07-15 11:24:08.470467] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:13:24.882 [2024-07-15 11:24:08.470531] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:25.140 [2024-07-15 11:24:08.601784] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:25.140 [2024-07-15 11:24:08.708018] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:13:25.398 [2024-07-15 11:24:08.773182] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:25.398 [2024-07-15 11:24:08.773221] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:25.964 11:24:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:13:25.965 11:24:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@862 -- # return 0 00:13:25.965 11:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:13:26.223 [2024-07-15 11:24:09.628824] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:26.223 [2024-07-15 11:24:09.628867] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:26.223 [2024-07-15 11:24:09.628879] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:26.223 [2024-07-15 11:24:09.628891] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:26.223 [2024-07-15 11:24:09.628904] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:26.223 [2024-07-15 11:24:09.628923] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:26.223 11:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:13:26.223 11:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:13:26.223 11:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:13:26.223 11:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:13:26.223 11:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:13:26.223 11:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:13:26.223 11:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:26.223 11:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:26.223 11:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:26.223 11:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:26.223 11:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:26.223 11:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:26.481 11:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:26.481 "name": "Existed_Raid", 00:13:26.481 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:26.481 "strip_size_kb": 64, 00:13:26.481 "state": "configuring", 00:13:26.481 "raid_level": "raid0", 00:13:26.481 "superblock": false, 00:13:26.481 "num_base_bdevs": 3, 00:13:26.481 "num_base_bdevs_discovered": 0, 00:13:26.481 "num_base_bdevs_operational": 3, 00:13:26.481 "base_bdevs_list": [ 00:13:26.481 { 00:13:26.481 "name": "BaseBdev1", 00:13:26.481 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:26.481 "is_configured": false, 00:13:26.481 "data_offset": 0, 00:13:26.481 "data_size": 0 00:13:26.481 }, 00:13:26.481 { 00:13:26.482 "name": "BaseBdev2", 00:13:26.482 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:26.482 "is_configured": false, 00:13:26.482 "data_offset": 0, 00:13:26.482 "data_size": 0 00:13:26.482 }, 00:13:26.482 { 00:13:26.482 "name": "BaseBdev3", 00:13:26.482 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:26.482 "is_configured": false, 00:13:26.482 "data_offset": 0, 00:13:26.482 "data_size": 0 00:13:26.482 } 00:13:26.482 ] 00:13:26.482 }' 00:13:26.482 11:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:26.482 11:24:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.046 11:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:13:27.304 [2024-07-15 11:24:10.723601] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:27.304 [2024-07-15 11:24:10.723637] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xfc3a80 name Existed_Raid, state configuring 00:13:27.304 11:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:13:27.563 [2024-07-15 11:24:10.964238] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:27.563 [2024-07-15 11:24:10.964269] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:27.563 [2024-07-15 11:24:10.964279] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:27.563 [2024-07-15 11:24:10.964290] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:27.563 [2024-07-15 11:24:10.964299] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:27.563 [2024-07-15 11:24:10.964310] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:27.563 11:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:13:27.822 [2024-07-15 11:24:11.214622] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:27.822 BaseBdev1 00:13:27.822 11:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:13:27.822 11:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:13:27.822 11:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:13:27.822 11:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:13:27.822 11:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:13:27.822 11:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:13:27.822 11:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:13:28.080 11:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:28.339 [ 00:13:28.339 { 00:13:28.339 "name": "BaseBdev1", 00:13:28.339 "aliases": [ 00:13:28.339 "7639f8b2-fe30-4f82-93d0-fe330e4c832c" 00:13:28.339 ], 00:13:28.339 "product_name": "Malloc disk", 00:13:28.339 "block_size": 512, 00:13:28.339 "num_blocks": 65536, 00:13:28.339 "uuid": "7639f8b2-fe30-4f82-93d0-fe330e4c832c", 00:13:28.339 "assigned_rate_limits": { 00:13:28.339 "rw_ios_per_sec": 0, 00:13:28.339 "rw_mbytes_per_sec": 0, 00:13:28.339 "r_mbytes_per_sec": 0, 00:13:28.339 "w_mbytes_per_sec": 0 00:13:28.339 }, 00:13:28.339 "claimed": true, 00:13:28.339 "claim_type": "exclusive_write", 00:13:28.339 "zoned": false, 00:13:28.339 "supported_io_types": { 00:13:28.339 "read": true, 00:13:28.339 "write": true, 00:13:28.339 "unmap": true, 00:13:28.339 "flush": true, 00:13:28.339 "reset": true, 00:13:28.339 "nvme_admin": false, 00:13:28.339 "nvme_io": false, 00:13:28.339 "nvme_io_md": false, 00:13:28.339 "write_zeroes": true, 00:13:28.339 "zcopy": true, 00:13:28.339 "get_zone_info": false, 00:13:28.339 "zone_management": false, 00:13:28.339 "zone_append": false, 00:13:28.339 "compare": false, 00:13:28.339 "compare_and_write": false, 00:13:28.339 "abort": true, 00:13:28.339 "seek_hole": false, 00:13:28.339 "seek_data": false, 00:13:28.339 "copy": true, 00:13:28.339 "nvme_iov_md": false 00:13:28.339 }, 00:13:28.339 "memory_domains": [ 00:13:28.339 { 00:13:28.339 "dma_device_id": "system", 00:13:28.339 "dma_device_type": 1 00:13:28.339 }, 00:13:28.339 { 00:13:28.339 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:28.339 "dma_device_type": 2 00:13:28.339 } 00:13:28.339 ], 00:13:28.339 "driver_specific": {} 00:13:28.339 } 00:13:28.339 ] 00:13:28.339 11:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:13:28.339 11:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:13:28.339 11:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:13:28.339 11:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:13:28.339 11:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:13:28.339 11:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:13:28.339 11:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:13:28.339 11:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:28.339 11:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:28.339 11:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:28.339 11:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:28.339 11:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:28.339 11:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:28.597 11:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:28.597 "name": "Existed_Raid", 00:13:28.597 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:28.597 "strip_size_kb": 64, 00:13:28.597 "state": "configuring", 00:13:28.597 "raid_level": "raid0", 00:13:28.597 "superblock": false, 00:13:28.597 "num_base_bdevs": 3, 00:13:28.597 "num_base_bdevs_discovered": 1, 00:13:28.597 "num_base_bdevs_operational": 3, 00:13:28.597 "base_bdevs_list": [ 00:13:28.597 { 00:13:28.597 "name": "BaseBdev1", 00:13:28.597 "uuid": "7639f8b2-fe30-4f82-93d0-fe330e4c832c", 00:13:28.597 "is_configured": true, 00:13:28.597 "data_offset": 0, 00:13:28.597 "data_size": 65536 00:13:28.597 }, 00:13:28.597 { 00:13:28.597 "name": "BaseBdev2", 00:13:28.597 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:28.597 "is_configured": false, 00:13:28.597 "data_offset": 0, 00:13:28.597 "data_size": 0 00:13:28.597 }, 00:13:28.597 { 00:13:28.597 "name": "BaseBdev3", 00:13:28.597 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:28.597 "is_configured": false, 00:13:28.597 "data_offset": 0, 00:13:28.597 "data_size": 0 00:13:28.597 } 00:13:28.597 ] 00:13:28.597 }' 00:13:28.597 11:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:28.597 11:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.162 11:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:13:29.420 [2024-07-15 11:24:12.794806] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:29.420 [2024-07-15 11:24:12.794846] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xfc3310 name Existed_Raid, state configuring 00:13:29.420 11:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:13:29.678 [2024-07-15 11:24:13.039480] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:29.678 [2024-07-15 11:24:13.040906] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:29.678 [2024-07-15 11:24:13.040947] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:29.678 [2024-07-15 11:24:13.040957] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:29.678 [2024-07-15 11:24:13.040969] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:29.678 11:24:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:13:29.678 11:24:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:13:29.678 11:24:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:13:29.678 11:24:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:13:29.678 11:24:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:13:29.678 11:24:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:13:29.678 11:24:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:13:29.678 11:24:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:13:29.678 11:24:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:29.678 11:24:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:29.678 11:24:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:29.678 11:24:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:29.678 11:24:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:29.678 11:24:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:29.934 11:24:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:29.934 "name": "Existed_Raid", 00:13:29.934 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:29.934 "strip_size_kb": 64, 00:13:29.934 "state": "configuring", 00:13:29.934 "raid_level": "raid0", 00:13:29.934 "superblock": false, 00:13:29.934 "num_base_bdevs": 3, 00:13:29.934 "num_base_bdevs_discovered": 1, 00:13:29.934 "num_base_bdevs_operational": 3, 00:13:29.934 "base_bdevs_list": [ 00:13:29.934 { 00:13:29.934 "name": "BaseBdev1", 00:13:29.934 "uuid": "7639f8b2-fe30-4f82-93d0-fe330e4c832c", 00:13:29.934 "is_configured": true, 00:13:29.934 "data_offset": 0, 00:13:29.934 "data_size": 65536 00:13:29.934 }, 00:13:29.934 { 00:13:29.934 "name": "BaseBdev2", 00:13:29.934 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:29.934 "is_configured": false, 00:13:29.934 "data_offset": 0, 00:13:29.934 "data_size": 0 00:13:29.934 }, 00:13:29.934 { 00:13:29.934 "name": "BaseBdev3", 00:13:29.934 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:29.934 "is_configured": false, 00:13:29.934 "data_offset": 0, 00:13:29.934 "data_size": 0 00:13:29.934 } 00:13:29.934 ] 00:13:29.934 }' 00:13:29.934 11:24:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:29.934 11:24:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.496 11:24:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:13:30.753 [2024-07-15 11:24:14.157903] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:30.753 BaseBdev2 00:13:30.753 11:24:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:13:30.753 11:24:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:13:30.753 11:24:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:13:30.753 11:24:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:13:30.753 11:24:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:13:30.753 11:24:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:13:30.753 11:24:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:13:31.010 11:24:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:31.268 [ 00:13:31.268 { 00:13:31.268 "name": "BaseBdev2", 00:13:31.268 "aliases": [ 00:13:31.268 "2c95bd9e-0b98-4fb2-b811-a916674c37e8" 00:13:31.268 ], 00:13:31.268 "product_name": "Malloc disk", 00:13:31.268 "block_size": 512, 00:13:31.268 "num_blocks": 65536, 00:13:31.268 "uuid": "2c95bd9e-0b98-4fb2-b811-a916674c37e8", 00:13:31.268 "assigned_rate_limits": { 00:13:31.268 "rw_ios_per_sec": 0, 00:13:31.268 "rw_mbytes_per_sec": 0, 00:13:31.268 "r_mbytes_per_sec": 0, 00:13:31.268 "w_mbytes_per_sec": 0 00:13:31.268 }, 00:13:31.268 "claimed": true, 00:13:31.268 "claim_type": "exclusive_write", 00:13:31.268 "zoned": false, 00:13:31.268 "supported_io_types": { 00:13:31.268 "read": true, 00:13:31.268 "write": true, 00:13:31.268 "unmap": true, 00:13:31.268 "flush": true, 00:13:31.268 "reset": true, 00:13:31.268 "nvme_admin": false, 00:13:31.268 "nvme_io": false, 00:13:31.268 "nvme_io_md": false, 00:13:31.268 "write_zeroes": true, 00:13:31.268 "zcopy": true, 00:13:31.268 "get_zone_info": false, 00:13:31.268 "zone_management": false, 00:13:31.268 "zone_append": false, 00:13:31.268 "compare": false, 00:13:31.268 "compare_and_write": false, 00:13:31.268 "abort": true, 00:13:31.268 "seek_hole": false, 00:13:31.268 "seek_data": false, 00:13:31.268 "copy": true, 00:13:31.268 "nvme_iov_md": false 00:13:31.268 }, 00:13:31.268 "memory_domains": [ 00:13:31.268 { 00:13:31.268 "dma_device_id": "system", 00:13:31.268 "dma_device_type": 1 00:13:31.268 }, 00:13:31.268 { 00:13:31.268 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:31.268 "dma_device_type": 2 00:13:31.268 } 00:13:31.268 ], 00:13:31.268 "driver_specific": {} 00:13:31.268 } 00:13:31.268 ] 00:13:31.268 11:24:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:13:31.268 11:24:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:13:31.268 11:24:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:13:31.268 11:24:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:13:31.268 11:24:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:13:31.268 11:24:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:13:31.268 11:24:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:13:31.268 11:24:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:13:31.268 11:24:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:13:31.268 11:24:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:31.268 11:24:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:31.268 11:24:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:31.268 11:24:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:31.268 11:24:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:31.268 11:24:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:31.526 11:24:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:31.526 "name": "Existed_Raid", 00:13:31.526 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:31.526 "strip_size_kb": 64, 00:13:31.526 "state": "configuring", 00:13:31.526 "raid_level": "raid0", 00:13:31.526 "superblock": false, 00:13:31.526 "num_base_bdevs": 3, 00:13:31.526 "num_base_bdevs_discovered": 2, 00:13:31.526 "num_base_bdevs_operational": 3, 00:13:31.526 "base_bdevs_list": [ 00:13:31.526 { 00:13:31.526 "name": "BaseBdev1", 00:13:31.526 "uuid": "7639f8b2-fe30-4f82-93d0-fe330e4c832c", 00:13:31.526 "is_configured": true, 00:13:31.526 "data_offset": 0, 00:13:31.526 "data_size": 65536 00:13:31.526 }, 00:13:31.526 { 00:13:31.526 "name": "BaseBdev2", 00:13:31.526 "uuid": "2c95bd9e-0b98-4fb2-b811-a916674c37e8", 00:13:31.526 "is_configured": true, 00:13:31.526 "data_offset": 0, 00:13:31.526 "data_size": 65536 00:13:31.526 }, 00:13:31.526 { 00:13:31.526 "name": "BaseBdev3", 00:13:31.526 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:31.526 "is_configured": false, 00:13:31.526 "data_offset": 0, 00:13:31.526 "data_size": 0 00:13:31.526 } 00:13:31.526 ] 00:13:31.526 }' 00:13:31.526 11:24:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:31.526 11:24:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.091 11:24:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:13:32.349 [2024-07-15 11:24:15.745587] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:32.349 [2024-07-15 11:24:15.745623] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xfc4400 00:13:32.349 [2024-07-15 11:24:15.745632] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:13:32.349 [2024-07-15 11:24:15.745880] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xfc3ef0 00:13:32.349 [2024-07-15 11:24:15.746007] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xfc4400 00:13:32.349 [2024-07-15 11:24:15.746018] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0xfc4400 00:13:32.349 [2024-07-15 11:24:15.746174] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:32.349 BaseBdev3 00:13:32.349 11:24:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev3 00:13:32.349 11:24:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:13:32.349 11:24:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:13:32.349 11:24:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:13:32.349 11:24:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:13:32.349 11:24:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:13:32.349 11:24:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:13:32.606 11:24:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:32.864 [ 00:13:32.864 { 00:13:32.864 "name": "BaseBdev3", 00:13:32.864 "aliases": [ 00:13:32.864 "e37d5fe0-92b2-4517-a8be-b7ecad38b278" 00:13:32.864 ], 00:13:32.864 "product_name": "Malloc disk", 00:13:32.864 "block_size": 512, 00:13:32.864 "num_blocks": 65536, 00:13:32.864 "uuid": "e37d5fe0-92b2-4517-a8be-b7ecad38b278", 00:13:32.864 "assigned_rate_limits": { 00:13:32.864 "rw_ios_per_sec": 0, 00:13:32.864 "rw_mbytes_per_sec": 0, 00:13:32.864 "r_mbytes_per_sec": 0, 00:13:32.864 "w_mbytes_per_sec": 0 00:13:32.864 }, 00:13:32.864 "claimed": true, 00:13:32.864 "claim_type": "exclusive_write", 00:13:32.864 "zoned": false, 00:13:32.864 "supported_io_types": { 00:13:32.864 "read": true, 00:13:32.864 "write": true, 00:13:32.864 "unmap": true, 00:13:32.864 "flush": true, 00:13:32.864 "reset": true, 00:13:32.864 "nvme_admin": false, 00:13:32.864 "nvme_io": false, 00:13:32.864 "nvme_io_md": false, 00:13:32.864 "write_zeroes": true, 00:13:32.864 "zcopy": true, 00:13:32.864 "get_zone_info": false, 00:13:32.864 "zone_management": false, 00:13:32.864 "zone_append": false, 00:13:32.864 "compare": false, 00:13:32.864 "compare_and_write": false, 00:13:32.864 "abort": true, 00:13:32.864 "seek_hole": false, 00:13:32.864 "seek_data": false, 00:13:32.864 "copy": true, 00:13:32.864 "nvme_iov_md": false 00:13:32.864 }, 00:13:32.864 "memory_domains": [ 00:13:32.864 { 00:13:32.864 "dma_device_id": "system", 00:13:32.864 "dma_device_type": 1 00:13:32.864 }, 00:13:32.864 { 00:13:32.864 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:32.864 "dma_device_type": 2 00:13:32.864 } 00:13:32.864 ], 00:13:32.864 "driver_specific": {} 00:13:32.864 } 00:13:32.864 ] 00:13:32.864 11:24:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:13:32.864 11:24:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:13:32.864 11:24:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:13:32.864 11:24:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:13:32.864 11:24:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:13:32.864 11:24:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:13:32.864 11:24:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:13:32.865 11:24:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:13:32.865 11:24:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:13:32.865 11:24:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:32.865 11:24:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:32.865 11:24:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:32.865 11:24:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:32.865 11:24:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:32.865 11:24:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:33.122 11:24:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:33.122 "name": "Existed_Raid", 00:13:33.122 "uuid": "ddeadf4b-90ac-466c-ace6-aeda6ae58f72", 00:13:33.122 "strip_size_kb": 64, 00:13:33.122 "state": "online", 00:13:33.122 "raid_level": "raid0", 00:13:33.122 "superblock": false, 00:13:33.122 "num_base_bdevs": 3, 00:13:33.122 "num_base_bdevs_discovered": 3, 00:13:33.122 "num_base_bdevs_operational": 3, 00:13:33.122 "base_bdevs_list": [ 00:13:33.122 { 00:13:33.122 "name": "BaseBdev1", 00:13:33.122 "uuid": "7639f8b2-fe30-4f82-93d0-fe330e4c832c", 00:13:33.122 "is_configured": true, 00:13:33.122 "data_offset": 0, 00:13:33.122 "data_size": 65536 00:13:33.122 }, 00:13:33.122 { 00:13:33.122 "name": "BaseBdev2", 00:13:33.122 "uuid": "2c95bd9e-0b98-4fb2-b811-a916674c37e8", 00:13:33.122 "is_configured": true, 00:13:33.122 "data_offset": 0, 00:13:33.122 "data_size": 65536 00:13:33.122 }, 00:13:33.122 { 00:13:33.122 "name": "BaseBdev3", 00:13:33.122 "uuid": "e37d5fe0-92b2-4517-a8be-b7ecad38b278", 00:13:33.122 "is_configured": true, 00:13:33.122 "data_offset": 0, 00:13:33.122 "data_size": 65536 00:13:33.122 } 00:13:33.122 ] 00:13:33.122 }' 00:13:33.122 11:24:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:33.122 11:24:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.688 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:13:33.688 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:13:33.688 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:13:33.688 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:13:33.688 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:13:33.688 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # local name 00:13:33.688 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:13:33.688 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:13:33.945 [2024-07-15 11:24:17.338124] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:33.945 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:13:33.945 "name": "Existed_Raid", 00:13:33.945 "aliases": [ 00:13:33.945 "ddeadf4b-90ac-466c-ace6-aeda6ae58f72" 00:13:33.945 ], 00:13:33.945 "product_name": "Raid Volume", 00:13:33.945 "block_size": 512, 00:13:33.945 "num_blocks": 196608, 00:13:33.945 "uuid": "ddeadf4b-90ac-466c-ace6-aeda6ae58f72", 00:13:33.945 "assigned_rate_limits": { 00:13:33.945 "rw_ios_per_sec": 0, 00:13:33.945 "rw_mbytes_per_sec": 0, 00:13:33.945 "r_mbytes_per_sec": 0, 00:13:33.945 "w_mbytes_per_sec": 0 00:13:33.945 }, 00:13:33.945 "claimed": false, 00:13:33.945 "zoned": false, 00:13:33.945 "supported_io_types": { 00:13:33.945 "read": true, 00:13:33.945 "write": true, 00:13:33.945 "unmap": true, 00:13:33.946 "flush": true, 00:13:33.946 "reset": true, 00:13:33.946 "nvme_admin": false, 00:13:33.946 "nvme_io": false, 00:13:33.946 "nvme_io_md": false, 00:13:33.946 "write_zeroes": true, 00:13:33.946 "zcopy": false, 00:13:33.946 "get_zone_info": false, 00:13:33.946 "zone_management": false, 00:13:33.946 "zone_append": false, 00:13:33.946 "compare": false, 00:13:33.946 "compare_and_write": false, 00:13:33.946 "abort": false, 00:13:33.946 "seek_hole": false, 00:13:33.946 "seek_data": false, 00:13:33.946 "copy": false, 00:13:33.946 "nvme_iov_md": false 00:13:33.946 }, 00:13:33.946 "memory_domains": [ 00:13:33.946 { 00:13:33.946 "dma_device_id": "system", 00:13:33.946 "dma_device_type": 1 00:13:33.946 }, 00:13:33.946 { 00:13:33.946 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:33.946 "dma_device_type": 2 00:13:33.946 }, 00:13:33.946 { 00:13:33.946 "dma_device_id": "system", 00:13:33.946 "dma_device_type": 1 00:13:33.946 }, 00:13:33.946 { 00:13:33.946 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:33.946 "dma_device_type": 2 00:13:33.946 }, 00:13:33.946 { 00:13:33.946 "dma_device_id": "system", 00:13:33.946 "dma_device_type": 1 00:13:33.946 }, 00:13:33.946 { 00:13:33.946 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:33.946 "dma_device_type": 2 00:13:33.946 } 00:13:33.946 ], 00:13:33.946 "driver_specific": { 00:13:33.946 "raid": { 00:13:33.946 "uuid": "ddeadf4b-90ac-466c-ace6-aeda6ae58f72", 00:13:33.946 "strip_size_kb": 64, 00:13:33.946 "state": "online", 00:13:33.946 "raid_level": "raid0", 00:13:33.946 "superblock": false, 00:13:33.946 "num_base_bdevs": 3, 00:13:33.946 "num_base_bdevs_discovered": 3, 00:13:33.946 "num_base_bdevs_operational": 3, 00:13:33.946 "base_bdevs_list": [ 00:13:33.946 { 00:13:33.946 "name": "BaseBdev1", 00:13:33.946 "uuid": "7639f8b2-fe30-4f82-93d0-fe330e4c832c", 00:13:33.946 "is_configured": true, 00:13:33.946 "data_offset": 0, 00:13:33.946 "data_size": 65536 00:13:33.946 }, 00:13:33.946 { 00:13:33.946 "name": "BaseBdev2", 00:13:33.946 "uuid": "2c95bd9e-0b98-4fb2-b811-a916674c37e8", 00:13:33.946 "is_configured": true, 00:13:33.946 "data_offset": 0, 00:13:33.946 "data_size": 65536 00:13:33.946 }, 00:13:33.946 { 00:13:33.946 "name": "BaseBdev3", 00:13:33.946 "uuid": "e37d5fe0-92b2-4517-a8be-b7ecad38b278", 00:13:33.946 "is_configured": true, 00:13:33.946 "data_offset": 0, 00:13:33.946 "data_size": 65536 00:13:33.946 } 00:13:33.946 ] 00:13:33.946 } 00:13:33.946 } 00:13:33.946 }' 00:13:33.946 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:33.946 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:13:33.946 BaseBdev2 00:13:33.946 BaseBdev3' 00:13:33.946 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:13:33.946 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:13:33.946 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:13:34.204 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:13:34.204 "name": "BaseBdev1", 00:13:34.204 "aliases": [ 00:13:34.204 "7639f8b2-fe30-4f82-93d0-fe330e4c832c" 00:13:34.204 ], 00:13:34.204 "product_name": "Malloc disk", 00:13:34.204 "block_size": 512, 00:13:34.204 "num_blocks": 65536, 00:13:34.204 "uuid": "7639f8b2-fe30-4f82-93d0-fe330e4c832c", 00:13:34.204 "assigned_rate_limits": { 00:13:34.204 "rw_ios_per_sec": 0, 00:13:34.204 "rw_mbytes_per_sec": 0, 00:13:34.204 "r_mbytes_per_sec": 0, 00:13:34.204 "w_mbytes_per_sec": 0 00:13:34.204 }, 00:13:34.204 "claimed": true, 00:13:34.204 "claim_type": "exclusive_write", 00:13:34.204 "zoned": false, 00:13:34.204 "supported_io_types": { 00:13:34.204 "read": true, 00:13:34.204 "write": true, 00:13:34.204 "unmap": true, 00:13:34.204 "flush": true, 00:13:34.204 "reset": true, 00:13:34.204 "nvme_admin": false, 00:13:34.204 "nvme_io": false, 00:13:34.204 "nvme_io_md": false, 00:13:34.204 "write_zeroes": true, 00:13:34.204 "zcopy": true, 00:13:34.204 "get_zone_info": false, 00:13:34.204 "zone_management": false, 00:13:34.204 "zone_append": false, 00:13:34.204 "compare": false, 00:13:34.204 "compare_and_write": false, 00:13:34.204 "abort": true, 00:13:34.204 "seek_hole": false, 00:13:34.204 "seek_data": false, 00:13:34.204 "copy": true, 00:13:34.204 "nvme_iov_md": false 00:13:34.204 }, 00:13:34.204 "memory_domains": [ 00:13:34.204 { 00:13:34.204 "dma_device_id": "system", 00:13:34.204 "dma_device_type": 1 00:13:34.204 }, 00:13:34.204 { 00:13:34.204 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:34.204 "dma_device_type": 2 00:13:34.204 } 00:13:34.204 ], 00:13:34.204 "driver_specific": {} 00:13:34.204 }' 00:13:34.204 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:13:34.204 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:13:34.204 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:13:34.204 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:13:34.204 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:13:34.204 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:13:34.204 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:13:34.204 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:13:34.461 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:13:34.461 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:13:34.461 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:13:34.461 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:13:34.461 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:13:34.461 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:13:34.461 11:24:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:13:34.722 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:13:34.722 "name": "BaseBdev2", 00:13:34.722 "aliases": [ 00:13:34.722 "2c95bd9e-0b98-4fb2-b811-a916674c37e8" 00:13:34.722 ], 00:13:34.722 "product_name": "Malloc disk", 00:13:34.722 "block_size": 512, 00:13:34.722 "num_blocks": 65536, 00:13:34.722 "uuid": "2c95bd9e-0b98-4fb2-b811-a916674c37e8", 00:13:34.722 "assigned_rate_limits": { 00:13:34.722 "rw_ios_per_sec": 0, 00:13:34.722 "rw_mbytes_per_sec": 0, 00:13:34.722 "r_mbytes_per_sec": 0, 00:13:34.722 "w_mbytes_per_sec": 0 00:13:34.722 }, 00:13:34.722 "claimed": true, 00:13:34.722 "claim_type": "exclusive_write", 00:13:34.722 "zoned": false, 00:13:34.722 "supported_io_types": { 00:13:34.722 "read": true, 00:13:34.722 "write": true, 00:13:34.722 "unmap": true, 00:13:34.722 "flush": true, 00:13:34.722 "reset": true, 00:13:34.722 "nvme_admin": false, 00:13:34.722 "nvme_io": false, 00:13:34.722 "nvme_io_md": false, 00:13:34.722 "write_zeroes": true, 00:13:34.722 "zcopy": true, 00:13:34.722 "get_zone_info": false, 00:13:34.722 "zone_management": false, 00:13:34.722 "zone_append": false, 00:13:34.722 "compare": false, 00:13:34.722 "compare_and_write": false, 00:13:34.722 "abort": true, 00:13:34.722 "seek_hole": false, 00:13:34.722 "seek_data": false, 00:13:34.722 "copy": true, 00:13:34.722 "nvme_iov_md": false 00:13:34.722 }, 00:13:34.722 "memory_domains": [ 00:13:34.722 { 00:13:34.722 "dma_device_id": "system", 00:13:34.722 "dma_device_type": 1 00:13:34.722 }, 00:13:34.722 { 00:13:34.722 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:34.722 "dma_device_type": 2 00:13:34.722 } 00:13:34.722 ], 00:13:34.722 "driver_specific": {} 00:13:34.722 }' 00:13:34.722 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:13:34.722 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:13:34.722 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:13:34.722 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:13:34.722 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:13:34.979 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:13:34.979 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:13:34.979 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:13:34.979 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:13:34.979 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:13:34.980 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:13:34.980 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:13:34.980 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:13:34.980 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:13:34.980 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:13:35.237 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:13:35.237 "name": "BaseBdev3", 00:13:35.237 "aliases": [ 00:13:35.237 "e37d5fe0-92b2-4517-a8be-b7ecad38b278" 00:13:35.237 ], 00:13:35.237 "product_name": "Malloc disk", 00:13:35.237 "block_size": 512, 00:13:35.237 "num_blocks": 65536, 00:13:35.237 "uuid": "e37d5fe0-92b2-4517-a8be-b7ecad38b278", 00:13:35.237 "assigned_rate_limits": { 00:13:35.237 "rw_ios_per_sec": 0, 00:13:35.237 "rw_mbytes_per_sec": 0, 00:13:35.237 "r_mbytes_per_sec": 0, 00:13:35.237 "w_mbytes_per_sec": 0 00:13:35.237 }, 00:13:35.237 "claimed": true, 00:13:35.237 "claim_type": "exclusive_write", 00:13:35.237 "zoned": false, 00:13:35.237 "supported_io_types": { 00:13:35.237 "read": true, 00:13:35.237 "write": true, 00:13:35.237 "unmap": true, 00:13:35.237 "flush": true, 00:13:35.237 "reset": true, 00:13:35.237 "nvme_admin": false, 00:13:35.237 "nvme_io": false, 00:13:35.237 "nvme_io_md": false, 00:13:35.237 "write_zeroes": true, 00:13:35.237 "zcopy": true, 00:13:35.237 "get_zone_info": false, 00:13:35.237 "zone_management": false, 00:13:35.237 "zone_append": false, 00:13:35.237 "compare": false, 00:13:35.237 "compare_and_write": false, 00:13:35.237 "abort": true, 00:13:35.237 "seek_hole": false, 00:13:35.237 "seek_data": false, 00:13:35.237 "copy": true, 00:13:35.237 "nvme_iov_md": false 00:13:35.237 }, 00:13:35.237 "memory_domains": [ 00:13:35.237 { 00:13:35.237 "dma_device_id": "system", 00:13:35.237 "dma_device_type": 1 00:13:35.237 }, 00:13:35.237 { 00:13:35.237 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:35.237 "dma_device_type": 2 00:13:35.237 } 00:13:35.237 ], 00:13:35.237 "driver_specific": {} 00:13:35.237 }' 00:13:35.237 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:13:35.237 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:13:35.494 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:13:35.494 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:13:35.494 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:13:35.494 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:13:35.494 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:13:35.494 11:24:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:13:35.494 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:13:35.494 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:13:35.494 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:13:35.494 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:13:35.494 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:13:35.752 [2024-07-15 11:24:19.303094] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:35.752 [2024-07-15 11:24:19.303120] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:35.752 [2024-07-15 11:24:19.303160] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:35.752 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@275 -- # local expected_state 00:13:35.752 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # has_redundancy raid0 00:13:35.752 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:13:35.752 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # return 1 00:13:35.752 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@277 -- # expected_state=offline 00:13:35.752 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:13:35.752 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:13:35.752 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=offline 00:13:35.752 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:13:35.752 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:13:35.752 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:13:35.752 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:35.752 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:35.752 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:35.752 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:35.752 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:35.752 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:36.010 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:36.010 "name": "Existed_Raid", 00:13:36.010 "uuid": "ddeadf4b-90ac-466c-ace6-aeda6ae58f72", 00:13:36.010 "strip_size_kb": 64, 00:13:36.010 "state": "offline", 00:13:36.010 "raid_level": "raid0", 00:13:36.010 "superblock": false, 00:13:36.010 "num_base_bdevs": 3, 00:13:36.010 "num_base_bdevs_discovered": 2, 00:13:36.010 "num_base_bdevs_operational": 2, 00:13:36.010 "base_bdevs_list": [ 00:13:36.010 { 00:13:36.010 "name": null, 00:13:36.010 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:36.010 "is_configured": false, 00:13:36.010 "data_offset": 0, 00:13:36.010 "data_size": 65536 00:13:36.010 }, 00:13:36.010 { 00:13:36.010 "name": "BaseBdev2", 00:13:36.010 "uuid": "2c95bd9e-0b98-4fb2-b811-a916674c37e8", 00:13:36.010 "is_configured": true, 00:13:36.010 "data_offset": 0, 00:13:36.010 "data_size": 65536 00:13:36.010 }, 00:13:36.010 { 00:13:36.010 "name": "BaseBdev3", 00:13:36.010 "uuid": "e37d5fe0-92b2-4517-a8be-b7ecad38b278", 00:13:36.010 "is_configured": true, 00:13:36.010 "data_offset": 0, 00:13:36.010 "data_size": 65536 00:13:36.010 } 00:13:36.010 ] 00:13:36.010 }' 00:13:36.010 11:24:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:36.010 11:24:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.576 11:24:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:13:36.576 11:24:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:13:36.576 11:24:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:13:36.576 11:24:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:36.841 11:24:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:13:36.841 11:24:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:36.841 11:24:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:13:37.100 [2024-07-15 11:24:20.596548] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:37.100 11:24:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:13:37.100 11:24:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:13:37.100 11:24:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:37.100 11:24:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:13:37.358 11:24:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:13:37.359 11:24:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:37.359 11:24:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev3 00:13:37.617 [2024-07-15 11:24:21.074169] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:37.617 [2024-07-15 11:24:21.074210] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xfc4400 name Existed_Raid, state offline 00:13:37.617 11:24:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:13:37.617 11:24:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:13:37.617 11:24:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:37.617 11:24:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:13:37.875 11:24:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:13:37.875 11:24:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:13:37.875 11:24:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # '[' 3 -gt 2 ']' 00:13:37.875 11:24:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i = 1 )) 00:13:37.875 11:24:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:13:37.875 11:24:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:13:38.134 BaseBdev2 00:13:38.134 11:24:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev2 00:13:38.134 11:24:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:13:38.134 11:24:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:13:38.134 11:24:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:13:38.134 11:24:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:13:38.134 11:24:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:13:38.134 11:24:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:13:38.392 11:24:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:38.392 [ 00:13:38.392 { 00:13:38.392 "name": "BaseBdev2", 00:13:38.392 "aliases": [ 00:13:38.392 "110240ba-ef9c-401c-97a5-6a1908352bdf" 00:13:38.392 ], 00:13:38.392 "product_name": "Malloc disk", 00:13:38.392 "block_size": 512, 00:13:38.392 "num_blocks": 65536, 00:13:38.392 "uuid": "110240ba-ef9c-401c-97a5-6a1908352bdf", 00:13:38.392 "assigned_rate_limits": { 00:13:38.392 "rw_ios_per_sec": 0, 00:13:38.392 "rw_mbytes_per_sec": 0, 00:13:38.392 "r_mbytes_per_sec": 0, 00:13:38.392 "w_mbytes_per_sec": 0 00:13:38.392 }, 00:13:38.392 "claimed": false, 00:13:38.392 "zoned": false, 00:13:38.393 "supported_io_types": { 00:13:38.393 "read": true, 00:13:38.393 "write": true, 00:13:38.393 "unmap": true, 00:13:38.393 "flush": true, 00:13:38.393 "reset": true, 00:13:38.393 "nvme_admin": false, 00:13:38.393 "nvme_io": false, 00:13:38.393 "nvme_io_md": false, 00:13:38.393 "write_zeroes": true, 00:13:38.393 "zcopy": true, 00:13:38.393 "get_zone_info": false, 00:13:38.393 "zone_management": false, 00:13:38.393 "zone_append": false, 00:13:38.393 "compare": false, 00:13:38.393 "compare_and_write": false, 00:13:38.393 "abort": true, 00:13:38.393 "seek_hole": false, 00:13:38.393 "seek_data": false, 00:13:38.393 "copy": true, 00:13:38.393 "nvme_iov_md": false 00:13:38.393 }, 00:13:38.393 "memory_domains": [ 00:13:38.393 { 00:13:38.393 "dma_device_id": "system", 00:13:38.393 "dma_device_type": 1 00:13:38.393 }, 00:13:38.393 { 00:13:38.393 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:38.393 "dma_device_type": 2 00:13:38.393 } 00:13:38.393 ], 00:13:38.393 "driver_specific": {} 00:13:38.393 } 00:13:38.393 ] 00:13:38.393 11:24:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:13:38.393 11:24:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:13:38.393 11:24:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:13:38.393 11:24:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:13:38.651 BaseBdev3 00:13:38.651 11:24:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev3 00:13:38.651 11:24:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:13:38.651 11:24:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:13:38.651 11:24:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:13:38.651 11:24:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:13:38.651 11:24:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:13:38.651 11:24:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:13:38.909 11:24:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:39.168 [ 00:13:39.168 { 00:13:39.168 "name": "BaseBdev3", 00:13:39.168 "aliases": [ 00:13:39.168 "c27280e8-ac90-4f30-b527-9137a1ac7b26" 00:13:39.168 ], 00:13:39.168 "product_name": "Malloc disk", 00:13:39.168 "block_size": 512, 00:13:39.168 "num_blocks": 65536, 00:13:39.168 "uuid": "c27280e8-ac90-4f30-b527-9137a1ac7b26", 00:13:39.168 "assigned_rate_limits": { 00:13:39.168 "rw_ios_per_sec": 0, 00:13:39.168 "rw_mbytes_per_sec": 0, 00:13:39.168 "r_mbytes_per_sec": 0, 00:13:39.168 "w_mbytes_per_sec": 0 00:13:39.168 }, 00:13:39.168 "claimed": false, 00:13:39.168 "zoned": false, 00:13:39.168 "supported_io_types": { 00:13:39.168 "read": true, 00:13:39.168 "write": true, 00:13:39.168 "unmap": true, 00:13:39.168 "flush": true, 00:13:39.168 "reset": true, 00:13:39.168 "nvme_admin": false, 00:13:39.168 "nvme_io": false, 00:13:39.168 "nvme_io_md": false, 00:13:39.168 "write_zeroes": true, 00:13:39.168 "zcopy": true, 00:13:39.168 "get_zone_info": false, 00:13:39.168 "zone_management": false, 00:13:39.168 "zone_append": false, 00:13:39.168 "compare": false, 00:13:39.168 "compare_and_write": false, 00:13:39.168 "abort": true, 00:13:39.168 "seek_hole": false, 00:13:39.168 "seek_data": false, 00:13:39.168 "copy": true, 00:13:39.168 "nvme_iov_md": false 00:13:39.168 }, 00:13:39.168 "memory_domains": [ 00:13:39.168 { 00:13:39.168 "dma_device_id": "system", 00:13:39.168 "dma_device_type": 1 00:13:39.168 }, 00:13:39.168 { 00:13:39.168 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:39.168 "dma_device_type": 2 00:13:39.168 } 00:13:39.168 ], 00:13:39.168 "driver_specific": {} 00:13:39.168 } 00:13:39.168 ] 00:13:39.168 11:24:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:13:39.168 11:24:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:13:39.168 11:24:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:13:39.168 11:24:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@305 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:13:39.427 [2024-07-15 11:24:22.825175] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:39.427 [2024-07-15 11:24:22.825215] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:39.427 [2024-07-15 11:24:22.825233] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:39.427 [2024-07-15 11:24:22.826585] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:39.427 11:24:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:13:39.427 11:24:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:13:39.427 11:24:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:13:39.427 11:24:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:13:39.427 11:24:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:13:39.427 11:24:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:13:39.427 11:24:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:39.427 11:24:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:39.427 11:24:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:39.427 11:24:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:39.427 11:24:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:39.427 11:24:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:39.686 11:24:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:39.686 "name": "Existed_Raid", 00:13:39.686 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:39.686 "strip_size_kb": 64, 00:13:39.686 "state": "configuring", 00:13:39.686 "raid_level": "raid0", 00:13:39.686 "superblock": false, 00:13:39.686 "num_base_bdevs": 3, 00:13:39.686 "num_base_bdevs_discovered": 2, 00:13:39.686 "num_base_bdevs_operational": 3, 00:13:39.686 "base_bdevs_list": [ 00:13:39.686 { 00:13:39.686 "name": "BaseBdev1", 00:13:39.686 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:39.686 "is_configured": false, 00:13:39.686 "data_offset": 0, 00:13:39.686 "data_size": 0 00:13:39.686 }, 00:13:39.686 { 00:13:39.686 "name": "BaseBdev2", 00:13:39.686 "uuid": "110240ba-ef9c-401c-97a5-6a1908352bdf", 00:13:39.686 "is_configured": true, 00:13:39.686 "data_offset": 0, 00:13:39.686 "data_size": 65536 00:13:39.686 }, 00:13:39.686 { 00:13:39.686 "name": "BaseBdev3", 00:13:39.686 "uuid": "c27280e8-ac90-4f30-b527-9137a1ac7b26", 00:13:39.686 "is_configured": true, 00:13:39.686 "data_offset": 0, 00:13:39.686 "data_size": 65536 00:13:39.686 } 00:13:39.686 ] 00:13:39.686 }' 00:13:39.686 11:24:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:39.686 11:24:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.308 11:24:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev2 00:13:40.579 [2024-07-15 11:24:23.912054] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:40.579 11:24:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@309 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:13:40.579 11:24:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:13:40.580 11:24:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:13:40.580 11:24:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:13:40.580 11:24:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:13:40.580 11:24:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:13:40.580 11:24:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:40.580 11:24:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:40.580 11:24:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:40.580 11:24:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:40.580 11:24:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:40.580 11:24:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:40.838 11:24:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:40.838 "name": "Existed_Raid", 00:13:40.838 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:40.838 "strip_size_kb": 64, 00:13:40.838 "state": "configuring", 00:13:40.838 "raid_level": "raid0", 00:13:40.838 "superblock": false, 00:13:40.838 "num_base_bdevs": 3, 00:13:40.838 "num_base_bdevs_discovered": 1, 00:13:40.838 "num_base_bdevs_operational": 3, 00:13:40.838 "base_bdevs_list": [ 00:13:40.838 { 00:13:40.838 "name": "BaseBdev1", 00:13:40.838 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:40.838 "is_configured": false, 00:13:40.838 "data_offset": 0, 00:13:40.838 "data_size": 0 00:13:40.838 }, 00:13:40.838 { 00:13:40.838 "name": null, 00:13:40.838 "uuid": "110240ba-ef9c-401c-97a5-6a1908352bdf", 00:13:40.838 "is_configured": false, 00:13:40.838 "data_offset": 0, 00:13:40.838 "data_size": 65536 00:13:40.838 }, 00:13:40.838 { 00:13:40.838 "name": "BaseBdev3", 00:13:40.838 "uuid": "c27280e8-ac90-4f30-b527-9137a1ac7b26", 00:13:40.838 "is_configured": true, 00:13:40.838 "data_offset": 0, 00:13:40.838 "data_size": 65536 00:13:40.838 } 00:13:40.838 ] 00:13:40.838 }' 00:13:40.838 11:24:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:40.838 11:24:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.403 11:24:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:41.403 11:24:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:41.661 11:24:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # [[ false == \f\a\l\s\e ]] 00:13:41.661 11:24:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:13:41.661 [2024-07-15 11:24:25.252186] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:41.661 BaseBdev1 00:13:41.919 11:24:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@313 -- # waitforbdev BaseBdev1 00:13:41.919 11:24:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:13:41.919 11:24:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:13:41.919 11:24:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:13:41.919 11:24:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:13:41.920 11:24:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:13:41.920 11:24:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:13:41.920 11:24:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:42.177 [ 00:13:42.177 { 00:13:42.177 "name": "BaseBdev1", 00:13:42.177 "aliases": [ 00:13:42.177 "56caf595-7156-4a78-8ca6-8846009478de" 00:13:42.177 ], 00:13:42.177 "product_name": "Malloc disk", 00:13:42.177 "block_size": 512, 00:13:42.177 "num_blocks": 65536, 00:13:42.177 "uuid": "56caf595-7156-4a78-8ca6-8846009478de", 00:13:42.177 "assigned_rate_limits": { 00:13:42.177 "rw_ios_per_sec": 0, 00:13:42.177 "rw_mbytes_per_sec": 0, 00:13:42.177 "r_mbytes_per_sec": 0, 00:13:42.177 "w_mbytes_per_sec": 0 00:13:42.177 }, 00:13:42.177 "claimed": true, 00:13:42.177 "claim_type": "exclusive_write", 00:13:42.177 "zoned": false, 00:13:42.177 "supported_io_types": { 00:13:42.177 "read": true, 00:13:42.177 "write": true, 00:13:42.177 "unmap": true, 00:13:42.177 "flush": true, 00:13:42.177 "reset": true, 00:13:42.177 "nvme_admin": false, 00:13:42.177 "nvme_io": false, 00:13:42.177 "nvme_io_md": false, 00:13:42.177 "write_zeroes": true, 00:13:42.177 "zcopy": true, 00:13:42.177 "get_zone_info": false, 00:13:42.177 "zone_management": false, 00:13:42.177 "zone_append": false, 00:13:42.177 "compare": false, 00:13:42.177 "compare_and_write": false, 00:13:42.177 "abort": true, 00:13:42.177 "seek_hole": false, 00:13:42.177 "seek_data": false, 00:13:42.177 "copy": true, 00:13:42.177 "nvme_iov_md": false 00:13:42.177 }, 00:13:42.177 "memory_domains": [ 00:13:42.177 { 00:13:42.177 "dma_device_id": "system", 00:13:42.177 "dma_device_type": 1 00:13:42.177 }, 00:13:42.177 { 00:13:42.177 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:42.177 "dma_device_type": 2 00:13:42.177 } 00:13:42.177 ], 00:13:42.177 "driver_specific": {} 00:13:42.177 } 00:13:42.177 ] 00:13:42.177 11:24:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:13:42.178 11:24:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:13:42.178 11:24:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:13:42.178 11:24:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:13:42.178 11:24:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:13:42.178 11:24:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:13:42.178 11:24:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:13:42.178 11:24:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:42.178 11:24:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:42.178 11:24:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:42.178 11:24:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:42.178 11:24:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:42.178 11:24:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:42.437 11:24:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:42.437 "name": "Existed_Raid", 00:13:42.437 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:42.437 "strip_size_kb": 64, 00:13:42.437 "state": "configuring", 00:13:42.437 "raid_level": "raid0", 00:13:42.437 "superblock": false, 00:13:42.437 "num_base_bdevs": 3, 00:13:42.437 "num_base_bdevs_discovered": 2, 00:13:42.437 "num_base_bdevs_operational": 3, 00:13:42.437 "base_bdevs_list": [ 00:13:42.437 { 00:13:42.437 "name": "BaseBdev1", 00:13:42.437 "uuid": "56caf595-7156-4a78-8ca6-8846009478de", 00:13:42.437 "is_configured": true, 00:13:42.437 "data_offset": 0, 00:13:42.437 "data_size": 65536 00:13:42.437 }, 00:13:42.437 { 00:13:42.437 "name": null, 00:13:42.437 "uuid": "110240ba-ef9c-401c-97a5-6a1908352bdf", 00:13:42.437 "is_configured": false, 00:13:42.437 "data_offset": 0, 00:13:42.437 "data_size": 65536 00:13:42.437 }, 00:13:42.437 { 00:13:42.437 "name": "BaseBdev3", 00:13:42.437 "uuid": "c27280e8-ac90-4f30-b527-9137a1ac7b26", 00:13:42.437 "is_configured": true, 00:13:42.437 "data_offset": 0, 00:13:42.437 "data_size": 65536 00:13:42.437 } 00:13:42.437 ] 00:13:42.437 }' 00:13:42.437 11:24:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:42.437 11:24:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.013 11:24:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:43.013 11:24:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:43.273 11:24:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # [[ true == \t\r\u\e ]] 00:13:43.273 11:24:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@317 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev3 00:13:43.531 [2024-07-15 11:24:26.880532] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:43.531 11:24:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:13:43.532 11:24:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:13:43.532 11:24:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:13:43.532 11:24:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:13:43.532 11:24:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:13:43.532 11:24:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:13:43.532 11:24:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:43.532 11:24:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:43.532 11:24:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:43.532 11:24:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:43.532 11:24:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:43.532 11:24:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:43.790 11:24:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:43.790 "name": "Existed_Raid", 00:13:43.790 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:43.790 "strip_size_kb": 64, 00:13:43.790 "state": "configuring", 00:13:43.790 "raid_level": "raid0", 00:13:43.790 "superblock": false, 00:13:43.790 "num_base_bdevs": 3, 00:13:43.790 "num_base_bdevs_discovered": 1, 00:13:43.790 "num_base_bdevs_operational": 3, 00:13:43.790 "base_bdevs_list": [ 00:13:43.790 { 00:13:43.790 "name": "BaseBdev1", 00:13:43.790 "uuid": "56caf595-7156-4a78-8ca6-8846009478de", 00:13:43.790 "is_configured": true, 00:13:43.790 "data_offset": 0, 00:13:43.790 "data_size": 65536 00:13:43.790 }, 00:13:43.790 { 00:13:43.790 "name": null, 00:13:43.790 "uuid": "110240ba-ef9c-401c-97a5-6a1908352bdf", 00:13:43.790 "is_configured": false, 00:13:43.790 "data_offset": 0, 00:13:43.790 "data_size": 65536 00:13:43.790 }, 00:13:43.790 { 00:13:43.790 "name": null, 00:13:43.790 "uuid": "c27280e8-ac90-4f30-b527-9137a1ac7b26", 00:13:43.790 "is_configured": false, 00:13:43.790 "data_offset": 0, 00:13:43.790 "data_size": 65536 00:13:43.790 } 00:13:43.790 ] 00:13:43.790 }' 00:13:43.790 11:24:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:43.790 11:24:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:44.356 11:24:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:44.356 11:24:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:44.614 11:24:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # [[ false == \f\a\l\s\e ]] 00:13:44.614 11:24:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:13:44.614 [2024-07-15 11:24:28.196034] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:44.872 11:24:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@322 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:13:44.872 11:24:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:13:44.872 11:24:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:13:44.872 11:24:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:13:44.872 11:24:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:13:44.872 11:24:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:13:44.872 11:24:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:44.872 11:24:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:44.872 11:24:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:44.872 11:24:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:44.872 11:24:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:44.872 11:24:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:44.872 11:24:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:44.872 "name": "Existed_Raid", 00:13:44.872 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:44.872 "strip_size_kb": 64, 00:13:44.872 "state": "configuring", 00:13:44.872 "raid_level": "raid0", 00:13:44.872 "superblock": false, 00:13:44.872 "num_base_bdevs": 3, 00:13:44.872 "num_base_bdevs_discovered": 2, 00:13:44.872 "num_base_bdevs_operational": 3, 00:13:44.872 "base_bdevs_list": [ 00:13:44.872 { 00:13:44.872 "name": "BaseBdev1", 00:13:44.873 "uuid": "56caf595-7156-4a78-8ca6-8846009478de", 00:13:44.873 "is_configured": true, 00:13:44.873 "data_offset": 0, 00:13:44.873 "data_size": 65536 00:13:44.873 }, 00:13:44.873 { 00:13:44.873 "name": null, 00:13:44.873 "uuid": "110240ba-ef9c-401c-97a5-6a1908352bdf", 00:13:44.873 "is_configured": false, 00:13:44.873 "data_offset": 0, 00:13:44.873 "data_size": 65536 00:13:44.873 }, 00:13:44.873 { 00:13:44.873 "name": "BaseBdev3", 00:13:44.873 "uuid": "c27280e8-ac90-4f30-b527-9137a1ac7b26", 00:13:44.873 "is_configured": true, 00:13:44.873 "data_offset": 0, 00:13:44.873 "data_size": 65536 00:13:44.873 } 00:13:44.873 ] 00:13:44.873 }' 00:13:44.873 11:24:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:44.873 11:24:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:45.437 11:24:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:45.437 11:24:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:45.696 11:24:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # [[ true == \t\r\u\e ]] 00:13:45.696 11:24:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@325 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:13:45.955 [2024-07-15 11:24:29.471442] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:45.955 11:24:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:13:45.955 11:24:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:13:45.955 11:24:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:13:45.955 11:24:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:13:45.955 11:24:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:13:45.955 11:24:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:13:45.955 11:24:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:45.955 11:24:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:45.955 11:24:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:45.955 11:24:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:45.955 11:24:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:45.955 11:24:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:46.213 11:24:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:46.213 "name": "Existed_Raid", 00:13:46.213 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:46.213 "strip_size_kb": 64, 00:13:46.213 "state": "configuring", 00:13:46.214 "raid_level": "raid0", 00:13:46.214 "superblock": false, 00:13:46.214 "num_base_bdevs": 3, 00:13:46.214 "num_base_bdevs_discovered": 1, 00:13:46.214 "num_base_bdevs_operational": 3, 00:13:46.214 "base_bdevs_list": [ 00:13:46.214 { 00:13:46.214 "name": null, 00:13:46.214 "uuid": "56caf595-7156-4a78-8ca6-8846009478de", 00:13:46.214 "is_configured": false, 00:13:46.214 "data_offset": 0, 00:13:46.214 "data_size": 65536 00:13:46.214 }, 00:13:46.214 { 00:13:46.214 "name": null, 00:13:46.214 "uuid": "110240ba-ef9c-401c-97a5-6a1908352bdf", 00:13:46.214 "is_configured": false, 00:13:46.214 "data_offset": 0, 00:13:46.214 "data_size": 65536 00:13:46.214 }, 00:13:46.214 { 00:13:46.214 "name": "BaseBdev3", 00:13:46.214 "uuid": "c27280e8-ac90-4f30-b527-9137a1ac7b26", 00:13:46.214 "is_configured": true, 00:13:46.214 "data_offset": 0, 00:13:46.214 "data_size": 65536 00:13:46.214 } 00:13:46.214 ] 00:13:46.214 }' 00:13:46.214 11:24:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:46.214 11:24:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:46.781 11:24:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@327 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:46.781 11:24:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@327 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:47.039 11:24:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@327 -- # [[ false == \f\a\l\s\e ]] 00:13:47.039 11:24:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@329 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:13:47.298 [2024-07-15 11:24:30.789493] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:47.298 11:24:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@330 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:13:47.298 11:24:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:13:47.298 11:24:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:13:47.298 11:24:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:13:47.298 11:24:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:13:47.298 11:24:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:13:47.298 11:24:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:47.298 11:24:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:47.298 11:24:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:47.298 11:24:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:47.298 11:24:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:47.298 11:24:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:47.556 11:24:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:47.556 "name": "Existed_Raid", 00:13:47.556 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:47.556 "strip_size_kb": 64, 00:13:47.556 "state": "configuring", 00:13:47.556 "raid_level": "raid0", 00:13:47.556 "superblock": false, 00:13:47.556 "num_base_bdevs": 3, 00:13:47.556 "num_base_bdevs_discovered": 2, 00:13:47.556 "num_base_bdevs_operational": 3, 00:13:47.556 "base_bdevs_list": [ 00:13:47.556 { 00:13:47.556 "name": null, 00:13:47.556 "uuid": "56caf595-7156-4a78-8ca6-8846009478de", 00:13:47.556 "is_configured": false, 00:13:47.556 "data_offset": 0, 00:13:47.556 "data_size": 65536 00:13:47.556 }, 00:13:47.556 { 00:13:47.556 "name": "BaseBdev2", 00:13:47.556 "uuid": "110240ba-ef9c-401c-97a5-6a1908352bdf", 00:13:47.556 "is_configured": true, 00:13:47.556 "data_offset": 0, 00:13:47.556 "data_size": 65536 00:13:47.556 }, 00:13:47.556 { 00:13:47.556 "name": "BaseBdev3", 00:13:47.556 "uuid": "c27280e8-ac90-4f30-b527-9137a1ac7b26", 00:13:47.556 "is_configured": true, 00:13:47.556 "data_offset": 0, 00:13:47.556 "data_size": 65536 00:13:47.556 } 00:13:47.556 ] 00:13:47.556 }' 00:13:47.556 11:24:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:47.556 11:24:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:48.123 11:24:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@331 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:48.123 11:24:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@331 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:48.382 11:24:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@331 -- # [[ true == \t\r\u\e ]] 00:13:48.382 11:24:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:48.382 11:24:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@333 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:13:48.950 11:24:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b NewBaseBdev -u 56caf595-7156-4a78-8ca6-8846009478de 00:13:49.208 [2024-07-15 11:24:32.625638] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:13:49.208 [2024-07-15 11:24:32.625675] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xfc2450 00:13:49.208 [2024-07-15 11:24:32.625683] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:13:49.208 [2024-07-15 11:24:32.625873] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xfc3a50 00:13:49.208 [2024-07-15 11:24:32.625998] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xfc2450 00:13:49.208 [2024-07-15 11:24:32.626008] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0xfc2450 00:13:49.208 [2024-07-15 11:24:32.626168] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:49.208 NewBaseBdev 00:13:49.208 11:24:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@334 -- # waitforbdev NewBaseBdev 00:13:49.208 11:24:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=NewBaseBdev 00:13:49.208 11:24:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:13:49.208 11:24:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:13:49.208 11:24:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:13:49.208 11:24:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:13:49.208 11:24:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:13:49.467 11:24:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev -t 2000 00:13:49.726 [ 00:13:49.726 { 00:13:49.726 "name": "NewBaseBdev", 00:13:49.726 "aliases": [ 00:13:49.726 "56caf595-7156-4a78-8ca6-8846009478de" 00:13:49.726 ], 00:13:49.726 "product_name": "Malloc disk", 00:13:49.726 "block_size": 512, 00:13:49.726 "num_blocks": 65536, 00:13:49.726 "uuid": "56caf595-7156-4a78-8ca6-8846009478de", 00:13:49.726 "assigned_rate_limits": { 00:13:49.726 "rw_ios_per_sec": 0, 00:13:49.726 "rw_mbytes_per_sec": 0, 00:13:49.726 "r_mbytes_per_sec": 0, 00:13:49.726 "w_mbytes_per_sec": 0 00:13:49.726 }, 00:13:49.726 "claimed": true, 00:13:49.726 "claim_type": "exclusive_write", 00:13:49.726 "zoned": false, 00:13:49.726 "supported_io_types": { 00:13:49.726 "read": true, 00:13:49.726 "write": true, 00:13:49.726 "unmap": true, 00:13:49.726 "flush": true, 00:13:49.726 "reset": true, 00:13:49.726 "nvme_admin": false, 00:13:49.726 "nvme_io": false, 00:13:49.726 "nvme_io_md": false, 00:13:49.726 "write_zeroes": true, 00:13:49.726 "zcopy": true, 00:13:49.726 "get_zone_info": false, 00:13:49.726 "zone_management": false, 00:13:49.726 "zone_append": false, 00:13:49.726 "compare": false, 00:13:49.726 "compare_and_write": false, 00:13:49.726 "abort": true, 00:13:49.726 "seek_hole": false, 00:13:49.726 "seek_data": false, 00:13:49.726 "copy": true, 00:13:49.726 "nvme_iov_md": false 00:13:49.726 }, 00:13:49.726 "memory_domains": [ 00:13:49.726 { 00:13:49.726 "dma_device_id": "system", 00:13:49.726 "dma_device_type": 1 00:13:49.726 }, 00:13:49.726 { 00:13:49.726 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:49.726 "dma_device_type": 2 00:13:49.726 } 00:13:49.726 ], 00:13:49.726 "driver_specific": {} 00:13:49.726 } 00:13:49.726 ] 00:13:49.726 11:24:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:13:49.726 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@335 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:13:49.726 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:13:49.726 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:13:49.726 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:13:49.726 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:13:49.726 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:13:49.726 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:49.726 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:49.726 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:49.726 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:49.726 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:49.726 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:49.985 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:49.985 "name": "Existed_Raid", 00:13:49.985 "uuid": "7d827979-f34e-4e8c-a932-892afae65d6e", 00:13:49.985 "strip_size_kb": 64, 00:13:49.985 "state": "online", 00:13:49.985 "raid_level": "raid0", 00:13:49.985 "superblock": false, 00:13:49.985 "num_base_bdevs": 3, 00:13:49.985 "num_base_bdevs_discovered": 3, 00:13:49.985 "num_base_bdevs_operational": 3, 00:13:49.985 "base_bdevs_list": [ 00:13:49.985 { 00:13:49.985 "name": "NewBaseBdev", 00:13:49.985 "uuid": "56caf595-7156-4a78-8ca6-8846009478de", 00:13:49.985 "is_configured": true, 00:13:49.985 "data_offset": 0, 00:13:49.985 "data_size": 65536 00:13:49.985 }, 00:13:49.985 { 00:13:49.985 "name": "BaseBdev2", 00:13:49.985 "uuid": "110240ba-ef9c-401c-97a5-6a1908352bdf", 00:13:49.985 "is_configured": true, 00:13:49.985 "data_offset": 0, 00:13:49.985 "data_size": 65536 00:13:49.985 }, 00:13:49.985 { 00:13:49.985 "name": "BaseBdev3", 00:13:49.985 "uuid": "c27280e8-ac90-4f30-b527-9137a1ac7b26", 00:13:49.985 "is_configured": true, 00:13:49.985 "data_offset": 0, 00:13:49.985 "data_size": 65536 00:13:49.985 } 00:13:49.985 ] 00:13:49.985 }' 00:13:49.985 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:49.985 11:24:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.550 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@336 -- # verify_raid_bdev_properties Existed_Raid 00:13:50.550 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:13:50.550 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:13:50.550 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:13:50.550 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:13:50.550 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # local name 00:13:50.550 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:13:50.550 11:24:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:13:50.550 [2024-07-15 11:24:34.133952] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:50.808 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:13:50.808 "name": "Existed_Raid", 00:13:50.808 "aliases": [ 00:13:50.808 "7d827979-f34e-4e8c-a932-892afae65d6e" 00:13:50.808 ], 00:13:50.808 "product_name": "Raid Volume", 00:13:50.808 "block_size": 512, 00:13:50.808 "num_blocks": 196608, 00:13:50.808 "uuid": "7d827979-f34e-4e8c-a932-892afae65d6e", 00:13:50.808 "assigned_rate_limits": { 00:13:50.808 "rw_ios_per_sec": 0, 00:13:50.808 "rw_mbytes_per_sec": 0, 00:13:50.808 "r_mbytes_per_sec": 0, 00:13:50.808 "w_mbytes_per_sec": 0 00:13:50.808 }, 00:13:50.808 "claimed": false, 00:13:50.808 "zoned": false, 00:13:50.808 "supported_io_types": { 00:13:50.808 "read": true, 00:13:50.808 "write": true, 00:13:50.808 "unmap": true, 00:13:50.808 "flush": true, 00:13:50.808 "reset": true, 00:13:50.808 "nvme_admin": false, 00:13:50.808 "nvme_io": false, 00:13:50.808 "nvme_io_md": false, 00:13:50.808 "write_zeroes": true, 00:13:50.808 "zcopy": false, 00:13:50.808 "get_zone_info": false, 00:13:50.808 "zone_management": false, 00:13:50.808 "zone_append": false, 00:13:50.808 "compare": false, 00:13:50.808 "compare_and_write": false, 00:13:50.808 "abort": false, 00:13:50.808 "seek_hole": false, 00:13:50.808 "seek_data": false, 00:13:50.808 "copy": false, 00:13:50.808 "nvme_iov_md": false 00:13:50.808 }, 00:13:50.808 "memory_domains": [ 00:13:50.808 { 00:13:50.808 "dma_device_id": "system", 00:13:50.808 "dma_device_type": 1 00:13:50.808 }, 00:13:50.808 { 00:13:50.808 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:50.808 "dma_device_type": 2 00:13:50.808 }, 00:13:50.808 { 00:13:50.808 "dma_device_id": "system", 00:13:50.808 "dma_device_type": 1 00:13:50.808 }, 00:13:50.808 { 00:13:50.808 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:50.808 "dma_device_type": 2 00:13:50.808 }, 00:13:50.808 { 00:13:50.808 "dma_device_id": "system", 00:13:50.808 "dma_device_type": 1 00:13:50.808 }, 00:13:50.808 { 00:13:50.808 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:50.808 "dma_device_type": 2 00:13:50.808 } 00:13:50.808 ], 00:13:50.808 "driver_specific": { 00:13:50.808 "raid": { 00:13:50.808 "uuid": "7d827979-f34e-4e8c-a932-892afae65d6e", 00:13:50.808 "strip_size_kb": 64, 00:13:50.808 "state": "online", 00:13:50.808 "raid_level": "raid0", 00:13:50.808 "superblock": false, 00:13:50.808 "num_base_bdevs": 3, 00:13:50.808 "num_base_bdevs_discovered": 3, 00:13:50.808 "num_base_bdevs_operational": 3, 00:13:50.808 "base_bdevs_list": [ 00:13:50.808 { 00:13:50.808 "name": "NewBaseBdev", 00:13:50.808 "uuid": "56caf595-7156-4a78-8ca6-8846009478de", 00:13:50.808 "is_configured": true, 00:13:50.808 "data_offset": 0, 00:13:50.808 "data_size": 65536 00:13:50.808 }, 00:13:50.808 { 00:13:50.808 "name": "BaseBdev2", 00:13:50.808 "uuid": "110240ba-ef9c-401c-97a5-6a1908352bdf", 00:13:50.808 "is_configured": true, 00:13:50.808 "data_offset": 0, 00:13:50.808 "data_size": 65536 00:13:50.808 }, 00:13:50.808 { 00:13:50.808 "name": "BaseBdev3", 00:13:50.808 "uuid": "c27280e8-ac90-4f30-b527-9137a1ac7b26", 00:13:50.808 "is_configured": true, 00:13:50.808 "data_offset": 0, 00:13:50.808 "data_size": 65536 00:13:50.808 } 00:13:50.808 ] 00:13:50.808 } 00:13:50.808 } 00:13:50.808 }' 00:13:50.808 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:50.808 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='NewBaseBdev 00:13:50.808 BaseBdev2 00:13:50.808 BaseBdev3' 00:13:50.808 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:13:50.808 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:13:50.808 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev 00:13:51.067 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:13:51.067 "name": "NewBaseBdev", 00:13:51.067 "aliases": [ 00:13:51.067 "56caf595-7156-4a78-8ca6-8846009478de" 00:13:51.067 ], 00:13:51.067 "product_name": "Malloc disk", 00:13:51.067 "block_size": 512, 00:13:51.067 "num_blocks": 65536, 00:13:51.067 "uuid": "56caf595-7156-4a78-8ca6-8846009478de", 00:13:51.067 "assigned_rate_limits": { 00:13:51.067 "rw_ios_per_sec": 0, 00:13:51.067 "rw_mbytes_per_sec": 0, 00:13:51.067 "r_mbytes_per_sec": 0, 00:13:51.067 "w_mbytes_per_sec": 0 00:13:51.067 }, 00:13:51.067 "claimed": true, 00:13:51.067 "claim_type": "exclusive_write", 00:13:51.067 "zoned": false, 00:13:51.067 "supported_io_types": { 00:13:51.067 "read": true, 00:13:51.067 "write": true, 00:13:51.067 "unmap": true, 00:13:51.067 "flush": true, 00:13:51.067 "reset": true, 00:13:51.067 "nvme_admin": false, 00:13:51.067 "nvme_io": false, 00:13:51.067 "nvme_io_md": false, 00:13:51.067 "write_zeroes": true, 00:13:51.067 "zcopy": true, 00:13:51.067 "get_zone_info": false, 00:13:51.067 "zone_management": false, 00:13:51.067 "zone_append": false, 00:13:51.067 "compare": false, 00:13:51.067 "compare_and_write": false, 00:13:51.067 "abort": true, 00:13:51.067 "seek_hole": false, 00:13:51.067 "seek_data": false, 00:13:51.067 "copy": true, 00:13:51.067 "nvme_iov_md": false 00:13:51.067 }, 00:13:51.067 "memory_domains": [ 00:13:51.067 { 00:13:51.067 "dma_device_id": "system", 00:13:51.067 "dma_device_type": 1 00:13:51.067 }, 00:13:51.067 { 00:13:51.067 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:51.067 "dma_device_type": 2 00:13:51.067 } 00:13:51.067 ], 00:13:51.067 "driver_specific": {} 00:13:51.067 }' 00:13:51.067 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:13:51.067 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:13:51.067 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:13:51.067 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:13:51.067 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:13:51.067 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:13:51.067 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:13:51.067 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:13:51.326 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:13:51.326 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:13:51.326 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:13:51.326 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:13:51.326 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:13:51.326 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:13:51.326 11:24:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:13:51.584 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:13:51.584 "name": "BaseBdev2", 00:13:51.584 "aliases": [ 00:13:51.584 "110240ba-ef9c-401c-97a5-6a1908352bdf" 00:13:51.584 ], 00:13:51.584 "product_name": "Malloc disk", 00:13:51.584 "block_size": 512, 00:13:51.584 "num_blocks": 65536, 00:13:51.584 "uuid": "110240ba-ef9c-401c-97a5-6a1908352bdf", 00:13:51.584 "assigned_rate_limits": { 00:13:51.584 "rw_ios_per_sec": 0, 00:13:51.584 "rw_mbytes_per_sec": 0, 00:13:51.584 "r_mbytes_per_sec": 0, 00:13:51.584 "w_mbytes_per_sec": 0 00:13:51.584 }, 00:13:51.584 "claimed": true, 00:13:51.584 "claim_type": "exclusive_write", 00:13:51.584 "zoned": false, 00:13:51.584 "supported_io_types": { 00:13:51.584 "read": true, 00:13:51.584 "write": true, 00:13:51.584 "unmap": true, 00:13:51.584 "flush": true, 00:13:51.584 "reset": true, 00:13:51.584 "nvme_admin": false, 00:13:51.584 "nvme_io": false, 00:13:51.584 "nvme_io_md": false, 00:13:51.584 "write_zeroes": true, 00:13:51.584 "zcopy": true, 00:13:51.584 "get_zone_info": false, 00:13:51.584 "zone_management": false, 00:13:51.584 "zone_append": false, 00:13:51.584 "compare": false, 00:13:51.584 "compare_and_write": false, 00:13:51.584 "abort": true, 00:13:51.584 "seek_hole": false, 00:13:51.584 "seek_data": false, 00:13:51.584 "copy": true, 00:13:51.584 "nvme_iov_md": false 00:13:51.584 }, 00:13:51.584 "memory_domains": [ 00:13:51.584 { 00:13:51.584 "dma_device_id": "system", 00:13:51.584 "dma_device_type": 1 00:13:51.584 }, 00:13:51.584 { 00:13:51.584 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:51.584 "dma_device_type": 2 00:13:51.584 } 00:13:51.584 ], 00:13:51.584 "driver_specific": {} 00:13:51.584 }' 00:13:51.584 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:13:51.584 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:13:51.584 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:13:51.584 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:13:51.584 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:13:51.843 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:13:51.843 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:13:51.843 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:13:51.843 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:13:51.843 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:13:51.843 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:13:51.843 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:13:51.843 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:13:51.843 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:13:51.843 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:13:52.102 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:13:52.102 "name": "BaseBdev3", 00:13:52.102 "aliases": [ 00:13:52.102 "c27280e8-ac90-4f30-b527-9137a1ac7b26" 00:13:52.102 ], 00:13:52.102 "product_name": "Malloc disk", 00:13:52.102 "block_size": 512, 00:13:52.102 "num_blocks": 65536, 00:13:52.102 "uuid": "c27280e8-ac90-4f30-b527-9137a1ac7b26", 00:13:52.102 "assigned_rate_limits": { 00:13:52.102 "rw_ios_per_sec": 0, 00:13:52.102 "rw_mbytes_per_sec": 0, 00:13:52.102 "r_mbytes_per_sec": 0, 00:13:52.102 "w_mbytes_per_sec": 0 00:13:52.102 }, 00:13:52.102 "claimed": true, 00:13:52.102 "claim_type": "exclusive_write", 00:13:52.102 "zoned": false, 00:13:52.102 "supported_io_types": { 00:13:52.102 "read": true, 00:13:52.102 "write": true, 00:13:52.102 "unmap": true, 00:13:52.102 "flush": true, 00:13:52.102 "reset": true, 00:13:52.102 "nvme_admin": false, 00:13:52.102 "nvme_io": false, 00:13:52.102 "nvme_io_md": false, 00:13:52.102 "write_zeroes": true, 00:13:52.102 "zcopy": true, 00:13:52.102 "get_zone_info": false, 00:13:52.102 "zone_management": false, 00:13:52.102 "zone_append": false, 00:13:52.102 "compare": false, 00:13:52.102 "compare_and_write": false, 00:13:52.102 "abort": true, 00:13:52.102 "seek_hole": false, 00:13:52.102 "seek_data": false, 00:13:52.102 "copy": true, 00:13:52.102 "nvme_iov_md": false 00:13:52.102 }, 00:13:52.102 "memory_domains": [ 00:13:52.102 { 00:13:52.102 "dma_device_id": "system", 00:13:52.102 "dma_device_type": 1 00:13:52.102 }, 00:13:52.102 { 00:13:52.102 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:52.102 "dma_device_type": 2 00:13:52.102 } 00:13:52.102 ], 00:13:52.102 "driver_specific": {} 00:13:52.102 }' 00:13:52.102 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:13:52.102 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:13:52.102 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:13:52.102 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:13:52.102 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:13:52.360 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:13:52.360 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:13:52.360 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:13:52.360 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:13:52.360 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:13:52.360 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:13:52.360 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:13:52.360 11:24:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@338 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:13:52.619 [2024-07-15 11:24:36.098867] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:52.619 [2024-07-15 11:24:36.098894] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:52.619 [2024-07-15 11:24:36.098955] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:52.619 [2024-07-15 11:24:36.099005] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:52.619 [2024-07-15 11:24:36.099017] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xfc2450 name Existed_Raid, state offline 00:13:52.619 11:24:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@341 -- # killprocess 882792 00:13:52.619 11:24:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@948 -- # '[' -z 882792 ']' 00:13:52.619 11:24:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@952 -- # kill -0 882792 00:13:52.619 11:24:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@953 -- # uname 00:13:52.619 11:24:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:13:52.619 11:24:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 882792 00:13:52.619 11:24:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:13:52.619 11:24:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:13:52.619 11:24:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 882792' 00:13:52.619 killing process with pid 882792 00:13:52.619 11:24:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@967 -- # kill 882792 00:13:52.619 [2024-07-15 11:24:36.164768] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:52.619 11:24:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # wait 882792 00:13:52.619 [2024-07-15 11:24:36.195786] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:52.877 11:24:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@343 -- # return 0 00:13:52.877 00:13:52.877 real 0m28.018s 00:13:52.877 user 0m51.430s 00:13:52.877 sys 0m4.992s 00:13:52.877 11:24:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:13:52.877 11:24:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.877 ************************************ 00:13:52.877 END TEST raid_state_function_test 00:13:52.877 ************************************ 00:13:52.877 11:24:36 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:13:52.877 11:24:36 bdev_raid -- bdev/bdev_raid.sh@868 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 3 true 00:13:52.877 11:24:36 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:13:52.877 11:24:36 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:13:52.877 11:24:36 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:53.135 ************************************ 00:13:53.135 START TEST raid_state_function_test_sb 00:13:53.135 ************************************ 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1123 -- # raid_state_function_test raid0 3 true 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@220 -- # local raid_level=raid0 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=3 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # local superblock=true 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev3 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # local strip_size 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # '[' raid0 '!=' raid1 ']' 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # strip_size=64 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@232 -- # strip_size_create_arg='-z 64' 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # '[' true = true ']' 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@238 -- # superblock_create_arg=-s 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # raid_pid=887080 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 887080' 00:13:53.135 Process raid pid: 887080 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@246 -- # waitforlisten 887080 /var/tmp/spdk-raid.sock 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@829 -- # '[' -z 887080 ']' 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@834 -- # local max_retries=100 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:13:53.135 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # xtrace_disable 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:13:53.135 11:24:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:53.135 [2024-07-15 11:24:36.567274] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:13:53.135 [2024-07-15 11:24:36.567338] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:53.135 [2024-07-15 11:24:36.697517] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:53.393 [2024-07-15 11:24:36.804783] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:13:53.393 [2024-07-15 11:24:36.871108] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:53.393 [2024-07-15 11:24:36.871142] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:53.959 11:24:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:13:53.959 11:24:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@862 -- # return 0 00:13:53.959 11:24:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r raid0 -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:13:54.218 [2024-07-15 11:24:37.702662] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:54.218 [2024-07-15 11:24:37.702704] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:54.218 [2024-07-15 11:24:37.702715] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:54.218 [2024-07-15 11:24:37.702727] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:54.218 [2024-07-15 11:24:37.702736] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:54.218 [2024-07-15 11:24:37.702748] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:54.218 11:24:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:13:54.218 11:24:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:13:54.218 11:24:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:13:54.218 11:24:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:13:54.218 11:24:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:13:54.218 11:24:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:13:54.218 11:24:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:54.218 11:24:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:54.218 11:24:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:54.218 11:24:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:54.218 11:24:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:54.218 11:24:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:54.477 11:24:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:54.477 "name": "Existed_Raid", 00:13:54.477 "uuid": "d68e5954-63bc-428a-afdd-5c98cc9ccce8", 00:13:54.477 "strip_size_kb": 64, 00:13:54.477 "state": "configuring", 00:13:54.477 "raid_level": "raid0", 00:13:54.477 "superblock": true, 00:13:54.477 "num_base_bdevs": 3, 00:13:54.477 "num_base_bdevs_discovered": 0, 00:13:54.477 "num_base_bdevs_operational": 3, 00:13:54.477 "base_bdevs_list": [ 00:13:54.477 { 00:13:54.477 "name": "BaseBdev1", 00:13:54.477 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:54.477 "is_configured": false, 00:13:54.477 "data_offset": 0, 00:13:54.477 "data_size": 0 00:13:54.477 }, 00:13:54.477 { 00:13:54.477 "name": "BaseBdev2", 00:13:54.477 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:54.477 "is_configured": false, 00:13:54.477 "data_offset": 0, 00:13:54.477 "data_size": 0 00:13:54.477 }, 00:13:54.477 { 00:13:54.477 "name": "BaseBdev3", 00:13:54.477 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:54.477 "is_configured": false, 00:13:54.477 "data_offset": 0, 00:13:54.477 "data_size": 0 00:13:54.477 } 00:13:54.477 ] 00:13:54.477 }' 00:13:54.477 11:24:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:54.477 11:24:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:55.044 11:24:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:13:55.303 [2024-07-15 11:24:38.729215] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:55.303 [2024-07-15 11:24:38.729245] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xd32a80 name Existed_Raid, state configuring 00:13:55.303 11:24:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r raid0 -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:13:55.562 [2024-07-15 11:24:38.977904] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:55.562 [2024-07-15 11:24:38.977944] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:55.562 [2024-07-15 11:24:38.977954] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:55.562 [2024-07-15 11:24:38.977965] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:55.562 [2024-07-15 11:24:38.977974] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:55.562 [2024-07-15 11:24:38.977985] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:55.562 11:24:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:13:55.820 [2024-07-15 11:24:39.236513] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:55.820 BaseBdev1 00:13:55.820 11:24:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:13:55.820 11:24:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:13:55.820 11:24:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:13:55.820 11:24:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:13:55.820 11:24:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:13:55.820 11:24:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:13:55.820 11:24:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:13:56.080 11:24:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:56.338 [ 00:13:56.338 { 00:13:56.338 "name": "BaseBdev1", 00:13:56.338 "aliases": [ 00:13:56.338 "5e7413d9-8fd8-4b88-8647-1119da1d2597" 00:13:56.338 ], 00:13:56.338 "product_name": "Malloc disk", 00:13:56.338 "block_size": 512, 00:13:56.338 "num_blocks": 65536, 00:13:56.338 "uuid": "5e7413d9-8fd8-4b88-8647-1119da1d2597", 00:13:56.338 "assigned_rate_limits": { 00:13:56.338 "rw_ios_per_sec": 0, 00:13:56.338 "rw_mbytes_per_sec": 0, 00:13:56.338 "r_mbytes_per_sec": 0, 00:13:56.338 "w_mbytes_per_sec": 0 00:13:56.338 }, 00:13:56.338 "claimed": true, 00:13:56.338 "claim_type": "exclusive_write", 00:13:56.338 "zoned": false, 00:13:56.338 "supported_io_types": { 00:13:56.338 "read": true, 00:13:56.338 "write": true, 00:13:56.338 "unmap": true, 00:13:56.338 "flush": true, 00:13:56.338 "reset": true, 00:13:56.338 "nvme_admin": false, 00:13:56.338 "nvme_io": false, 00:13:56.338 "nvme_io_md": false, 00:13:56.338 "write_zeroes": true, 00:13:56.338 "zcopy": true, 00:13:56.338 "get_zone_info": false, 00:13:56.338 "zone_management": false, 00:13:56.338 "zone_append": false, 00:13:56.338 "compare": false, 00:13:56.338 "compare_and_write": false, 00:13:56.338 "abort": true, 00:13:56.338 "seek_hole": false, 00:13:56.338 "seek_data": false, 00:13:56.338 "copy": true, 00:13:56.338 "nvme_iov_md": false 00:13:56.338 }, 00:13:56.338 "memory_domains": [ 00:13:56.338 { 00:13:56.338 "dma_device_id": "system", 00:13:56.338 "dma_device_type": 1 00:13:56.338 }, 00:13:56.338 { 00:13:56.338 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:56.339 "dma_device_type": 2 00:13:56.339 } 00:13:56.339 ], 00:13:56.339 "driver_specific": {} 00:13:56.339 } 00:13:56.339 ] 00:13:56.339 11:24:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:13:56.339 11:24:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:13:56.339 11:24:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:13:56.339 11:24:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:13:56.339 11:24:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:13:56.339 11:24:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:13:56.339 11:24:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:13:56.339 11:24:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:56.339 11:24:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:56.339 11:24:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:56.339 11:24:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:56.339 11:24:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:56.339 11:24:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:56.598 11:24:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:56.598 "name": "Existed_Raid", 00:13:56.598 "uuid": "98354f09-971e-40f5-a84d-1681d74e8e1f", 00:13:56.598 "strip_size_kb": 64, 00:13:56.598 "state": "configuring", 00:13:56.598 "raid_level": "raid0", 00:13:56.598 "superblock": true, 00:13:56.598 "num_base_bdevs": 3, 00:13:56.598 "num_base_bdevs_discovered": 1, 00:13:56.598 "num_base_bdevs_operational": 3, 00:13:56.598 "base_bdevs_list": [ 00:13:56.598 { 00:13:56.598 "name": "BaseBdev1", 00:13:56.598 "uuid": "5e7413d9-8fd8-4b88-8647-1119da1d2597", 00:13:56.598 "is_configured": true, 00:13:56.598 "data_offset": 2048, 00:13:56.598 "data_size": 63488 00:13:56.598 }, 00:13:56.598 { 00:13:56.598 "name": "BaseBdev2", 00:13:56.598 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:56.598 "is_configured": false, 00:13:56.598 "data_offset": 0, 00:13:56.598 "data_size": 0 00:13:56.598 }, 00:13:56.598 { 00:13:56.598 "name": "BaseBdev3", 00:13:56.598 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:56.598 "is_configured": false, 00:13:56.598 "data_offset": 0, 00:13:56.598 "data_size": 0 00:13:56.598 } 00:13:56.598 ] 00:13:56.598 }' 00:13:56.598 11:24:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:56.598 11:24:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.211 11:24:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:13:57.211 [2024-07-15 11:24:40.800656] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:57.211 [2024-07-15 11:24:40.800694] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xd32310 name Existed_Raid, state configuring 00:13:57.468 11:24:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r raid0 -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:13:57.468 [2024-07-15 11:24:41.045339] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:57.468 [2024-07-15 11:24:41.046769] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:57.468 [2024-07-15 11:24:41.046800] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:57.468 [2024-07-15 11:24:41.046811] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:57.468 [2024-07-15 11:24:41.046823] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:57.725 11:24:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:13:57.725 11:24:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:13:57.725 11:24:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:13:57.725 11:24:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:13:57.725 11:24:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:13:57.725 11:24:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:13:57.725 11:24:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:13:57.725 11:24:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:13:57.725 11:24:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:57.725 11:24:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:57.725 11:24:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:57.725 11:24:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:57.725 11:24:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:57.725 11:24:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:57.725 11:24:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:57.725 "name": "Existed_Raid", 00:13:57.725 "uuid": "2a4e606d-9cac-4f9d-99eb-3e9f87fa2c11", 00:13:57.725 "strip_size_kb": 64, 00:13:57.725 "state": "configuring", 00:13:57.725 "raid_level": "raid0", 00:13:57.725 "superblock": true, 00:13:57.725 "num_base_bdevs": 3, 00:13:57.725 "num_base_bdevs_discovered": 1, 00:13:57.725 "num_base_bdevs_operational": 3, 00:13:57.725 "base_bdevs_list": [ 00:13:57.725 { 00:13:57.725 "name": "BaseBdev1", 00:13:57.725 "uuid": "5e7413d9-8fd8-4b88-8647-1119da1d2597", 00:13:57.725 "is_configured": true, 00:13:57.725 "data_offset": 2048, 00:13:57.725 "data_size": 63488 00:13:57.725 }, 00:13:57.725 { 00:13:57.725 "name": "BaseBdev2", 00:13:57.725 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:57.725 "is_configured": false, 00:13:57.725 "data_offset": 0, 00:13:57.725 "data_size": 0 00:13:57.725 }, 00:13:57.725 { 00:13:57.725 "name": "BaseBdev3", 00:13:57.725 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:57.725 "is_configured": false, 00:13:57.725 "data_offset": 0, 00:13:57.725 "data_size": 0 00:13:57.725 } 00:13:57.725 ] 00:13:57.725 }' 00:13:57.725 11:24:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:57.725 11:24:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:58.657 11:24:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:13:58.657 [2024-07-15 11:24:42.147710] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:58.657 BaseBdev2 00:13:58.657 11:24:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:13:58.657 11:24:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:13:58.657 11:24:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:13:58.657 11:24:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:13:58.657 11:24:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:13:58.657 11:24:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:13:58.657 11:24:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:13:58.915 11:24:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:59.173 [ 00:13:59.173 { 00:13:59.173 "name": "BaseBdev2", 00:13:59.173 "aliases": [ 00:13:59.173 "25bd68cb-11a8-4769-894d-e352e9d79f37" 00:13:59.173 ], 00:13:59.173 "product_name": "Malloc disk", 00:13:59.173 "block_size": 512, 00:13:59.173 "num_blocks": 65536, 00:13:59.173 "uuid": "25bd68cb-11a8-4769-894d-e352e9d79f37", 00:13:59.173 "assigned_rate_limits": { 00:13:59.173 "rw_ios_per_sec": 0, 00:13:59.173 "rw_mbytes_per_sec": 0, 00:13:59.173 "r_mbytes_per_sec": 0, 00:13:59.173 "w_mbytes_per_sec": 0 00:13:59.173 }, 00:13:59.173 "claimed": true, 00:13:59.173 "claim_type": "exclusive_write", 00:13:59.173 "zoned": false, 00:13:59.173 "supported_io_types": { 00:13:59.173 "read": true, 00:13:59.173 "write": true, 00:13:59.173 "unmap": true, 00:13:59.173 "flush": true, 00:13:59.173 "reset": true, 00:13:59.173 "nvme_admin": false, 00:13:59.173 "nvme_io": false, 00:13:59.173 "nvme_io_md": false, 00:13:59.173 "write_zeroes": true, 00:13:59.173 "zcopy": true, 00:13:59.173 "get_zone_info": false, 00:13:59.173 "zone_management": false, 00:13:59.173 "zone_append": false, 00:13:59.173 "compare": false, 00:13:59.173 "compare_and_write": false, 00:13:59.173 "abort": true, 00:13:59.173 "seek_hole": false, 00:13:59.173 "seek_data": false, 00:13:59.173 "copy": true, 00:13:59.173 "nvme_iov_md": false 00:13:59.173 }, 00:13:59.173 "memory_domains": [ 00:13:59.173 { 00:13:59.173 "dma_device_id": "system", 00:13:59.173 "dma_device_type": 1 00:13:59.173 }, 00:13:59.173 { 00:13:59.173 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:59.173 "dma_device_type": 2 00:13:59.173 } 00:13:59.173 ], 00:13:59.173 "driver_specific": {} 00:13:59.173 } 00:13:59.173 ] 00:13:59.173 11:24:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:13:59.173 11:24:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:13:59.173 11:24:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:13:59.173 11:24:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:13:59.173 11:24:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:13:59.173 11:24:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:13:59.173 11:24:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:13:59.173 11:24:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:13:59.173 11:24:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:13:59.173 11:24:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:13:59.173 11:24:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:13:59.173 11:24:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:13:59.173 11:24:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:13:59.173 11:24:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:13:59.173 11:24:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:59.430 11:24:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:13:59.430 "name": "Existed_Raid", 00:13:59.430 "uuid": "2a4e606d-9cac-4f9d-99eb-3e9f87fa2c11", 00:13:59.430 "strip_size_kb": 64, 00:13:59.430 "state": "configuring", 00:13:59.430 "raid_level": "raid0", 00:13:59.430 "superblock": true, 00:13:59.430 "num_base_bdevs": 3, 00:13:59.430 "num_base_bdevs_discovered": 2, 00:13:59.430 "num_base_bdevs_operational": 3, 00:13:59.430 "base_bdevs_list": [ 00:13:59.430 { 00:13:59.430 "name": "BaseBdev1", 00:13:59.430 "uuid": "5e7413d9-8fd8-4b88-8647-1119da1d2597", 00:13:59.430 "is_configured": true, 00:13:59.430 "data_offset": 2048, 00:13:59.430 "data_size": 63488 00:13:59.430 }, 00:13:59.430 { 00:13:59.430 "name": "BaseBdev2", 00:13:59.430 "uuid": "25bd68cb-11a8-4769-894d-e352e9d79f37", 00:13:59.430 "is_configured": true, 00:13:59.430 "data_offset": 2048, 00:13:59.430 "data_size": 63488 00:13:59.430 }, 00:13:59.430 { 00:13:59.430 "name": "BaseBdev3", 00:13:59.430 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:59.430 "is_configured": false, 00:13:59.430 "data_offset": 0, 00:13:59.430 "data_size": 0 00:13:59.430 } 00:13:59.430 ] 00:13:59.430 }' 00:13:59.430 11:24:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:13:59.430 11:24:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.994 11:24:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:14:00.251 [2024-07-15 11:24:43.735270] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:00.251 [2024-07-15 11:24:43.735429] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xd33400 00:14:00.251 [2024-07-15 11:24:43.735443] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:00.251 [2024-07-15 11:24:43.735614] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xd32ef0 00:14:00.251 [2024-07-15 11:24:43.735725] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xd33400 00:14:00.251 [2024-07-15 11:24:43.735735] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0xd33400 00:14:00.251 [2024-07-15 11:24:43.735824] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:00.251 BaseBdev3 00:14:00.251 11:24:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev3 00:14:00.251 11:24:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:14:00.251 11:24:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:14:00.251 11:24:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:14:00.251 11:24:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:14:00.251 11:24:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:14:00.251 11:24:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:14:00.508 11:24:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:00.764 [ 00:14:00.764 { 00:14:00.764 "name": "BaseBdev3", 00:14:00.764 "aliases": [ 00:14:00.764 "2f9b7058-43d9-4fa3-986b-5467472f2578" 00:14:00.764 ], 00:14:00.764 "product_name": "Malloc disk", 00:14:00.764 "block_size": 512, 00:14:00.764 "num_blocks": 65536, 00:14:00.764 "uuid": "2f9b7058-43d9-4fa3-986b-5467472f2578", 00:14:00.764 "assigned_rate_limits": { 00:14:00.764 "rw_ios_per_sec": 0, 00:14:00.764 "rw_mbytes_per_sec": 0, 00:14:00.764 "r_mbytes_per_sec": 0, 00:14:00.764 "w_mbytes_per_sec": 0 00:14:00.764 }, 00:14:00.764 "claimed": true, 00:14:00.764 "claim_type": "exclusive_write", 00:14:00.764 "zoned": false, 00:14:00.764 "supported_io_types": { 00:14:00.764 "read": true, 00:14:00.764 "write": true, 00:14:00.764 "unmap": true, 00:14:00.764 "flush": true, 00:14:00.764 "reset": true, 00:14:00.764 "nvme_admin": false, 00:14:00.764 "nvme_io": false, 00:14:00.764 "nvme_io_md": false, 00:14:00.764 "write_zeroes": true, 00:14:00.764 "zcopy": true, 00:14:00.764 "get_zone_info": false, 00:14:00.764 "zone_management": false, 00:14:00.764 "zone_append": false, 00:14:00.764 "compare": false, 00:14:00.764 "compare_and_write": false, 00:14:00.764 "abort": true, 00:14:00.764 "seek_hole": false, 00:14:00.764 "seek_data": false, 00:14:00.764 "copy": true, 00:14:00.764 "nvme_iov_md": false 00:14:00.764 }, 00:14:00.764 "memory_domains": [ 00:14:00.764 { 00:14:00.764 "dma_device_id": "system", 00:14:00.764 "dma_device_type": 1 00:14:00.764 }, 00:14:00.764 { 00:14:00.764 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:00.764 "dma_device_type": 2 00:14:00.764 } 00:14:00.764 ], 00:14:00.764 "driver_specific": {} 00:14:00.764 } 00:14:00.764 ] 00:14:00.764 11:24:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:14:00.764 11:24:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:14:00.765 11:24:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:14:00.765 11:24:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:14:00.765 11:24:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:14:00.765 11:24:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:14:00.765 11:24:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:14:00.765 11:24:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:00.765 11:24:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:00.765 11:24:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:00.765 11:24:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:00.765 11:24:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:00.765 11:24:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:00.765 11:24:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:00.765 11:24:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:01.021 11:24:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:01.021 "name": "Existed_Raid", 00:14:01.021 "uuid": "2a4e606d-9cac-4f9d-99eb-3e9f87fa2c11", 00:14:01.021 "strip_size_kb": 64, 00:14:01.021 "state": "online", 00:14:01.021 "raid_level": "raid0", 00:14:01.021 "superblock": true, 00:14:01.021 "num_base_bdevs": 3, 00:14:01.021 "num_base_bdevs_discovered": 3, 00:14:01.021 "num_base_bdevs_operational": 3, 00:14:01.021 "base_bdevs_list": [ 00:14:01.021 { 00:14:01.021 "name": "BaseBdev1", 00:14:01.021 "uuid": "5e7413d9-8fd8-4b88-8647-1119da1d2597", 00:14:01.021 "is_configured": true, 00:14:01.021 "data_offset": 2048, 00:14:01.021 "data_size": 63488 00:14:01.021 }, 00:14:01.021 { 00:14:01.021 "name": "BaseBdev2", 00:14:01.021 "uuid": "25bd68cb-11a8-4769-894d-e352e9d79f37", 00:14:01.021 "is_configured": true, 00:14:01.021 "data_offset": 2048, 00:14:01.021 "data_size": 63488 00:14:01.021 }, 00:14:01.021 { 00:14:01.021 "name": "BaseBdev3", 00:14:01.021 "uuid": "2f9b7058-43d9-4fa3-986b-5467472f2578", 00:14:01.021 "is_configured": true, 00:14:01.021 "data_offset": 2048, 00:14:01.021 "data_size": 63488 00:14:01.021 } 00:14:01.021 ] 00:14:01.021 }' 00:14:01.021 11:24:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:01.021 11:24:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.584 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:14:01.584 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:14:01.584 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:14:01.584 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:14:01.584 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:14:01.584 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # local name 00:14:01.584 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:14:01.584 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:14:01.842 [2024-07-15 11:24:45.227520] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:01.842 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:14:01.842 "name": "Existed_Raid", 00:14:01.842 "aliases": [ 00:14:01.842 "2a4e606d-9cac-4f9d-99eb-3e9f87fa2c11" 00:14:01.842 ], 00:14:01.842 "product_name": "Raid Volume", 00:14:01.842 "block_size": 512, 00:14:01.842 "num_blocks": 190464, 00:14:01.842 "uuid": "2a4e606d-9cac-4f9d-99eb-3e9f87fa2c11", 00:14:01.842 "assigned_rate_limits": { 00:14:01.842 "rw_ios_per_sec": 0, 00:14:01.842 "rw_mbytes_per_sec": 0, 00:14:01.842 "r_mbytes_per_sec": 0, 00:14:01.842 "w_mbytes_per_sec": 0 00:14:01.842 }, 00:14:01.842 "claimed": false, 00:14:01.842 "zoned": false, 00:14:01.842 "supported_io_types": { 00:14:01.842 "read": true, 00:14:01.842 "write": true, 00:14:01.842 "unmap": true, 00:14:01.842 "flush": true, 00:14:01.842 "reset": true, 00:14:01.842 "nvme_admin": false, 00:14:01.842 "nvme_io": false, 00:14:01.842 "nvme_io_md": false, 00:14:01.842 "write_zeroes": true, 00:14:01.842 "zcopy": false, 00:14:01.842 "get_zone_info": false, 00:14:01.842 "zone_management": false, 00:14:01.842 "zone_append": false, 00:14:01.842 "compare": false, 00:14:01.842 "compare_and_write": false, 00:14:01.842 "abort": false, 00:14:01.842 "seek_hole": false, 00:14:01.842 "seek_data": false, 00:14:01.842 "copy": false, 00:14:01.842 "nvme_iov_md": false 00:14:01.842 }, 00:14:01.842 "memory_domains": [ 00:14:01.842 { 00:14:01.842 "dma_device_id": "system", 00:14:01.842 "dma_device_type": 1 00:14:01.842 }, 00:14:01.842 { 00:14:01.842 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:01.842 "dma_device_type": 2 00:14:01.842 }, 00:14:01.842 { 00:14:01.842 "dma_device_id": "system", 00:14:01.842 "dma_device_type": 1 00:14:01.842 }, 00:14:01.842 { 00:14:01.842 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:01.842 "dma_device_type": 2 00:14:01.842 }, 00:14:01.842 { 00:14:01.842 "dma_device_id": "system", 00:14:01.842 "dma_device_type": 1 00:14:01.842 }, 00:14:01.842 { 00:14:01.842 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:01.842 "dma_device_type": 2 00:14:01.842 } 00:14:01.842 ], 00:14:01.842 "driver_specific": { 00:14:01.842 "raid": { 00:14:01.842 "uuid": "2a4e606d-9cac-4f9d-99eb-3e9f87fa2c11", 00:14:01.842 "strip_size_kb": 64, 00:14:01.842 "state": "online", 00:14:01.842 "raid_level": "raid0", 00:14:01.842 "superblock": true, 00:14:01.842 "num_base_bdevs": 3, 00:14:01.842 "num_base_bdevs_discovered": 3, 00:14:01.842 "num_base_bdevs_operational": 3, 00:14:01.842 "base_bdevs_list": [ 00:14:01.842 { 00:14:01.842 "name": "BaseBdev1", 00:14:01.842 "uuid": "5e7413d9-8fd8-4b88-8647-1119da1d2597", 00:14:01.842 "is_configured": true, 00:14:01.842 "data_offset": 2048, 00:14:01.842 "data_size": 63488 00:14:01.842 }, 00:14:01.842 { 00:14:01.842 "name": "BaseBdev2", 00:14:01.842 "uuid": "25bd68cb-11a8-4769-894d-e352e9d79f37", 00:14:01.842 "is_configured": true, 00:14:01.842 "data_offset": 2048, 00:14:01.842 "data_size": 63488 00:14:01.842 }, 00:14:01.842 { 00:14:01.842 "name": "BaseBdev3", 00:14:01.842 "uuid": "2f9b7058-43d9-4fa3-986b-5467472f2578", 00:14:01.842 "is_configured": true, 00:14:01.842 "data_offset": 2048, 00:14:01.842 "data_size": 63488 00:14:01.842 } 00:14:01.842 ] 00:14:01.842 } 00:14:01.842 } 00:14:01.842 }' 00:14:01.842 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:01.842 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:14:01.842 BaseBdev2 00:14:01.842 BaseBdev3' 00:14:01.842 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:14:01.842 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:14:01.842 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:14:02.100 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:14:02.100 "name": "BaseBdev1", 00:14:02.100 "aliases": [ 00:14:02.100 "5e7413d9-8fd8-4b88-8647-1119da1d2597" 00:14:02.100 ], 00:14:02.100 "product_name": "Malloc disk", 00:14:02.100 "block_size": 512, 00:14:02.100 "num_blocks": 65536, 00:14:02.100 "uuid": "5e7413d9-8fd8-4b88-8647-1119da1d2597", 00:14:02.100 "assigned_rate_limits": { 00:14:02.100 "rw_ios_per_sec": 0, 00:14:02.100 "rw_mbytes_per_sec": 0, 00:14:02.100 "r_mbytes_per_sec": 0, 00:14:02.100 "w_mbytes_per_sec": 0 00:14:02.100 }, 00:14:02.100 "claimed": true, 00:14:02.100 "claim_type": "exclusive_write", 00:14:02.100 "zoned": false, 00:14:02.100 "supported_io_types": { 00:14:02.100 "read": true, 00:14:02.100 "write": true, 00:14:02.100 "unmap": true, 00:14:02.100 "flush": true, 00:14:02.100 "reset": true, 00:14:02.100 "nvme_admin": false, 00:14:02.100 "nvme_io": false, 00:14:02.100 "nvme_io_md": false, 00:14:02.100 "write_zeroes": true, 00:14:02.100 "zcopy": true, 00:14:02.100 "get_zone_info": false, 00:14:02.100 "zone_management": false, 00:14:02.100 "zone_append": false, 00:14:02.100 "compare": false, 00:14:02.100 "compare_and_write": false, 00:14:02.100 "abort": true, 00:14:02.100 "seek_hole": false, 00:14:02.100 "seek_data": false, 00:14:02.100 "copy": true, 00:14:02.100 "nvme_iov_md": false 00:14:02.100 }, 00:14:02.100 "memory_domains": [ 00:14:02.100 { 00:14:02.100 "dma_device_id": "system", 00:14:02.100 "dma_device_type": 1 00:14:02.100 }, 00:14:02.100 { 00:14:02.100 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:02.100 "dma_device_type": 2 00:14:02.100 } 00:14:02.100 ], 00:14:02.100 "driver_specific": {} 00:14:02.100 }' 00:14:02.100 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:02.100 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:02.100 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:14:02.100 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:02.100 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:02.358 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:14:02.358 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:02.358 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:02.358 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:14:02.358 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:02.358 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:02.358 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:14:02.358 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:14:02.358 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:14:02.358 11:24:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:14:02.616 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:14:02.616 "name": "BaseBdev2", 00:14:02.616 "aliases": [ 00:14:02.616 "25bd68cb-11a8-4769-894d-e352e9d79f37" 00:14:02.616 ], 00:14:02.616 "product_name": "Malloc disk", 00:14:02.616 "block_size": 512, 00:14:02.616 "num_blocks": 65536, 00:14:02.616 "uuid": "25bd68cb-11a8-4769-894d-e352e9d79f37", 00:14:02.616 "assigned_rate_limits": { 00:14:02.616 "rw_ios_per_sec": 0, 00:14:02.616 "rw_mbytes_per_sec": 0, 00:14:02.616 "r_mbytes_per_sec": 0, 00:14:02.616 "w_mbytes_per_sec": 0 00:14:02.616 }, 00:14:02.616 "claimed": true, 00:14:02.616 "claim_type": "exclusive_write", 00:14:02.616 "zoned": false, 00:14:02.616 "supported_io_types": { 00:14:02.616 "read": true, 00:14:02.616 "write": true, 00:14:02.616 "unmap": true, 00:14:02.616 "flush": true, 00:14:02.616 "reset": true, 00:14:02.616 "nvme_admin": false, 00:14:02.616 "nvme_io": false, 00:14:02.616 "nvme_io_md": false, 00:14:02.616 "write_zeroes": true, 00:14:02.616 "zcopy": true, 00:14:02.616 "get_zone_info": false, 00:14:02.616 "zone_management": false, 00:14:02.616 "zone_append": false, 00:14:02.616 "compare": false, 00:14:02.616 "compare_and_write": false, 00:14:02.616 "abort": true, 00:14:02.616 "seek_hole": false, 00:14:02.616 "seek_data": false, 00:14:02.616 "copy": true, 00:14:02.616 "nvme_iov_md": false 00:14:02.616 }, 00:14:02.616 "memory_domains": [ 00:14:02.616 { 00:14:02.616 "dma_device_id": "system", 00:14:02.616 "dma_device_type": 1 00:14:02.616 }, 00:14:02.616 { 00:14:02.616 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:02.616 "dma_device_type": 2 00:14:02.616 } 00:14:02.616 ], 00:14:02.616 "driver_specific": {} 00:14:02.616 }' 00:14:02.617 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:02.617 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:02.617 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:14:02.617 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:02.874 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:02.874 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:14:02.874 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:02.874 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:02.874 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:14:02.874 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:02.874 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:02.874 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:14:02.874 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:14:02.874 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:14:02.874 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:14:03.132 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:14:03.132 "name": "BaseBdev3", 00:14:03.132 "aliases": [ 00:14:03.132 "2f9b7058-43d9-4fa3-986b-5467472f2578" 00:14:03.132 ], 00:14:03.132 "product_name": "Malloc disk", 00:14:03.132 "block_size": 512, 00:14:03.132 "num_blocks": 65536, 00:14:03.132 "uuid": "2f9b7058-43d9-4fa3-986b-5467472f2578", 00:14:03.132 "assigned_rate_limits": { 00:14:03.132 "rw_ios_per_sec": 0, 00:14:03.132 "rw_mbytes_per_sec": 0, 00:14:03.132 "r_mbytes_per_sec": 0, 00:14:03.132 "w_mbytes_per_sec": 0 00:14:03.132 }, 00:14:03.132 "claimed": true, 00:14:03.132 "claim_type": "exclusive_write", 00:14:03.132 "zoned": false, 00:14:03.132 "supported_io_types": { 00:14:03.132 "read": true, 00:14:03.132 "write": true, 00:14:03.132 "unmap": true, 00:14:03.132 "flush": true, 00:14:03.132 "reset": true, 00:14:03.132 "nvme_admin": false, 00:14:03.132 "nvme_io": false, 00:14:03.132 "nvme_io_md": false, 00:14:03.132 "write_zeroes": true, 00:14:03.132 "zcopy": true, 00:14:03.132 "get_zone_info": false, 00:14:03.132 "zone_management": false, 00:14:03.132 "zone_append": false, 00:14:03.132 "compare": false, 00:14:03.132 "compare_and_write": false, 00:14:03.132 "abort": true, 00:14:03.132 "seek_hole": false, 00:14:03.132 "seek_data": false, 00:14:03.132 "copy": true, 00:14:03.132 "nvme_iov_md": false 00:14:03.132 }, 00:14:03.132 "memory_domains": [ 00:14:03.132 { 00:14:03.132 "dma_device_id": "system", 00:14:03.132 "dma_device_type": 1 00:14:03.132 }, 00:14:03.132 { 00:14:03.132 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:03.132 "dma_device_type": 2 00:14:03.132 } 00:14:03.132 ], 00:14:03.132 "driver_specific": {} 00:14:03.132 }' 00:14:03.132 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:03.132 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:03.132 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:14:03.132 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:03.390 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:03.390 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:14:03.390 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:03.390 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:03.390 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:14:03.390 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:03.390 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:03.390 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:14:03.390 11:24:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:14:03.649 [2024-07-15 11:24:47.180444] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:03.649 [2024-07-15 11:24:47.180470] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:03.649 [2024-07-15 11:24:47.180511] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:03.649 11:24:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@275 -- # local expected_state 00:14:03.649 11:24:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # has_redundancy raid0 00:14:03.649 11:24:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # case $1 in 00:14:03.649 11:24:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # return 1 00:14:03.649 11:24:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@277 -- # expected_state=offline 00:14:03.649 11:24:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:14:03.649 11:24:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:14:03.649 11:24:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=offline 00:14:03.649 11:24:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:14:03.649 11:24:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:03.649 11:24:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:14:03.649 11:24:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:03.649 11:24:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:03.649 11:24:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:03.649 11:24:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:03.649 11:24:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:03.649 11:24:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:03.908 11:24:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:03.908 "name": "Existed_Raid", 00:14:03.908 "uuid": "2a4e606d-9cac-4f9d-99eb-3e9f87fa2c11", 00:14:03.908 "strip_size_kb": 64, 00:14:03.908 "state": "offline", 00:14:03.908 "raid_level": "raid0", 00:14:03.908 "superblock": true, 00:14:03.908 "num_base_bdevs": 3, 00:14:03.908 "num_base_bdevs_discovered": 2, 00:14:03.908 "num_base_bdevs_operational": 2, 00:14:03.908 "base_bdevs_list": [ 00:14:03.908 { 00:14:03.908 "name": null, 00:14:03.908 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:03.908 "is_configured": false, 00:14:03.908 "data_offset": 2048, 00:14:03.908 "data_size": 63488 00:14:03.908 }, 00:14:03.908 { 00:14:03.908 "name": "BaseBdev2", 00:14:03.908 "uuid": "25bd68cb-11a8-4769-894d-e352e9d79f37", 00:14:03.908 "is_configured": true, 00:14:03.908 "data_offset": 2048, 00:14:03.908 "data_size": 63488 00:14:03.908 }, 00:14:03.908 { 00:14:03.908 "name": "BaseBdev3", 00:14:03.908 "uuid": "2f9b7058-43d9-4fa3-986b-5467472f2578", 00:14:03.908 "is_configured": true, 00:14:03.908 "data_offset": 2048, 00:14:03.908 "data_size": 63488 00:14:03.908 } 00:14:03.908 ] 00:14:03.908 }' 00:14:03.908 11:24:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:03.908 11:24:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.475 11:24:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:14:04.475 11:24:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:14:04.475 11:24:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:14:04.475 11:24:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:04.734 11:24:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:14:04.734 11:24:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:04.734 11:24:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:14:04.992 [2024-07-15 11:24:48.497618] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:04.992 11:24:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:14:04.992 11:24:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:14:04.992 11:24:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:04.992 11:24:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:14:05.249 11:24:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:14:05.250 11:24:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:05.250 11:24:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev3 00:14:05.508 [2024-07-15 11:24:49.003374] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:05.508 [2024-07-15 11:24:49.003423] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xd33400 name Existed_Raid, state offline 00:14:05.508 11:24:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:14:05.508 11:24:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:14:05.508 11:24:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:05.508 11:24:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:14:05.766 11:24:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:14:05.766 11:24:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:14:05.766 11:24:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # '[' 3 -gt 2 ']' 00:14:05.766 11:24:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i = 1 )) 00:14:05.766 11:24:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:14:05.766 11:24:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:14:06.024 BaseBdev2 00:14:06.024 11:24:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev2 00:14:06.024 11:24:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:14:06.024 11:24:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:14:06.024 11:24:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:14:06.024 11:24:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:14:06.024 11:24:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:14:06.024 11:24:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:14:06.283 11:24:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:06.541 [ 00:14:06.541 { 00:14:06.541 "name": "BaseBdev2", 00:14:06.541 "aliases": [ 00:14:06.541 "cbf7a58c-1289-4db7-9c8c-02c58647961e" 00:14:06.541 ], 00:14:06.541 "product_name": "Malloc disk", 00:14:06.541 "block_size": 512, 00:14:06.541 "num_blocks": 65536, 00:14:06.541 "uuid": "cbf7a58c-1289-4db7-9c8c-02c58647961e", 00:14:06.541 "assigned_rate_limits": { 00:14:06.541 "rw_ios_per_sec": 0, 00:14:06.541 "rw_mbytes_per_sec": 0, 00:14:06.541 "r_mbytes_per_sec": 0, 00:14:06.541 "w_mbytes_per_sec": 0 00:14:06.541 }, 00:14:06.541 "claimed": false, 00:14:06.541 "zoned": false, 00:14:06.541 "supported_io_types": { 00:14:06.541 "read": true, 00:14:06.541 "write": true, 00:14:06.541 "unmap": true, 00:14:06.541 "flush": true, 00:14:06.541 "reset": true, 00:14:06.541 "nvme_admin": false, 00:14:06.541 "nvme_io": false, 00:14:06.541 "nvme_io_md": false, 00:14:06.541 "write_zeroes": true, 00:14:06.541 "zcopy": true, 00:14:06.541 "get_zone_info": false, 00:14:06.541 "zone_management": false, 00:14:06.541 "zone_append": false, 00:14:06.541 "compare": false, 00:14:06.541 "compare_and_write": false, 00:14:06.541 "abort": true, 00:14:06.541 "seek_hole": false, 00:14:06.541 "seek_data": false, 00:14:06.541 "copy": true, 00:14:06.541 "nvme_iov_md": false 00:14:06.541 }, 00:14:06.541 "memory_domains": [ 00:14:06.541 { 00:14:06.541 "dma_device_id": "system", 00:14:06.541 "dma_device_type": 1 00:14:06.541 }, 00:14:06.541 { 00:14:06.541 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:06.541 "dma_device_type": 2 00:14:06.541 } 00:14:06.541 ], 00:14:06.541 "driver_specific": {} 00:14:06.541 } 00:14:06.541 ] 00:14:06.541 11:24:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:14:06.541 11:24:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:14:06.541 11:24:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:14:06.541 11:24:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:14:06.800 BaseBdev3 00:14:06.800 11:24:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev3 00:14:06.800 11:24:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:14:06.800 11:24:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:14:06.800 11:24:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:14:06.800 11:24:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:14:06.800 11:24:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:14:06.800 11:24:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:14:07.058 11:24:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:07.316 [ 00:14:07.316 { 00:14:07.316 "name": "BaseBdev3", 00:14:07.316 "aliases": [ 00:14:07.316 "16f39641-0697-473a-8035-ff1b1c52b8a0" 00:14:07.316 ], 00:14:07.316 "product_name": "Malloc disk", 00:14:07.316 "block_size": 512, 00:14:07.316 "num_blocks": 65536, 00:14:07.316 "uuid": "16f39641-0697-473a-8035-ff1b1c52b8a0", 00:14:07.316 "assigned_rate_limits": { 00:14:07.316 "rw_ios_per_sec": 0, 00:14:07.316 "rw_mbytes_per_sec": 0, 00:14:07.316 "r_mbytes_per_sec": 0, 00:14:07.316 "w_mbytes_per_sec": 0 00:14:07.316 }, 00:14:07.316 "claimed": false, 00:14:07.316 "zoned": false, 00:14:07.316 "supported_io_types": { 00:14:07.316 "read": true, 00:14:07.316 "write": true, 00:14:07.316 "unmap": true, 00:14:07.316 "flush": true, 00:14:07.316 "reset": true, 00:14:07.316 "nvme_admin": false, 00:14:07.316 "nvme_io": false, 00:14:07.316 "nvme_io_md": false, 00:14:07.316 "write_zeroes": true, 00:14:07.316 "zcopy": true, 00:14:07.316 "get_zone_info": false, 00:14:07.316 "zone_management": false, 00:14:07.316 "zone_append": false, 00:14:07.316 "compare": false, 00:14:07.316 "compare_and_write": false, 00:14:07.316 "abort": true, 00:14:07.316 "seek_hole": false, 00:14:07.316 "seek_data": false, 00:14:07.316 "copy": true, 00:14:07.316 "nvme_iov_md": false 00:14:07.316 }, 00:14:07.316 "memory_domains": [ 00:14:07.316 { 00:14:07.316 "dma_device_id": "system", 00:14:07.316 "dma_device_type": 1 00:14:07.316 }, 00:14:07.316 { 00:14:07.316 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:07.316 "dma_device_type": 2 00:14:07.316 } 00:14:07.316 ], 00:14:07.316 "driver_specific": {} 00:14:07.316 } 00:14:07.316 ] 00:14:07.316 11:24:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:14:07.316 11:24:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:14:07.316 11:24:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:14:07.316 11:24:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@305 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r raid0 -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:14:07.574 [2024-07-15 11:24:50.995151] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:07.574 [2024-07-15 11:24:50.995190] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:07.574 [2024-07-15 11:24:50.995209] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:07.574 [2024-07-15 11:24:50.996558] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:07.574 11:24:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:14:07.574 11:24:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:14:07.574 11:24:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:14:07.574 11:24:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:14:07.574 11:24:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:07.574 11:24:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:07.574 11:24:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:07.574 11:24:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:07.574 11:24:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:07.574 11:24:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:07.574 11:24:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:07.574 11:24:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:07.832 11:24:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:07.832 "name": "Existed_Raid", 00:14:07.832 "uuid": "36187ade-b189-4a8e-aba2-c5c774536705", 00:14:07.832 "strip_size_kb": 64, 00:14:07.832 "state": "configuring", 00:14:07.832 "raid_level": "raid0", 00:14:07.832 "superblock": true, 00:14:07.832 "num_base_bdevs": 3, 00:14:07.832 "num_base_bdevs_discovered": 2, 00:14:07.832 "num_base_bdevs_operational": 3, 00:14:07.832 "base_bdevs_list": [ 00:14:07.833 { 00:14:07.833 "name": "BaseBdev1", 00:14:07.833 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:07.833 "is_configured": false, 00:14:07.833 "data_offset": 0, 00:14:07.833 "data_size": 0 00:14:07.833 }, 00:14:07.833 { 00:14:07.833 "name": "BaseBdev2", 00:14:07.833 "uuid": "cbf7a58c-1289-4db7-9c8c-02c58647961e", 00:14:07.833 "is_configured": true, 00:14:07.833 "data_offset": 2048, 00:14:07.833 "data_size": 63488 00:14:07.833 }, 00:14:07.833 { 00:14:07.833 "name": "BaseBdev3", 00:14:07.833 "uuid": "16f39641-0697-473a-8035-ff1b1c52b8a0", 00:14:07.833 "is_configured": true, 00:14:07.833 "data_offset": 2048, 00:14:07.833 "data_size": 63488 00:14:07.833 } 00:14:07.833 ] 00:14:07.833 }' 00:14:07.833 11:24:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:07.833 11:24:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:08.398 11:24:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev2 00:14:08.657 [2024-07-15 11:24:52.041885] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:08.657 11:24:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@309 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:14:08.657 11:24:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:14:08.657 11:24:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:14:08.657 11:24:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:14:08.657 11:24:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:08.657 11:24:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:08.657 11:24:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:08.657 11:24:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:08.657 11:24:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:08.657 11:24:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:08.657 11:24:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:08.657 11:24:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:08.915 11:24:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:08.915 "name": "Existed_Raid", 00:14:08.915 "uuid": "36187ade-b189-4a8e-aba2-c5c774536705", 00:14:08.915 "strip_size_kb": 64, 00:14:08.915 "state": "configuring", 00:14:08.915 "raid_level": "raid0", 00:14:08.915 "superblock": true, 00:14:08.915 "num_base_bdevs": 3, 00:14:08.915 "num_base_bdevs_discovered": 1, 00:14:08.915 "num_base_bdevs_operational": 3, 00:14:08.915 "base_bdevs_list": [ 00:14:08.915 { 00:14:08.915 "name": "BaseBdev1", 00:14:08.915 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:08.915 "is_configured": false, 00:14:08.915 "data_offset": 0, 00:14:08.915 "data_size": 0 00:14:08.915 }, 00:14:08.915 { 00:14:08.915 "name": null, 00:14:08.915 "uuid": "cbf7a58c-1289-4db7-9c8c-02c58647961e", 00:14:08.915 "is_configured": false, 00:14:08.915 "data_offset": 2048, 00:14:08.915 "data_size": 63488 00:14:08.915 }, 00:14:08.915 { 00:14:08.915 "name": "BaseBdev3", 00:14:08.915 "uuid": "16f39641-0697-473a-8035-ff1b1c52b8a0", 00:14:08.915 "is_configured": true, 00:14:08.915 "data_offset": 2048, 00:14:08.915 "data_size": 63488 00:14:08.915 } 00:14:08.915 ] 00:14:08.915 }' 00:14:08.915 11:24:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:08.915 11:24:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:09.482 11:24:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:09.482 11:24:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:09.748 11:24:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # [[ false == \f\a\l\s\e ]] 00:14:09.748 11:24:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:14:10.315 [2024-07-15 11:24:53.658528] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:10.315 BaseBdev1 00:14:10.315 11:24:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@313 -- # waitforbdev BaseBdev1 00:14:10.315 11:24:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:14:10.315 11:24:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:14:10.315 11:24:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:14:10.315 11:24:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:14:10.315 11:24:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:14:10.315 11:24:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:14:10.573 11:24:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:10.573 [ 00:14:10.573 { 00:14:10.573 "name": "BaseBdev1", 00:14:10.573 "aliases": [ 00:14:10.573 "16770936-d48d-439b-a3f5-95f22c40a73c" 00:14:10.573 ], 00:14:10.573 "product_name": "Malloc disk", 00:14:10.573 "block_size": 512, 00:14:10.573 "num_blocks": 65536, 00:14:10.573 "uuid": "16770936-d48d-439b-a3f5-95f22c40a73c", 00:14:10.573 "assigned_rate_limits": { 00:14:10.573 "rw_ios_per_sec": 0, 00:14:10.573 "rw_mbytes_per_sec": 0, 00:14:10.573 "r_mbytes_per_sec": 0, 00:14:10.573 "w_mbytes_per_sec": 0 00:14:10.573 }, 00:14:10.573 "claimed": true, 00:14:10.573 "claim_type": "exclusive_write", 00:14:10.573 "zoned": false, 00:14:10.573 "supported_io_types": { 00:14:10.573 "read": true, 00:14:10.574 "write": true, 00:14:10.574 "unmap": true, 00:14:10.574 "flush": true, 00:14:10.574 "reset": true, 00:14:10.574 "nvme_admin": false, 00:14:10.574 "nvme_io": false, 00:14:10.574 "nvme_io_md": false, 00:14:10.574 "write_zeroes": true, 00:14:10.574 "zcopy": true, 00:14:10.574 "get_zone_info": false, 00:14:10.574 "zone_management": false, 00:14:10.574 "zone_append": false, 00:14:10.574 "compare": false, 00:14:10.574 "compare_and_write": false, 00:14:10.574 "abort": true, 00:14:10.574 "seek_hole": false, 00:14:10.574 "seek_data": false, 00:14:10.574 "copy": true, 00:14:10.574 "nvme_iov_md": false 00:14:10.574 }, 00:14:10.574 "memory_domains": [ 00:14:10.574 { 00:14:10.574 "dma_device_id": "system", 00:14:10.574 "dma_device_type": 1 00:14:10.574 }, 00:14:10.574 { 00:14:10.574 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:10.574 "dma_device_type": 2 00:14:10.574 } 00:14:10.574 ], 00:14:10.574 "driver_specific": {} 00:14:10.574 } 00:14:10.574 ] 00:14:10.574 11:24:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:14:10.574 11:24:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:14:10.574 11:24:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:14:10.574 11:24:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:14:10.574 11:24:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:14:10.574 11:24:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:10.574 11:24:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:10.574 11:24:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:10.574 11:24:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:10.574 11:24:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:10.574 11:24:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:10.574 11:24:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:10.574 11:24:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:10.833 11:24:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:10.833 "name": "Existed_Raid", 00:14:10.833 "uuid": "36187ade-b189-4a8e-aba2-c5c774536705", 00:14:10.833 "strip_size_kb": 64, 00:14:10.833 "state": "configuring", 00:14:10.833 "raid_level": "raid0", 00:14:10.833 "superblock": true, 00:14:10.833 "num_base_bdevs": 3, 00:14:10.833 "num_base_bdevs_discovered": 2, 00:14:10.833 "num_base_bdevs_operational": 3, 00:14:10.833 "base_bdevs_list": [ 00:14:10.833 { 00:14:10.833 "name": "BaseBdev1", 00:14:10.833 "uuid": "16770936-d48d-439b-a3f5-95f22c40a73c", 00:14:10.833 "is_configured": true, 00:14:10.833 "data_offset": 2048, 00:14:10.833 "data_size": 63488 00:14:10.833 }, 00:14:10.833 { 00:14:10.833 "name": null, 00:14:10.833 "uuid": "cbf7a58c-1289-4db7-9c8c-02c58647961e", 00:14:10.833 "is_configured": false, 00:14:10.833 "data_offset": 2048, 00:14:10.833 "data_size": 63488 00:14:10.833 }, 00:14:10.833 { 00:14:10.833 "name": "BaseBdev3", 00:14:10.833 "uuid": "16f39641-0697-473a-8035-ff1b1c52b8a0", 00:14:10.833 "is_configured": true, 00:14:10.833 "data_offset": 2048, 00:14:10.833 "data_size": 63488 00:14:10.833 } 00:14:10.833 ] 00:14:10.833 }' 00:14:10.833 11:24:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:10.833 11:24:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:11.769 11:24:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:11.769 11:24:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:11.769 11:24:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # [[ true == \t\r\u\e ]] 00:14:11.769 11:24:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@317 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev3 00:14:12.027 [2024-07-15 11:24:55.471389] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:12.027 11:24:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:14:12.027 11:24:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:14:12.027 11:24:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:14:12.027 11:24:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:14:12.027 11:24:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:12.027 11:24:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:12.027 11:24:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:12.027 11:24:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:12.027 11:24:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:12.027 11:24:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:12.027 11:24:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:12.027 11:24:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:12.285 11:24:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:12.285 "name": "Existed_Raid", 00:14:12.285 "uuid": "36187ade-b189-4a8e-aba2-c5c774536705", 00:14:12.285 "strip_size_kb": 64, 00:14:12.285 "state": "configuring", 00:14:12.285 "raid_level": "raid0", 00:14:12.285 "superblock": true, 00:14:12.285 "num_base_bdevs": 3, 00:14:12.285 "num_base_bdevs_discovered": 1, 00:14:12.285 "num_base_bdevs_operational": 3, 00:14:12.285 "base_bdevs_list": [ 00:14:12.285 { 00:14:12.285 "name": "BaseBdev1", 00:14:12.285 "uuid": "16770936-d48d-439b-a3f5-95f22c40a73c", 00:14:12.285 "is_configured": true, 00:14:12.285 "data_offset": 2048, 00:14:12.285 "data_size": 63488 00:14:12.285 }, 00:14:12.285 { 00:14:12.285 "name": null, 00:14:12.285 "uuid": "cbf7a58c-1289-4db7-9c8c-02c58647961e", 00:14:12.285 "is_configured": false, 00:14:12.285 "data_offset": 2048, 00:14:12.285 "data_size": 63488 00:14:12.285 }, 00:14:12.285 { 00:14:12.285 "name": null, 00:14:12.285 "uuid": "16f39641-0697-473a-8035-ff1b1c52b8a0", 00:14:12.285 "is_configured": false, 00:14:12.285 "data_offset": 2048, 00:14:12.285 "data_size": 63488 00:14:12.285 } 00:14:12.285 ] 00:14:12.285 }' 00:14:12.285 11:24:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:12.285 11:24:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.850 11:24:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:12.850 11:24:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:13.108 11:24:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # [[ false == \f\a\l\s\e ]] 00:14:13.108 11:24:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:14:13.108 [2024-07-15 11:24:56.666586] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:13.108 11:24:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@322 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:14:13.108 11:24:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:14:13.108 11:24:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:14:13.108 11:24:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:14:13.108 11:24:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:13.108 11:24:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:13.108 11:24:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:13.108 11:24:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:13.108 11:24:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:13.108 11:24:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:13.108 11:24:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:13.108 11:24:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:13.366 11:24:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:13.366 "name": "Existed_Raid", 00:14:13.367 "uuid": "36187ade-b189-4a8e-aba2-c5c774536705", 00:14:13.367 "strip_size_kb": 64, 00:14:13.367 "state": "configuring", 00:14:13.367 "raid_level": "raid0", 00:14:13.367 "superblock": true, 00:14:13.367 "num_base_bdevs": 3, 00:14:13.367 "num_base_bdevs_discovered": 2, 00:14:13.367 "num_base_bdevs_operational": 3, 00:14:13.367 "base_bdevs_list": [ 00:14:13.367 { 00:14:13.367 "name": "BaseBdev1", 00:14:13.367 "uuid": "16770936-d48d-439b-a3f5-95f22c40a73c", 00:14:13.367 "is_configured": true, 00:14:13.367 "data_offset": 2048, 00:14:13.367 "data_size": 63488 00:14:13.367 }, 00:14:13.367 { 00:14:13.367 "name": null, 00:14:13.367 "uuid": "cbf7a58c-1289-4db7-9c8c-02c58647961e", 00:14:13.367 "is_configured": false, 00:14:13.367 "data_offset": 2048, 00:14:13.367 "data_size": 63488 00:14:13.367 }, 00:14:13.367 { 00:14:13.367 "name": "BaseBdev3", 00:14:13.367 "uuid": "16f39641-0697-473a-8035-ff1b1c52b8a0", 00:14:13.367 "is_configured": true, 00:14:13.367 "data_offset": 2048, 00:14:13.367 "data_size": 63488 00:14:13.367 } 00:14:13.367 ] 00:14:13.367 }' 00:14:13.367 11:24:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:13.367 11:24:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.967 11:24:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:13.967 11:24:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:14.224 11:24:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # [[ true == \t\r\u\e ]] 00:14:14.224 11:24:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@325 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:14:14.482 [2024-07-15 11:24:57.857771] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:14.482 11:24:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:14:14.482 11:24:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:14:14.482 11:24:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:14:14.482 11:24:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:14:14.482 11:24:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:14.482 11:24:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:14.482 11:24:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:14.482 11:24:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:14.482 11:24:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:14.482 11:24:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:14.482 11:24:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:14.482 11:24:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:14.482 11:24:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:14.482 "name": "Existed_Raid", 00:14:14.482 "uuid": "36187ade-b189-4a8e-aba2-c5c774536705", 00:14:14.482 "strip_size_kb": 64, 00:14:14.482 "state": "configuring", 00:14:14.482 "raid_level": "raid0", 00:14:14.482 "superblock": true, 00:14:14.482 "num_base_bdevs": 3, 00:14:14.482 "num_base_bdevs_discovered": 1, 00:14:14.482 "num_base_bdevs_operational": 3, 00:14:14.482 "base_bdevs_list": [ 00:14:14.482 { 00:14:14.482 "name": null, 00:14:14.482 "uuid": "16770936-d48d-439b-a3f5-95f22c40a73c", 00:14:14.482 "is_configured": false, 00:14:14.482 "data_offset": 2048, 00:14:14.482 "data_size": 63488 00:14:14.482 }, 00:14:14.482 { 00:14:14.482 "name": null, 00:14:14.482 "uuid": "cbf7a58c-1289-4db7-9c8c-02c58647961e", 00:14:14.482 "is_configured": false, 00:14:14.482 "data_offset": 2048, 00:14:14.482 "data_size": 63488 00:14:14.482 }, 00:14:14.482 { 00:14:14.482 "name": "BaseBdev3", 00:14:14.482 "uuid": "16f39641-0697-473a-8035-ff1b1c52b8a0", 00:14:14.482 "is_configured": true, 00:14:14.482 "data_offset": 2048, 00:14:14.482 "data_size": 63488 00:14:14.482 } 00:14:14.482 ] 00:14:14.482 }' 00:14:14.482 11:24:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:14.482 11:24:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:15.069 11:24:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@327 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:15.069 11:24:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@327 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:15.327 11:24:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@327 -- # [[ false == \f\a\l\s\e ]] 00:14:15.327 11:24:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@329 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:14:15.585 [2024-07-15 11:24:59.055534] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:15.585 11:24:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@330 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:14:15.585 11:24:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:14:15.585 11:24:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:14:15.585 11:24:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:14:15.585 11:24:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:15.585 11:24:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:15.585 11:24:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:15.585 11:24:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:15.585 11:24:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:15.585 11:24:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:15.585 11:24:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:15.585 11:24:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:15.843 11:24:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:15.843 "name": "Existed_Raid", 00:14:15.843 "uuid": "36187ade-b189-4a8e-aba2-c5c774536705", 00:14:15.843 "strip_size_kb": 64, 00:14:15.843 "state": "configuring", 00:14:15.843 "raid_level": "raid0", 00:14:15.843 "superblock": true, 00:14:15.843 "num_base_bdevs": 3, 00:14:15.843 "num_base_bdevs_discovered": 2, 00:14:15.843 "num_base_bdevs_operational": 3, 00:14:15.843 "base_bdevs_list": [ 00:14:15.843 { 00:14:15.843 "name": null, 00:14:15.843 "uuid": "16770936-d48d-439b-a3f5-95f22c40a73c", 00:14:15.843 "is_configured": false, 00:14:15.843 "data_offset": 2048, 00:14:15.843 "data_size": 63488 00:14:15.843 }, 00:14:15.843 { 00:14:15.843 "name": "BaseBdev2", 00:14:15.843 "uuid": "cbf7a58c-1289-4db7-9c8c-02c58647961e", 00:14:15.843 "is_configured": true, 00:14:15.843 "data_offset": 2048, 00:14:15.843 "data_size": 63488 00:14:15.843 }, 00:14:15.843 { 00:14:15.843 "name": "BaseBdev3", 00:14:15.843 "uuid": "16f39641-0697-473a-8035-ff1b1c52b8a0", 00:14:15.843 "is_configured": true, 00:14:15.843 "data_offset": 2048, 00:14:15.843 "data_size": 63488 00:14:15.843 } 00:14:15.843 ] 00:14:15.843 }' 00:14:15.843 11:24:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:15.843 11:24:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:16.410 11:24:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@331 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:16.410 11:24:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@331 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:16.668 11:25:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@331 -- # [[ true == \t\r\u\e ]] 00:14:16.668 11:25:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:16.668 11:25:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@333 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:14:16.927 11:25:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b NewBaseBdev -u 16770936-d48d-439b-a3f5-95f22c40a73c 00:14:17.185 [2024-07-15 11:25:00.562965] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:14:17.185 [2024-07-15 11:25:00.563117] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xd31e90 00:14:17.185 [2024-07-15 11:25:00.563131] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:17.185 [2024-07-15 11:25:00.563304] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xa38940 00:14:17.185 [2024-07-15 11:25:00.563417] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xd31e90 00:14:17.185 [2024-07-15 11:25:00.563427] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0xd31e90 00:14:17.185 [2024-07-15 11:25:00.563515] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:17.185 NewBaseBdev 00:14:17.185 11:25:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@334 -- # waitforbdev NewBaseBdev 00:14:17.185 11:25:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=NewBaseBdev 00:14:17.185 11:25:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:14:17.185 11:25:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:14:17.185 11:25:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:14:17.185 11:25:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:14:17.185 11:25:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:14:17.444 11:25:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev -t 2000 00:14:17.702 [ 00:14:17.702 { 00:14:17.702 "name": "NewBaseBdev", 00:14:17.702 "aliases": [ 00:14:17.702 "16770936-d48d-439b-a3f5-95f22c40a73c" 00:14:17.702 ], 00:14:17.702 "product_name": "Malloc disk", 00:14:17.702 "block_size": 512, 00:14:17.702 "num_blocks": 65536, 00:14:17.702 "uuid": "16770936-d48d-439b-a3f5-95f22c40a73c", 00:14:17.702 "assigned_rate_limits": { 00:14:17.702 "rw_ios_per_sec": 0, 00:14:17.702 "rw_mbytes_per_sec": 0, 00:14:17.702 "r_mbytes_per_sec": 0, 00:14:17.702 "w_mbytes_per_sec": 0 00:14:17.702 }, 00:14:17.702 "claimed": true, 00:14:17.702 "claim_type": "exclusive_write", 00:14:17.702 "zoned": false, 00:14:17.702 "supported_io_types": { 00:14:17.702 "read": true, 00:14:17.702 "write": true, 00:14:17.702 "unmap": true, 00:14:17.702 "flush": true, 00:14:17.702 "reset": true, 00:14:17.702 "nvme_admin": false, 00:14:17.702 "nvme_io": false, 00:14:17.702 "nvme_io_md": false, 00:14:17.702 "write_zeroes": true, 00:14:17.702 "zcopy": true, 00:14:17.702 "get_zone_info": false, 00:14:17.702 "zone_management": false, 00:14:17.702 "zone_append": false, 00:14:17.702 "compare": false, 00:14:17.702 "compare_and_write": false, 00:14:17.702 "abort": true, 00:14:17.702 "seek_hole": false, 00:14:17.702 "seek_data": false, 00:14:17.702 "copy": true, 00:14:17.702 "nvme_iov_md": false 00:14:17.702 }, 00:14:17.702 "memory_domains": [ 00:14:17.702 { 00:14:17.702 "dma_device_id": "system", 00:14:17.702 "dma_device_type": 1 00:14:17.702 }, 00:14:17.702 { 00:14:17.702 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:17.702 "dma_device_type": 2 00:14:17.702 } 00:14:17.702 ], 00:14:17.702 "driver_specific": {} 00:14:17.702 } 00:14:17.702 ] 00:14:17.702 11:25:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:14:17.702 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@335 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:14:17.702 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:14:17.702 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:14:17.702 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:14:17.702 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:17.702 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:17.702 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:17.702 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:17.702 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:17.702 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:17.702 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:17.702 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:17.960 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:17.960 "name": "Existed_Raid", 00:14:17.960 "uuid": "36187ade-b189-4a8e-aba2-c5c774536705", 00:14:17.960 "strip_size_kb": 64, 00:14:17.960 "state": "online", 00:14:17.960 "raid_level": "raid0", 00:14:17.960 "superblock": true, 00:14:17.960 "num_base_bdevs": 3, 00:14:17.960 "num_base_bdevs_discovered": 3, 00:14:17.960 "num_base_bdevs_operational": 3, 00:14:17.960 "base_bdevs_list": [ 00:14:17.960 { 00:14:17.960 "name": "NewBaseBdev", 00:14:17.960 "uuid": "16770936-d48d-439b-a3f5-95f22c40a73c", 00:14:17.960 "is_configured": true, 00:14:17.960 "data_offset": 2048, 00:14:17.960 "data_size": 63488 00:14:17.960 }, 00:14:17.960 { 00:14:17.960 "name": "BaseBdev2", 00:14:17.960 "uuid": "cbf7a58c-1289-4db7-9c8c-02c58647961e", 00:14:17.960 "is_configured": true, 00:14:17.960 "data_offset": 2048, 00:14:17.960 "data_size": 63488 00:14:17.960 }, 00:14:17.960 { 00:14:17.960 "name": "BaseBdev3", 00:14:17.960 "uuid": "16f39641-0697-473a-8035-ff1b1c52b8a0", 00:14:17.960 "is_configured": true, 00:14:17.960 "data_offset": 2048, 00:14:17.960 "data_size": 63488 00:14:17.960 } 00:14:17.960 ] 00:14:17.960 }' 00:14:17.960 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:17.960 11:25:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:18.527 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@336 -- # verify_raid_bdev_properties Existed_Raid 00:14:18.527 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:14:18.527 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:14:18.527 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:14:18.527 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:14:18.527 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # local name 00:14:18.527 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:14:18.527 11:25:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:14:18.786 [2024-07-15 11:25:02.147473] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:18.786 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:14:18.786 "name": "Existed_Raid", 00:14:18.786 "aliases": [ 00:14:18.786 "36187ade-b189-4a8e-aba2-c5c774536705" 00:14:18.786 ], 00:14:18.786 "product_name": "Raid Volume", 00:14:18.786 "block_size": 512, 00:14:18.786 "num_blocks": 190464, 00:14:18.786 "uuid": "36187ade-b189-4a8e-aba2-c5c774536705", 00:14:18.786 "assigned_rate_limits": { 00:14:18.786 "rw_ios_per_sec": 0, 00:14:18.786 "rw_mbytes_per_sec": 0, 00:14:18.786 "r_mbytes_per_sec": 0, 00:14:18.786 "w_mbytes_per_sec": 0 00:14:18.786 }, 00:14:18.786 "claimed": false, 00:14:18.786 "zoned": false, 00:14:18.786 "supported_io_types": { 00:14:18.786 "read": true, 00:14:18.786 "write": true, 00:14:18.786 "unmap": true, 00:14:18.786 "flush": true, 00:14:18.786 "reset": true, 00:14:18.786 "nvme_admin": false, 00:14:18.786 "nvme_io": false, 00:14:18.786 "nvme_io_md": false, 00:14:18.786 "write_zeroes": true, 00:14:18.786 "zcopy": false, 00:14:18.786 "get_zone_info": false, 00:14:18.786 "zone_management": false, 00:14:18.786 "zone_append": false, 00:14:18.786 "compare": false, 00:14:18.786 "compare_and_write": false, 00:14:18.786 "abort": false, 00:14:18.786 "seek_hole": false, 00:14:18.786 "seek_data": false, 00:14:18.786 "copy": false, 00:14:18.786 "nvme_iov_md": false 00:14:18.786 }, 00:14:18.786 "memory_domains": [ 00:14:18.786 { 00:14:18.786 "dma_device_id": "system", 00:14:18.786 "dma_device_type": 1 00:14:18.786 }, 00:14:18.786 { 00:14:18.786 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:18.786 "dma_device_type": 2 00:14:18.786 }, 00:14:18.786 { 00:14:18.786 "dma_device_id": "system", 00:14:18.786 "dma_device_type": 1 00:14:18.786 }, 00:14:18.786 { 00:14:18.786 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:18.786 "dma_device_type": 2 00:14:18.786 }, 00:14:18.786 { 00:14:18.786 "dma_device_id": "system", 00:14:18.786 "dma_device_type": 1 00:14:18.786 }, 00:14:18.786 { 00:14:18.786 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:18.786 "dma_device_type": 2 00:14:18.786 } 00:14:18.786 ], 00:14:18.786 "driver_specific": { 00:14:18.786 "raid": { 00:14:18.786 "uuid": "36187ade-b189-4a8e-aba2-c5c774536705", 00:14:18.786 "strip_size_kb": 64, 00:14:18.786 "state": "online", 00:14:18.786 "raid_level": "raid0", 00:14:18.786 "superblock": true, 00:14:18.786 "num_base_bdevs": 3, 00:14:18.786 "num_base_bdevs_discovered": 3, 00:14:18.786 "num_base_bdevs_operational": 3, 00:14:18.786 "base_bdevs_list": [ 00:14:18.786 { 00:14:18.786 "name": "NewBaseBdev", 00:14:18.786 "uuid": "16770936-d48d-439b-a3f5-95f22c40a73c", 00:14:18.786 "is_configured": true, 00:14:18.786 "data_offset": 2048, 00:14:18.786 "data_size": 63488 00:14:18.786 }, 00:14:18.786 { 00:14:18.786 "name": "BaseBdev2", 00:14:18.786 "uuid": "cbf7a58c-1289-4db7-9c8c-02c58647961e", 00:14:18.786 "is_configured": true, 00:14:18.786 "data_offset": 2048, 00:14:18.786 "data_size": 63488 00:14:18.786 }, 00:14:18.786 { 00:14:18.786 "name": "BaseBdev3", 00:14:18.786 "uuid": "16f39641-0697-473a-8035-ff1b1c52b8a0", 00:14:18.786 "is_configured": true, 00:14:18.786 "data_offset": 2048, 00:14:18.786 "data_size": 63488 00:14:18.786 } 00:14:18.786 ] 00:14:18.786 } 00:14:18.786 } 00:14:18.786 }' 00:14:18.786 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:18.786 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # base_bdev_names='NewBaseBdev 00:14:18.786 BaseBdev2 00:14:18.786 BaseBdev3' 00:14:18.786 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:14:18.786 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev 00:14:18.786 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:14:19.045 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:14:19.045 "name": "NewBaseBdev", 00:14:19.045 "aliases": [ 00:14:19.045 "16770936-d48d-439b-a3f5-95f22c40a73c" 00:14:19.045 ], 00:14:19.045 "product_name": "Malloc disk", 00:14:19.045 "block_size": 512, 00:14:19.045 "num_blocks": 65536, 00:14:19.045 "uuid": "16770936-d48d-439b-a3f5-95f22c40a73c", 00:14:19.045 "assigned_rate_limits": { 00:14:19.045 "rw_ios_per_sec": 0, 00:14:19.045 "rw_mbytes_per_sec": 0, 00:14:19.045 "r_mbytes_per_sec": 0, 00:14:19.045 "w_mbytes_per_sec": 0 00:14:19.045 }, 00:14:19.045 "claimed": true, 00:14:19.045 "claim_type": "exclusive_write", 00:14:19.045 "zoned": false, 00:14:19.045 "supported_io_types": { 00:14:19.045 "read": true, 00:14:19.045 "write": true, 00:14:19.045 "unmap": true, 00:14:19.045 "flush": true, 00:14:19.045 "reset": true, 00:14:19.045 "nvme_admin": false, 00:14:19.045 "nvme_io": false, 00:14:19.045 "nvme_io_md": false, 00:14:19.045 "write_zeroes": true, 00:14:19.045 "zcopy": true, 00:14:19.045 "get_zone_info": false, 00:14:19.045 "zone_management": false, 00:14:19.045 "zone_append": false, 00:14:19.045 "compare": false, 00:14:19.045 "compare_and_write": false, 00:14:19.045 "abort": true, 00:14:19.045 "seek_hole": false, 00:14:19.045 "seek_data": false, 00:14:19.045 "copy": true, 00:14:19.045 "nvme_iov_md": false 00:14:19.045 }, 00:14:19.045 "memory_domains": [ 00:14:19.045 { 00:14:19.045 "dma_device_id": "system", 00:14:19.045 "dma_device_type": 1 00:14:19.045 }, 00:14:19.045 { 00:14:19.045 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:19.045 "dma_device_type": 2 00:14:19.045 } 00:14:19.045 ], 00:14:19.045 "driver_specific": {} 00:14:19.045 }' 00:14:19.045 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:19.045 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:19.045 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:14:19.045 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:19.045 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:19.304 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:14:19.304 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:19.304 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:19.304 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:14:19.304 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:19.304 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:19.304 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:14:19.304 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:14:19.304 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:14:19.304 11:25:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:14:19.563 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:14:19.563 "name": "BaseBdev2", 00:14:19.563 "aliases": [ 00:14:19.563 "cbf7a58c-1289-4db7-9c8c-02c58647961e" 00:14:19.563 ], 00:14:19.563 "product_name": "Malloc disk", 00:14:19.563 "block_size": 512, 00:14:19.563 "num_blocks": 65536, 00:14:19.563 "uuid": "cbf7a58c-1289-4db7-9c8c-02c58647961e", 00:14:19.563 "assigned_rate_limits": { 00:14:19.563 "rw_ios_per_sec": 0, 00:14:19.563 "rw_mbytes_per_sec": 0, 00:14:19.563 "r_mbytes_per_sec": 0, 00:14:19.563 "w_mbytes_per_sec": 0 00:14:19.563 }, 00:14:19.563 "claimed": true, 00:14:19.563 "claim_type": "exclusive_write", 00:14:19.563 "zoned": false, 00:14:19.563 "supported_io_types": { 00:14:19.563 "read": true, 00:14:19.563 "write": true, 00:14:19.563 "unmap": true, 00:14:19.563 "flush": true, 00:14:19.563 "reset": true, 00:14:19.563 "nvme_admin": false, 00:14:19.563 "nvme_io": false, 00:14:19.563 "nvme_io_md": false, 00:14:19.563 "write_zeroes": true, 00:14:19.563 "zcopy": true, 00:14:19.563 "get_zone_info": false, 00:14:19.563 "zone_management": false, 00:14:19.563 "zone_append": false, 00:14:19.563 "compare": false, 00:14:19.563 "compare_and_write": false, 00:14:19.563 "abort": true, 00:14:19.563 "seek_hole": false, 00:14:19.563 "seek_data": false, 00:14:19.563 "copy": true, 00:14:19.563 "nvme_iov_md": false 00:14:19.563 }, 00:14:19.563 "memory_domains": [ 00:14:19.563 { 00:14:19.563 "dma_device_id": "system", 00:14:19.563 "dma_device_type": 1 00:14:19.563 }, 00:14:19.563 { 00:14:19.563 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:19.563 "dma_device_type": 2 00:14:19.563 } 00:14:19.563 ], 00:14:19.563 "driver_specific": {} 00:14:19.563 }' 00:14:19.563 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:19.563 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:19.563 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:14:19.822 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:19.822 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:19.822 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:14:19.822 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:19.822 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:19.822 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:14:19.822 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:19.822 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:20.081 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:14:20.081 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:14:20.081 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:14:20.081 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:14:20.081 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:14:20.081 "name": "BaseBdev3", 00:14:20.081 "aliases": [ 00:14:20.081 "16f39641-0697-473a-8035-ff1b1c52b8a0" 00:14:20.081 ], 00:14:20.081 "product_name": "Malloc disk", 00:14:20.081 "block_size": 512, 00:14:20.081 "num_blocks": 65536, 00:14:20.081 "uuid": "16f39641-0697-473a-8035-ff1b1c52b8a0", 00:14:20.081 "assigned_rate_limits": { 00:14:20.081 "rw_ios_per_sec": 0, 00:14:20.081 "rw_mbytes_per_sec": 0, 00:14:20.081 "r_mbytes_per_sec": 0, 00:14:20.081 "w_mbytes_per_sec": 0 00:14:20.081 }, 00:14:20.081 "claimed": true, 00:14:20.081 "claim_type": "exclusive_write", 00:14:20.081 "zoned": false, 00:14:20.081 "supported_io_types": { 00:14:20.081 "read": true, 00:14:20.081 "write": true, 00:14:20.081 "unmap": true, 00:14:20.081 "flush": true, 00:14:20.081 "reset": true, 00:14:20.081 "nvme_admin": false, 00:14:20.081 "nvme_io": false, 00:14:20.081 "nvme_io_md": false, 00:14:20.081 "write_zeroes": true, 00:14:20.081 "zcopy": true, 00:14:20.081 "get_zone_info": false, 00:14:20.081 "zone_management": false, 00:14:20.081 "zone_append": false, 00:14:20.081 "compare": false, 00:14:20.081 "compare_and_write": false, 00:14:20.081 "abort": true, 00:14:20.081 "seek_hole": false, 00:14:20.081 "seek_data": false, 00:14:20.081 "copy": true, 00:14:20.081 "nvme_iov_md": false 00:14:20.081 }, 00:14:20.081 "memory_domains": [ 00:14:20.081 { 00:14:20.081 "dma_device_id": "system", 00:14:20.081 "dma_device_type": 1 00:14:20.081 }, 00:14:20.081 { 00:14:20.081 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:20.081 "dma_device_type": 2 00:14:20.081 } 00:14:20.081 ], 00:14:20.081 "driver_specific": {} 00:14:20.081 }' 00:14:20.081 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:20.339 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:20.339 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:14:20.339 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:20.339 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:20.339 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:14:20.339 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:20.339 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:20.598 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:14:20.598 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:20.598 11:25:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:20.598 11:25:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:14:20.598 11:25:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@338 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:14:20.857 [2024-07-15 11:25:04.260846] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:20.857 [2024-07-15 11:25:04.260872] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:20.857 [2024-07-15 11:25:04.260932] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:20.857 [2024-07-15 11:25:04.260984] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:20.857 [2024-07-15 11:25:04.260996] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xd31e90 name Existed_Raid, state offline 00:14:20.857 11:25:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@341 -- # killprocess 887080 00:14:20.857 11:25:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@948 -- # '[' -z 887080 ']' 00:14:20.857 11:25:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@952 -- # kill -0 887080 00:14:20.857 11:25:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@953 -- # uname 00:14:20.857 11:25:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:14:20.857 11:25:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 887080 00:14:20.857 11:25:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:14:20.857 11:25:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:14:20.857 11:25:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@966 -- # echo 'killing process with pid 887080' 00:14:20.857 killing process with pid 887080 00:14:20.857 11:25:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@967 -- # kill 887080 00:14:20.857 [2024-07-15 11:25:04.328187] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:20.857 11:25:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # wait 887080 00:14:20.857 [2024-07-15 11:25:04.358662] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:21.116 11:25:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@343 -- # return 0 00:14:21.116 00:14:21.116 real 0m28.082s 00:14:21.116 user 0m51.387s 00:14:21.116 sys 0m5.146s 00:14:21.116 11:25:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1124 -- # xtrace_disable 00:14:21.116 11:25:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.116 ************************************ 00:14:21.116 END TEST raid_state_function_test_sb 00:14:21.116 ************************************ 00:14:21.116 11:25:04 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:14:21.116 11:25:04 bdev_raid -- bdev/bdev_raid.sh@869 -- # run_test raid_superblock_test raid_superblock_test raid0 3 00:14:21.116 11:25:04 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:14:21.117 11:25:04 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:14:21.117 11:25:04 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:21.117 ************************************ 00:14:21.117 START TEST raid_superblock_test 00:14:21.117 ************************************ 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1123 -- # raid_superblock_test raid0 3 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@392 -- # local raid_level=raid0 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local num_base_bdevs=3 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # base_bdevs_malloc=() 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local base_bdevs_malloc 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_pt=() 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_pt 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt_uuid=() 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt_uuid 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local raid_bdev_name=raid_bdev1 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local strip_size 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size_create_arg 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local raid_bdev_uuid 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@403 -- # '[' raid0 '!=' raid1 ']' 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # strip_size=64 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size_create_arg='-z 64' 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # raid_pid=891380 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@410 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -L bdev_raid 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # waitforlisten 891380 /var/tmp/spdk-raid.sock 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@829 -- # '[' -z 891380 ']' 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:14:21.117 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:14:21.117 11:25:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:21.376 [2024-07-15 11:25:04.721816] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:14:21.376 [2024-07-15 11:25:04.721878] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid891380 ] 00:14:21.376 [2024-07-15 11:25:04.847742] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:21.376 [2024-07-15 11:25:04.951586] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:14:21.635 [2024-07-15 11:25:05.023920] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:21.635 [2024-07-15 11:25:05.023967] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:22.202 11:25:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:14:22.202 11:25:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@862 -- # return 0 00:14:22.202 11:25:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i = 1 )) 00:14:22.202 11:25:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:14:22.202 11:25:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc1 00:14:22.202 11:25:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt1 00:14:22.202 11:25:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:14:22.202 11:25:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:22.202 11:25:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:14:22.202 11:25:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:22.202 11:25:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc1 00:14:22.462 malloc1 00:14:22.462 11:25:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:22.721 [2024-07-15 11:25:06.147710] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:22.721 [2024-07-15 11:25:06.147757] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:22.721 [2024-07-15 11:25:06.147779] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x14b2570 00:14:22.721 [2024-07-15 11:25:06.147791] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:22.721 [2024-07-15 11:25:06.149481] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:22.721 [2024-07-15 11:25:06.149508] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:22.721 pt1 00:14:22.721 11:25:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:14:22.721 11:25:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:14:22.721 11:25:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc2 00:14:22.721 11:25:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt2 00:14:22.721 11:25:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:14:22.721 11:25:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:22.721 11:25:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:14:22.721 11:25:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:22.721 11:25:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc2 00:14:22.980 malloc2 00:14:22.981 11:25:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:23.240 [2024-07-15 11:25:06.642772] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:23.240 [2024-07-15 11:25:06.642820] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:23.240 [2024-07-15 11:25:06.642844] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x14b3970 00:14:23.240 [2024-07-15 11:25:06.642857] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:23.240 [2024-07-15 11:25:06.644547] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:23.240 [2024-07-15 11:25:06.644574] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:23.240 pt2 00:14:23.240 11:25:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:14:23.240 11:25:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:14:23.240 11:25:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc3 00:14:23.240 11:25:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt3 00:14:23.240 11:25:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:14:23.240 11:25:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:23.240 11:25:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:14:23.240 11:25:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:23.240 11:25:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc3 00:14:23.499 malloc3 00:14:23.499 11:25:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:23.758 [2024-07-15 11:25:07.121867] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:23.758 [2024-07-15 11:25:07.121911] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:23.758 [2024-07-15 11:25:07.121934] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x164a340 00:14:23.758 [2024-07-15 11:25:07.121946] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:23.758 [2024-07-15 11:25:07.123491] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:23.758 [2024-07-15 11:25:07.123518] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:23.758 pt3 00:14:23.758 11:25:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:14:23.758 11:25:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:14:23.758 11:25:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@429 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'pt1 pt2 pt3' -n raid_bdev1 -s 00:14:24.017 [2024-07-15 11:25:07.366536] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:24.017 [2024-07-15 11:25:07.367874] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:24.017 [2024-07-15 11:25:07.367940] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:24.017 [2024-07-15 11:25:07.368092] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x14aaea0 00:14:24.017 [2024-07-15 11:25:07.368104] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:24.017 [2024-07-15 11:25:07.368304] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x14b2240 00:14:24.017 [2024-07-15 11:25:07.368445] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x14aaea0 00:14:24.017 [2024-07-15 11:25:07.368455] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x14aaea0 00:14:24.018 [2024-07-15 11:25:07.368553] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:24.018 11:25:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:14:24.018 11:25:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:14:24.018 11:25:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:14:24.018 11:25:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:14:24.018 11:25:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:24.018 11:25:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:24.018 11:25:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:24.018 11:25:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:24.018 11:25:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:24.018 11:25:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:24.018 11:25:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:24.018 11:25:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:24.276 11:25:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:24.276 "name": "raid_bdev1", 00:14:24.276 "uuid": "a29d8dc6-e7ef-4cc0-92ab-c81e8df169f4", 00:14:24.276 "strip_size_kb": 64, 00:14:24.276 "state": "online", 00:14:24.276 "raid_level": "raid0", 00:14:24.276 "superblock": true, 00:14:24.276 "num_base_bdevs": 3, 00:14:24.276 "num_base_bdevs_discovered": 3, 00:14:24.276 "num_base_bdevs_operational": 3, 00:14:24.276 "base_bdevs_list": [ 00:14:24.276 { 00:14:24.276 "name": "pt1", 00:14:24.276 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:24.276 "is_configured": true, 00:14:24.276 "data_offset": 2048, 00:14:24.276 "data_size": 63488 00:14:24.276 }, 00:14:24.276 { 00:14:24.276 "name": "pt2", 00:14:24.276 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:24.276 "is_configured": true, 00:14:24.276 "data_offset": 2048, 00:14:24.276 "data_size": 63488 00:14:24.276 }, 00:14:24.276 { 00:14:24.276 "name": "pt3", 00:14:24.276 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:24.276 "is_configured": true, 00:14:24.276 "data_offset": 2048, 00:14:24.276 "data_size": 63488 00:14:24.276 } 00:14:24.276 ] 00:14:24.276 }' 00:14:24.276 11:25:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:24.276 11:25:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:24.843 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_properties raid_bdev1 00:14:24.843 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:14:24.843 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:14:24.843 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:14:24.843 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:14:24.843 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # local name 00:14:24.843 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:14:24.843 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:14:24.843 [2024-07-15 11:25:08.365427] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:24.843 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:14:24.843 "name": "raid_bdev1", 00:14:24.843 "aliases": [ 00:14:24.843 "a29d8dc6-e7ef-4cc0-92ab-c81e8df169f4" 00:14:24.843 ], 00:14:24.843 "product_name": "Raid Volume", 00:14:24.843 "block_size": 512, 00:14:24.843 "num_blocks": 190464, 00:14:24.843 "uuid": "a29d8dc6-e7ef-4cc0-92ab-c81e8df169f4", 00:14:24.843 "assigned_rate_limits": { 00:14:24.843 "rw_ios_per_sec": 0, 00:14:24.843 "rw_mbytes_per_sec": 0, 00:14:24.843 "r_mbytes_per_sec": 0, 00:14:24.843 "w_mbytes_per_sec": 0 00:14:24.843 }, 00:14:24.843 "claimed": false, 00:14:24.843 "zoned": false, 00:14:24.843 "supported_io_types": { 00:14:24.843 "read": true, 00:14:24.843 "write": true, 00:14:24.843 "unmap": true, 00:14:24.843 "flush": true, 00:14:24.843 "reset": true, 00:14:24.843 "nvme_admin": false, 00:14:24.843 "nvme_io": false, 00:14:24.843 "nvme_io_md": false, 00:14:24.843 "write_zeroes": true, 00:14:24.843 "zcopy": false, 00:14:24.843 "get_zone_info": false, 00:14:24.843 "zone_management": false, 00:14:24.843 "zone_append": false, 00:14:24.843 "compare": false, 00:14:24.843 "compare_and_write": false, 00:14:24.843 "abort": false, 00:14:24.843 "seek_hole": false, 00:14:24.843 "seek_data": false, 00:14:24.843 "copy": false, 00:14:24.843 "nvme_iov_md": false 00:14:24.843 }, 00:14:24.843 "memory_domains": [ 00:14:24.843 { 00:14:24.843 "dma_device_id": "system", 00:14:24.843 "dma_device_type": 1 00:14:24.843 }, 00:14:24.843 { 00:14:24.843 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:24.843 "dma_device_type": 2 00:14:24.843 }, 00:14:24.843 { 00:14:24.843 "dma_device_id": "system", 00:14:24.843 "dma_device_type": 1 00:14:24.843 }, 00:14:24.843 { 00:14:24.843 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:24.843 "dma_device_type": 2 00:14:24.843 }, 00:14:24.843 { 00:14:24.843 "dma_device_id": "system", 00:14:24.843 "dma_device_type": 1 00:14:24.843 }, 00:14:24.843 { 00:14:24.843 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:24.843 "dma_device_type": 2 00:14:24.843 } 00:14:24.843 ], 00:14:24.843 "driver_specific": { 00:14:24.843 "raid": { 00:14:24.843 "uuid": "a29d8dc6-e7ef-4cc0-92ab-c81e8df169f4", 00:14:24.843 "strip_size_kb": 64, 00:14:24.843 "state": "online", 00:14:24.843 "raid_level": "raid0", 00:14:24.843 "superblock": true, 00:14:24.843 "num_base_bdevs": 3, 00:14:24.843 "num_base_bdevs_discovered": 3, 00:14:24.843 "num_base_bdevs_operational": 3, 00:14:24.843 "base_bdevs_list": [ 00:14:24.843 { 00:14:24.843 "name": "pt1", 00:14:24.843 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:24.843 "is_configured": true, 00:14:24.843 "data_offset": 2048, 00:14:24.843 "data_size": 63488 00:14:24.843 }, 00:14:24.843 { 00:14:24.843 "name": "pt2", 00:14:24.843 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:24.843 "is_configured": true, 00:14:24.843 "data_offset": 2048, 00:14:24.843 "data_size": 63488 00:14:24.843 }, 00:14:24.843 { 00:14:24.843 "name": "pt3", 00:14:24.843 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:24.843 "is_configured": true, 00:14:24.843 "data_offset": 2048, 00:14:24.843 "data_size": 63488 00:14:24.843 } 00:14:24.843 ] 00:14:24.843 } 00:14:24.843 } 00:14:24.843 }' 00:14:24.843 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:24.843 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:14:24.843 pt2 00:14:24.843 pt3' 00:14:24.843 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:14:25.102 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:14:25.102 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:14:25.102 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:14:25.102 "name": "pt1", 00:14:25.102 "aliases": [ 00:14:25.102 "00000000-0000-0000-0000-000000000001" 00:14:25.102 ], 00:14:25.102 "product_name": "passthru", 00:14:25.102 "block_size": 512, 00:14:25.102 "num_blocks": 65536, 00:14:25.102 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:25.102 "assigned_rate_limits": { 00:14:25.102 "rw_ios_per_sec": 0, 00:14:25.102 "rw_mbytes_per_sec": 0, 00:14:25.102 "r_mbytes_per_sec": 0, 00:14:25.102 "w_mbytes_per_sec": 0 00:14:25.102 }, 00:14:25.102 "claimed": true, 00:14:25.102 "claim_type": "exclusive_write", 00:14:25.102 "zoned": false, 00:14:25.102 "supported_io_types": { 00:14:25.102 "read": true, 00:14:25.102 "write": true, 00:14:25.102 "unmap": true, 00:14:25.102 "flush": true, 00:14:25.102 "reset": true, 00:14:25.102 "nvme_admin": false, 00:14:25.102 "nvme_io": false, 00:14:25.102 "nvme_io_md": false, 00:14:25.102 "write_zeroes": true, 00:14:25.102 "zcopy": true, 00:14:25.102 "get_zone_info": false, 00:14:25.102 "zone_management": false, 00:14:25.102 "zone_append": false, 00:14:25.102 "compare": false, 00:14:25.102 "compare_and_write": false, 00:14:25.102 "abort": true, 00:14:25.102 "seek_hole": false, 00:14:25.102 "seek_data": false, 00:14:25.102 "copy": true, 00:14:25.102 "nvme_iov_md": false 00:14:25.102 }, 00:14:25.102 "memory_domains": [ 00:14:25.102 { 00:14:25.102 "dma_device_id": "system", 00:14:25.102 "dma_device_type": 1 00:14:25.102 }, 00:14:25.102 { 00:14:25.102 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:25.102 "dma_device_type": 2 00:14:25.102 } 00:14:25.102 ], 00:14:25.102 "driver_specific": { 00:14:25.102 "passthru": { 00:14:25.102 "name": "pt1", 00:14:25.102 "base_bdev_name": "malloc1" 00:14:25.102 } 00:14:25.102 } 00:14:25.102 }' 00:14:25.102 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:25.361 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:25.361 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:14:25.361 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:25.361 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:25.361 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:14:25.361 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:25.361 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:25.361 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:14:25.361 11:25:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:25.620 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:25.620 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:14:25.620 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:14:25.620 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:14:25.620 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:14:25.879 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:14:25.879 "name": "pt2", 00:14:25.879 "aliases": [ 00:14:25.879 "00000000-0000-0000-0000-000000000002" 00:14:25.879 ], 00:14:25.879 "product_name": "passthru", 00:14:25.879 "block_size": 512, 00:14:25.879 "num_blocks": 65536, 00:14:25.879 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:25.879 "assigned_rate_limits": { 00:14:25.879 "rw_ios_per_sec": 0, 00:14:25.879 "rw_mbytes_per_sec": 0, 00:14:25.879 "r_mbytes_per_sec": 0, 00:14:25.879 "w_mbytes_per_sec": 0 00:14:25.879 }, 00:14:25.879 "claimed": true, 00:14:25.879 "claim_type": "exclusive_write", 00:14:25.879 "zoned": false, 00:14:25.879 "supported_io_types": { 00:14:25.879 "read": true, 00:14:25.879 "write": true, 00:14:25.879 "unmap": true, 00:14:25.879 "flush": true, 00:14:25.879 "reset": true, 00:14:25.879 "nvme_admin": false, 00:14:25.879 "nvme_io": false, 00:14:25.879 "nvme_io_md": false, 00:14:25.879 "write_zeroes": true, 00:14:25.879 "zcopy": true, 00:14:25.879 "get_zone_info": false, 00:14:25.879 "zone_management": false, 00:14:25.879 "zone_append": false, 00:14:25.879 "compare": false, 00:14:25.879 "compare_and_write": false, 00:14:25.879 "abort": true, 00:14:25.879 "seek_hole": false, 00:14:25.879 "seek_data": false, 00:14:25.879 "copy": true, 00:14:25.879 "nvme_iov_md": false 00:14:25.879 }, 00:14:25.879 "memory_domains": [ 00:14:25.879 { 00:14:25.879 "dma_device_id": "system", 00:14:25.879 "dma_device_type": 1 00:14:25.879 }, 00:14:25.879 { 00:14:25.880 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:25.880 "dma_device_type": 2 00:14:25.880 } 00:14:25.880 ], 00:14:25.880 "driver_specific": { 00:14:25.880 "passthru": { 00:14:25.880 "name": "pt2", 00:14:25.880 "base_bdev_name": "malloc2" 00:14:25.880 } 00:14:25.880 } 00:14:25.880 }' 00:14:25.880 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:25.880 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:25.880 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:14:25.880 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:25.880 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:25.880 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:14:25.880 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:25.880 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:26.139 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:14:26.139 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:26.139 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:26.139 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:14:26.139 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:14:26.139 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt3 00:14:26.139 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:14:26.397 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:14:26.397 "name": "pt3", 00:14:26.397 "aliases": [ 00:14:26.397 "00000000-0000-0000-0000-000000000003" 00:14:26.397 ], 00:14:26.397 "product_name": "passthru", 00:14:26.397 "block_size": 512, 00:14:26.397 "num_blocks": 65536, 00:14:26.397 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:26.397 "assigned_rate_limits": { 00:14:26.397 "rw_ios_per_sec": 0, 00:14:26.397 "rw_mbytes_per_sec": 0, 00:14:26.397 "r_mbytes_per_sec": 0, 00:14:26.397 "w_mbytes_per_sec": 0 00:14:26.397 }, 00:14:26.397 "claimed": true, 00:14:26.397 "claim_type": "exclusive_write", 00:14:26.397 "zoned": false, 00:14:26.397 "supported_io_types": { 00:14:26.397 "read": true, 00:14:26.397 "write": true, 00:14:26.397 "unmap": true, 00:14:26.397 "flush": true, 00:14:26.397 "reset": true, 00:14:26.397 "nvme_admin": false, 00:14:26.397 "nvme_io": false, 00:14:26.397 "nvme_io_md": false, 00:14:26.397 "write_zeroes": true, 00:14:26.397 "zcopy": true, 00:14:26.397 "get_zone_info": false, 00:14:26.397 "zone_management": false, 00:14:26.397 "zone_append": false, 00:14:26.397 "compare": false, 00:14:26.397 "compare_and_write": false, 00:14:26.397 "abort": true, 00:14:26.397 "seek_hole": false, 00:14:26.397 "seek_data": false, 00:14:26.397 "copy": true, 00:14:26.397 "nvme_iov_md": false 00:14:26.397 }, 00:14:26.397 "memory_domains": [ 00:14:26.397 { 00:14:26.397 "dma_device_id": "system", 00:14:26.397 "dma_device_type": 1 00:14:26.397 }, 00:14:26.397 { 00:14:26.397 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:26.397 "dma_device_type": 2 00:14:26.397 } 00:14:26.397 ], 00:14:26.397 "driver_specific": { 00:14:26.397 "passthru": { 00:14:26.397 "name": "pt3", 00:14:26.397 "base_bdev_name": "malloc3" 00:14:26.397 } 00:14:26.397 } 00:14:26.397 }' 00:14:26.397 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:26.397 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:26.397 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:14:26.397 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:26.397 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:26.397 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:14:26.397 11:25:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:26.655 11:25:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:26.655 11:25:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:14:26.655 11:25:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:26.655 11:25:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:26.655 11:25:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:14:26.655 11:25:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:14:26.655 11:25:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # jq -r '.[] | .uuid' 00:14:26.928 [2024-07-15 11:25:10.374813] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:26.928 11:25:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # raid_bdev_uuid=a29d8dc6-e7ef-4cc0-92ab-c81e8df169f4 00:14:26.928 11:25:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # '[' -z a29d8dc6-e7ef-4cc0-92ab-c81e8df169f4 ']' 00:14:26.928 11:25:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@440 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:14:27.197 [2024-07-15 11:25:10.619180] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:27.197 [2024-07-15 11:25:10.619202] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:27.197 [2024-07-15 11:25:10.619251] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:27.197 [2024-07-15 11:25:10.619303] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:27.197 [2024-07-15 11:25:10.619315] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x14aaea0 name raid_bdev1, state offline 00:14:27.197 11:25:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:27.197 11:25:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # jq -r '.[]' 00:14:27.455 11:25:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # raid_bdev= 00:14:27.455 11:25:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # '[' -n '' ']' 00:14:27.455 11:25:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:14:27.455 11:25:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt1 00:14:27.712 11:25:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:14:27.712 11:25:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:14:28.003 11:25:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:14:28.003 11:25:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt3 00:14:28.260 11:25:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs 00:14:28.260 11:25:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:14:28.260 11:25:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # '[' false == true ']' 00:14:28.260 11:25:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@456 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'malloc1 malloc2 malloc3' -n raid_bdev1 00:14:28.260 11:25:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@648 -- # local es=0 00:14:28.260 11:25:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'malloc1 malloc2 malloc3' -n raid_bdev1 00:14:28.260 11:25:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:14:28.260 11:25:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:14:28.260 11:25:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:14:28.260 11:25:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:14:28.260 11:25:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:14:28.260 11:25:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:14:28.260 11:25:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:14:28.516 11:25:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py ]] 00:14:28.516 11:25:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'malloc1 malloc2 malloc3' -n raid_bdev1 00:14:28.516 [2024-07-15 11:25:12.079111] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:14:28.516 [2024-07-15 11:25:12.080443] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:14:28.516 [2024-07-15 11:25:12.080484] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:14:28.516 [2024-07-15 11:25:12.080529] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:14:28.516 [2024-07-15 11:25:12.080566] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:14:28.516 [2024-07-15 11:25:12.080589] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:14:28.516 [2024-07-15 11:25:12.080606] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:28.516 [2024-07-15 11:25:12.080617] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1655ff0 name raid_bdev1, state configuring 00:14:28.516 request: 00:14:28.516 { 00:14:28.516 "name": "raid_bdev1", 00:14:28.516 "raid_level": "raid0", 00:14:28.516 "base_bdevs": [ 00:14:28.516 "malloc1", 00:14:28.516 "malloc2", 00:14:28.516 "malloc3" 00:14:28.516 ], 00:14:28.516 "strip_size_kb": 64, 00:14:28.516 "superblock": false, 00:14:28.516 "method": "bdev_raid_create", 00:14:28.516 "req_id": 1 00:14:28.516 } 00:14:28.516 Got JSON-RPC error response 00:14:28.516 response: 00:14:28.516 { 00:14:28.516 "code": -17, 00:14:28.516 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:14:28.516 } 00:14:28.516 11:25:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@651 -- # es=1 00:14:28.516 11:25:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:14:28.516 11:25:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:14:28.516 11:25:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:14:28.516 11:25:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:28.516 11:25:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # jq -r '.[]' 00:14:28.772 11:25:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # raid_bdev= 00:14:28.772 11:25:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # '[' -n '' ']' 00:14:28.772 11:25:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@464 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:29.029 [2024-07-15 11:25:12.564379] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:29.029 [2024-07-15 11:25:12.564423] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:29.029 [2024-07-15 11:25:12.564444] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x14b27a0 00:14:29.029 [2024-07-15 11:25:12.564456] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:29.029 [2024-07-15 11:25:12.566093] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:29.029 [2024-07-15 11:25:12.566121] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:29.029 [2024-07-15 11:25:12.566181] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:14:29.029 [2024-07-15 11:25:12.566208] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:29.029 pt1 00:14:29.029 11:25:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@467 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:14:29.029 11:25:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:14:29.029 11:25:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:14:29.029 11:25:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:14:29.029 11:25:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:29.029 11:25:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:29.029 11:25:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:29.029 11:25:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:29.029 11:25:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:29.029 11:25:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:29.029 11:25:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:29.029 11:25:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:29.287 11:25:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:29.287 "name": "raid_bdev1", 00:14:29.287 "uuid": "a29d8dc6-e7ef-4cc0-92ab-c81e8df169f4", 00:14:29.287 "strip_size_kb": 64, 00:14:29.287 "state": "configuring", 00:14:29.287 "raid_level": "raid0", 00:14:29.287 "superblock": true, 00:14:29.287 "num_base_bdevs": 3, 00:14:29.287 "num_base_bdevs_discovered": 1, 00:14:29.287 "num_base_bdevs_operational": 3, 00:14:29.287 "base_bdevs_list": [ 00:14:29.287 { 00:14:29.287 "name": "pt1", 00:14:29.287 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:29.287 "is_configured": true, 00:14:29.287 "data_offset": 2048, 00:14:29.287 "data_size": 63488 00:14:29.287 }, 00:14:29.287 { 00:14:29.287 "name": null, 00:14:29.287 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:29.287 "is_configured": false, 00:14:29.287 "data_offset": 2048, 00:14:29.287 "data_size": 63488 00:14:29.287 }, 00:14:29.287 { 00:14:29.287 "name": null, 00:14:29.287 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:29.287 "is_configured": false, 00:14:29.287 "data_offset": 2048, 00:14:29.287 "data_size": 63488 00:14:29.287 } 00:14:29.287 ] 00:14:29.287 }' 00:14:29.287 11:25:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:29.287 11:25:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.220 11:25:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@469 -- # '[' 3 -gt 2 ']' 00:14:30.220 11:25:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@471 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:30.220 [2024-07-15 11:25:13.683373] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:30.220 [2024-07-15 11:25:13.683419] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:30.220 [2024-07-15 11:25:13.683438] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x14a9c70 00:14:30.220 [2024-07-15 11:25:13.683450] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:30.220 [2024-07-15 11:25:13.683793] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:30.220 [2024-07-15 11:25:13.683810] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:30.220 [2024-07-15 11:25:13.683869] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:30.220 [2024-07-15 11:25:13.683888] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:30.220 pt2 00:14:30.220 11:25:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:14:30.478 [2024-07-15 11:25:13.924035] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:14:30.478 11:25:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:14:30.478 11:25:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:14:30.478 11:25:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:14:30.478 11:25:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:14:30.478 11:25:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:30.478 11:25:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:30.478 11:25:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:30.478 11:25:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:30.478 11:25:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:30.478 11:25:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:30.478 11:25:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:30.478 11:25:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:30.737 11:25:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:30.737 "name": "raid_bdev1", 00:14:30.737 "uuid": "a29d8dc6-e7ef-4cc0-92ab-c81e8df169f4", 00:14:30.737 "strip_size_kb": 64, 00:14:30.737 "state": "configuring", 00:14:30.737 "raid_level": "raid0", 00:14:30.737 "superblock": true, 00:14:30.737 "num_base_bdevs": 3, 00:14:30.737 "num_base_bdevs_discovered": 1, 00:14:30.737 "num_base_bdevs_operational": 3, 00:14:30.737 "base_bdevs_list": [ 00:14:30.737 { 00:14:30.737 "name": "pt1", 00:14:30.737 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:30.737 "is_configured": true, 00:14:30.737 "data_offset": 2048, 00:14:30.737 "data_size": 63488 00:14:30.737 }, 00:14:30.737 { 00:14:30.737 "name": null, 00:14:30.737 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:30.737 "is_configured": false, 00:14:30.737 "data_offset": 2048, 00:14:30.737 "data_size": 63488 00:14:30.737 }, 00:14:30.737 { 00:14:30.737 "name": null, 00:14:30.737 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:30.737 "is_configured": false, 00:14:30.737 "data_offset": 2048, 00:14:30.737 "data_size": 63488 00:14:30.737 } 00:14:30.737 ] 00:14:30.737 }' 00:14:30.737 11:25:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:30.737 11:25:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.303 11:25:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i = 1 )) 00:14:31.303 11:25:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:14:31.303 11:25:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:31.562 [2024-07-15 11:25:15.010918] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:31.562 [2024-07-15 11:25:15.010969] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:31.562 [2024-07-15 11:25:15.010987] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x164afa0 00:14:31.562 [2024-07-15 11:25:15.011000] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:31.562 [2024-07-15 11:25:15.011324] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:31.562 [2024-07-15 11:25:15.011342] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:31.562 [2024-07-15 11:25:15.011403] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:31.562 [2024-07-15 11:25:15.011421] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:31.562 pt2 00:14:31.562 11:25:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:14:31.562 11:25:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:14:31.562 11:25:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:31.820 [2024-07-15 11:25:15.255565] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:31.820 [2024-07-15 11:25:15.255597] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:31.820 [2024-07-15 11:25:15.255614] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x164bb30 00:14:31.820 [2024-07-15 11:25:15.255626] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:31.820 [2024-07-15 11:25:15.255914] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:31.820 [2024-07-15 11:25:15.255943] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:31.820 [2024-07-15 11:25:15.255997] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:14:31.820 [2024-07-15 11:25:15.256014] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:31.820 [2024-07-15 11:25:15.256118] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x164cc00 00:14:31.820 [2024-07-15 11:25:15.256128] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:31.820 [2024-07-15 11:25:15.256292] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x16559b0 00:14:31.820 [2024-07-15 11:25:15.256414] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x164cc00 00:14:31.820 [2024-07-15 11:25:15.256424] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x164cc00 00:14:31.820 [2024-07-15 11:25:15.256517] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:31.820 pt3 00:14:31.820 11:25:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:14:31.820 11:25:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:14:31.820 11:25:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@482 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:14:31.820 11:25:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:14:31.820 11:25:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:14:31.820 11:25:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:14:31.820 11:25:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:31.820 11:25:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:31.820 11:25:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:31.820 11:25:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:31.820 11:25:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:31.820 11:25:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:31.820 11:25:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:31.820 11:25:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:32.078 11:25:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:32.078 "name": "raid_bdev1", 00:14:32.078 "uuid": "a29d8dc6-e7ef-4cc0-92ab-c81e8df169f4", 00:14:32.078 "strip_size_kb": 64, 00:14:32.078 "state": "online", 00:14:32.078 "raid_level": "raid0", 00:14:32.078 "superblock": true, 00:14:32.078 "num_base_bdevs": 3, 00:14:32.078 "num_base_bdevs_discovered": 3, 00:14:32.078 "num_base_bdevs_operational": 3, 00:14:32.078 "base_bdevs_list": [ 00:14:32.078 { 00:14:32.078 "name": "pt1", 00:14:32.078 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:32.078 "is_configured": true, 00:14:32.078 "data_offset": 2048, 00:14:32.078 "data_size": 63488 00:14:32.078 }, 00:14:32.078 { 00:14:32.078 "name": "pt2", 00:14:32.078 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:32.078 "is_configured": true, 00:14:32.078 "data_offset": 2048, 00:14:32.078 "data_size": 63488 00:14:32.078 }, 00:14:32.078 { 00:14:32.078 "name": "pt3", 00:14:32.079 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:32.079 "is_configured": true, 00:14:32.079 "data_offset": 2048, 00:14:32.079 "data_size": 63488 00:14:32.079 } 00:14:32.079 ] 00:14:32.079 }' 00:14:32.079 11:25:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:32.079 11:25:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.644 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_properties raid_bdev1 00:14:32.644 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:14:32.645 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:14:32.645 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:14:32.645 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:14:32.645 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # local name 00:14:32.645 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:14:32.645 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:14:32.902 [2024-07-15 11:25:16.290590] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:32.902 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:14:32.902 "name": "raid_bdev1", 00:14:32.902 "aliases": [ 00:14:32.902 "a29d8dc6-e7ef-4cc0-92ab-c81e8df169f4" 00:14:32.902 ], 00:14:32.902 "product_name": "Raid Volume", 00:14:32.902 "block_size": 512, 00:14:32.902 "num_blocks": 190464, 00:14:32.902 "uuid": "a29d8dc6-e7ef-4cc0-92ab-c81e8df169f4", 00:14:32.902 "assigned_rate_limits": { 00:14:32.902 "rw_ios_per_sec": 0, 00:14:32.902 "rw_mbytes_per_sec": 0, 00:14:32.902 "r_mbytes_per_sec": 0, 00:14:32.902 "w_mbytes_per_sec": 0 00:14:32.902 }, 00:14:32.902 "claimed": false, 00:14:32.902 "zoned": false, 00:14:32.902 "supported_io_types": { 00:14:32.902 "read": true, 00:14:32.902 "write": true, 00:14:32.902 "unmap": true, 00:14:32.902 "flush": true, 00:14:32.902 "reset": true, 00:14:32.902 "nvme_admin": false, 00:14:32.903 "nvme_io": false, 00:14:32.903 "nvme_io_md": false, 00:14:32.903 "write_zeroes": true, 00:14:32.903 "zcopy": false, 00:14:32.903 "get_zone_info": false, 00:14:32.903 "zone_management": false, 00:14:32.903 "zone_append": false, 00:14:32.903 "compare": false, 00:14:32.903 "compare_and_write": false, 00:14:32.903 "abort": false, 00:14:32.903 "seek_hole": false, 00:14:32.903 "seek_data": false, 00:14:32.903 "copy": false, 00:14:32.903 "nvme_iov_md": false 00:14:32.903 }, 00:14:32.903 "memory_domains": [ 00:14:32.903 { 00:14:32.903 "dma_device_id": "system", 00:14:32.903 "dma_device_type": 1 00:14:32.903 }, 00:14:32.903 { 00:14:32.903 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:32.903 "dma_device_type": 2 00:14:32.903 }, 00:14:32.903 { 00:14:32.903 "dma_device_id": "system", 00:14:32.903 "dma_device_type": 1 00:14:32.903 }, 00:14:32.903 { 00:14:32.903 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:32.903 "dma_device_type": 2 00:14:32.903 }, 00:14:32.903 { 00:14:32.903 "dma_device_id": "system", 00:14:32.903 "dma_device_type": 1 00:14:32.903 }, 00:14:32.903 { 00:14:32.903 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:32.903 "dma_device_type": 2 00:14:32.903 } 00:14:32.903 ], 00:14:32.903 "driver_specific": { 00:14:32.903 "raid": { 00:14:32.903 "uuid": "a29d8dc6-e7ef-4cc0-92ab-c81e8df169f4", 00:14:32.903 "strip_size_kb": 64, 00:14:32.903 "state": "online", 00:14:32.903 "raid_level": "raid0", 00:14:32.903 "superblock": true, 00:14:32.903 "num_base_bdevs": 3, 00:14:32.903 "num_base_bdevs_discovered": 3, 00:14:32.903 "num_base_bdevs_operational": 3, 00:14:32.903 "base_bdevs_list": [ 00:14:32.903 { 00:14:32.903 "name": "pt1", 00:14:32.903 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:32.903 "is_configured": true, 00:14:32.903 "data_offset": 2048, 00:14:32.903 "data_size": 63488 00:14:32.903 }, 00:14:32.903 { 00:14:32.903 "name": "pt2", 00:14:32.903 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:32.903 "is_configured": true, 00:14:32.903 "data_offset": 2048, 00:14:32.903 "data_size": 63488 00:14:32.903 }, 00:14:32.903 { 00:14:32.903 "name": "pt3", 00:14:32.903 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:32.903 "is_configured": true, 00:14:32.903 "data_offset": 2048, 00:14:32.903 "data_size": 63488 00:14:32.903 } 00:14:32.903 ] 00:14:32.903 } 00:14:32.903 } 00:14:32.903 }' 00:14:32.903 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:32.903 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:14:32.903 pt2 00:14:32.903 pt3' 00:14:32.903 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:14:32.903 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:14:32.903 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:14:33.161 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:14:33.161 "name": "pt1", 00:14:33.161 "aliases": [ 00:14:33.161 "00000000-0000-0000-0000-000000000001" 00:14:33.161 ], 00:14:33.161 "product_name": "passthru", 00:14:33.161 "block_size": 512, 00:14:33.161 "num_blocks": 65536, 00:14:33.161 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:33.161 "assigned_rate_limits": { 00:14:33.161 "rw_ios_per_sec": 0, 00:14:33.161 "rw_mbytes_per_sec": 0, 00:14:33.161 "r_mbytes_per_sec": 0, 00:14:33.161 "w_mbytes_per_sec": 0 00:14:33.161 }, 00:14:33.161 "claimed": true, 00:14:33.161 "claim_type": "exclusive_write", 00:14:33.161 "zoned": false, 00:14:33.161 "supported_io_types": { 00:14:33.161 "read": true, 00:14:33.161 "write": true, 00:14:33.161 "unmap": true, 00:14:33.161 "flush": true, 00:14:33.161 "reset": true, 00:14:33.161 "nvme_admin": false, 00:14:33.161 "nvme_io": false, 00:14:33.161 "nvme_io_md": false, 00:14:33.161 "write_zeroes": true, 00:14:33.161 "zcopy": true, 00:14:33.161 "get_zone_info": false, 00:14:33.161 "zone_management": false, 00:14:33.161 "zone_append": false, 00:14:33.161 "compare": false, 00:14:33.161 "compare_and_write": false, 00:14:33.161 "abort": true, 00:14:33.161 "seek_hole": false, 00:14:33.161 "seek_data": false, 00:14:33.161 "copy": true, 00:14:33.161 "nvme_iov_md": false 00:14:33.161 }, 00:14:33.161 "memory_domains": [ 00:14:33.161 { 00:14:33.161 "dma_device_id": "system", 00:14:33.161 "dma_device_type": 1 00:14:33.161 }, 00:14:33.161 { 00:14:33.161 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:33.161 "dma_device_type": 2 00:14:33.161 } 00:14:33.161 ], 00:14:33.161 "driver_specific": { 00:14:33.161 "passthru": { 00:14:33.161 "name": "pt1", 00:14:33.161 "base_bdev_name": "malloc1" 00:14:33.161 } 00:14:33.161 } 00:14:33.161 }' 00:14:33.161 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:33.161 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:33.161 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:14:33.161 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:33.162 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:33.420 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:14:33.420 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:33.420 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:33.420 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:14:33.420 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:33.420 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:33.420 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:14:33.420 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:14:33.420 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:14:33.420 11:25:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:14:33.678 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:14:33.678 "name": "pt2", 00:14:33.678 "aliases": [ 00:14:33.678 "00000000-0000-0000-0000-000000000002" 00:14:33.678 ], 00:14:33.678 "product_name": "passthru", 00:14:33.678 "block_size": 512, 00:14:33.678 "num_blocks": 65536, 00:14:33.678 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:33.678 "assigned_rate_limits": { 00:14:33.678 "rw_ios_per_sec": 0, 00:14:33.678 "rw_mbytes_per_sec": 0, 00:14:33.678 "r_mbytes_per_sec": 0, 00:14:33.678 "w_mbytes_per_sec": 0 00:14:33.678 }, 00:14:33.678 "claimed": true, 00:14:33.678 "claim_type": "exclusive_write", 00:14:33.678 "zoned": false, 00:14:33.678 "supported_io_types": { 00:14:33.678 "read": true, 00:14:33.678 "write": true, 00:14:33.678 "unmap": true, 00:14:33.678 "flush": true, 00:14:33.678 "reset": true, 00:14:33.678 "nvme_admin": false, 00:14:33.678 "nvme_io": false, 00:14:33.678 "nvme_io_md": false, 00:14:33.678 "write_zeroes": true, 00:14:33.678 "zcopy": true, 00:14:33.678 "get_zone_info": false, 00:14:33.678 "zone_management": false, 00:14:33.678 "zone_append": false, 00:14:33.678 "compare": false, 00:14:33.678 "compare_and_write": false, 00:14:33.678 "abort": true, 00:14:33.678 "seek_hole": false, 00:14:33.678 "seek_data": false, 00:14:33.678 "copy": true, 00:14:33.678 "nvme_iov_md": false 00:14:33.678 }, 00:14:33.678 "memory_domains": [ 00:14:33.678 { 00:14:33.678 "dma_device_id": "system", 00:14:33.678 "dma_device_type": 1 00:14:33.678 }, 00:14:33.678 { 00:14:33.678 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:33.678 "dma_device_type": 2 00:14:33.678 } 00:14:33.678 ], 00:14:33.678 "driver_specific": { 00:14:33.678 "passthru": { 00:14:33.678 "name": "pt2", 00:14:33.678 "base_bdev_name": "malloc2" 00:14:33.678 } 00:14:33.678 } 00:14:33.678 }' 00:14:33.678 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:33.678 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:33.935 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:14:33.935 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:33.935 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:33.935 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:14:33.935 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:33.935 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:33.935 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:14:33.936 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:33.936 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:34.193 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:14:34.193 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:14:34.193 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt3 00:14:34.193 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:14:34.451 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:14:34.451 "name": "pt3", 00:14:34.451 "aliases": [ 00:14:34.451 "00000000-0000-0000-0000-000000000003" 00:14:34.451 ], 00:14:34.451 "product_name": "passthru", 00:14:34.451 "block_size": 512, 00:14:34.451 "num_blocks": 65536, 00:14:34.451 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:34.451 "assigned_rate_limits": { 00:14:34.451 "rw_ios_per_sec": 0, 00:14:34.451 "rw_mbytes_per_sec": 0, 00:14:34.451 "r_mbytes_per_sec": 0, 00:14:34.451 "w_mbytes_per_sec": 0 00:14:34.451 }, 00:14:34.451 "claimed": true, 00:14:34.451 "claim_type": "exclusive_write", 00:14:34.451 "zoned": false, 00:14:34.451 "supported_io_types": { 00:14:34.451 "read": true, 00:14:34.451 "write": true, 00:14:34.451 "unmap": true, 00:14:34.451 "flush": true, 00:14:34.451 "reset": true, 00:14:34.451 "nvme_admin": false, 00:14:34.451 "nvme_io": false, 00:14:34.451 "nvme_io_md": false, 00:14:34.451 "write_zeroes": true, 00:14:34.451 "zcopy": true, 00:14:34.451 "get_zone_info": false, 00:14:34.451 "zone_management": false, 00:14:34.451 "zone_append": false, 00:14:34.451 "compare": false, 00:14:34.451 "compare_and_write": false, 00:14:34.451 "abort": true, 00:14:34.451 "seek_hole": false, 00:14:34.451 "seek_data": false, 00:14:34.451 "copy": true, 00:14:34.451 "nvme_iov_md": false 00:14:34.451 }, 00:14:34.451 "memory_domains": [ 00:14:34.451 { 00:14:34.451 "dma_device_id": "system", 00:14:34.451 "dma_device_type": 1 00:14:34.451 }, 00:14:34.451 { 00:14:34.451 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:34.451 "dma_device_type": 2 00:14:34.451 } 00:14:34.451 ], 00:14:34.451 "driver_specific": { 00:14:34.451 "passthru": { 00:14:34.451 "name": "pt3", 00:14:34.451 "base_bdev_name": "malloc3" 00:14:34.451 } 00:14:34.451 } 00:14:34.451 }' 00:14:34.451 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:34.451 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:34.451 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:14:34.451 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:34.451 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:34.451 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:14:34.451 11:25:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:34.451 11:25:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:34.451 11:25:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:14:34.708 11:25:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:34.709 11:25:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:34.709 11:25:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:14:34.709 11:25:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:14:34.709 11:25:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # jq -r '.[] | .uuid' 00:14:34.966 [2024-07-15 11:25:18.352056] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:34.966 11:25:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # '[' a29d8dc6-e7ef-4cc0-92ab-c81e8df169f4 '!=' a29d8dc6-e7ef-4cc0-92ab-c81e8df169f4 ']' 00:14:34.966 11:25:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@490 -- # has_redundancy raid0 00:14:34.966 11:25:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:14:34.966 11:25:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@215 -- # return 1 00:14:34.966 11:25:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@562 -- # killprocess 891380 00:14:34.966 11:25:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@948 -- # '[' -z 891380 ']' 00:14:34.966 11:25:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@952 -- # kill -0 891380 00:14:34.966 11:25:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@953 -- # uname 00:14:34.966 11:25:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:14:34.966 11:25:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 891380 00:14:34.966 11:25:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:14:34.966 11:25:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:14:34.966 11:25:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 891380' 00:14:34.966 killing process with pid 891380 00:14:34.966 11:25:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@967 -- # kill 891380 00:14:34.966 [2024-07-15 11:25:18.423999] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:34.966 [2024-07-15 11:25:18.424052] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:34.966 [2024-07-15 11:25:18.424104] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:34.966 [2024-07-15 11:25:18.424116] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x164cc00 name raid_bdev1, state offline 00:14:34.966 11:25:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # wait 891380 00:14:34.966 [2024-07-15 11:25:18.455030] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:35.222 11:25:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@564 -- # return 0 00:14:35.222 00:14:35.222 real 0m14.017s 00:14:35.222 user 0m25.157s 00:14:35.222 sys 0m2.607s 00:14:35.222 11:25:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:14:35.222 11:25:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.222 ************************************ 00:14:35.222 END TEST raid_superblock_test 00:14:35.222 ************************************ 00:14:35.222 11:25:18 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:14:35.222 11:25:18 bdev_raid -- bdev/bdev_raid.sh@870 -- # run_test raid_read_error_test raid_io_error_test raid0 3 read 00:14:35.222 11:25:18 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:14:35.222 11:25:18 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:14:35.222 11:25:18 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:35.222 ************************************ 00:14:35.222 START TEST raid_read_error_test 00:14:35.222 ************************************ 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1123 -- # raid_io_error_test raid0 3 read 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@788 -- # local raid_level=raid0 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@789 -- # local num_base_bdevs=3 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local error_io_type=read 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i = 1 )) 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev1 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev2 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev3 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local base_bdevs 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local raid_bdev_name=raid_bdev1 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local strip_size 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local create_arg 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local bdevperf_log 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local fail_per_s 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # '[' raid0 '!=' raid1 ']' 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@799 -- # strip_size=64 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # create_arg+=' -z 64' 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@805 -- # mktemp -p /raidtest 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@805 -- # bdevperf_log=/raidtest/tmp.R4qeT5E5TW 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@808 -- # raid_pid=893564 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # waitforlisten 893564 /var/tmp/spdk-raid.sock 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@829 -- # '[' -z 893564 ']' 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:14:35.222 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:14:35.222 11:25:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.479 [2024-07-15 11:25:18.824884] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:14:35.479 [2024-07-15 11:25:18.824958] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid893564 ] 00:14:35.479 [2024-07-15 11:25:18.952008] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:35.479 [2024-07-15 11:25:19.057729] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:14:35.736 [2024-07-15 11:25:19.126311] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:35.736 [2024-07-15 11:25:19.126348] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:36.302 11:25:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:14:36.302 11:25:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@862 -- # return 0 00:14:36.302 11:25:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:14:36.302 11:25:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:36.559 BaseBdev1_malloc 00:14:36.560 11:25:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev1_malloc 00:14:36.818 true 00:14:36.818 11:25:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:14:37.077 [2024-07-15 11:25:20.474456] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:14:37.077 [2024-07-15 11:25:20.474503] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:37.077 [2024-07-15 11:25:20.474527] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xdcb0d0 00:14:37.077 [2024-07-15 11:25:20.474540] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:37.077 [2024-07-15 11:25:20.476446] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:37.077 [2024-07-15 11:25:20.476478] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:37.077 BaseBdev1 00:14:37.077 11:25:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:14:37.077 11:25:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:37.335 BaseBdev2_malloc 00:14:37.335 11:25:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev2_malloc 00:14:37.592 true 00:14:37.592 11:25:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:14:37.851 [2024-07-15 11:25:21.189694] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:14:37.851 [2024-07-15 11:25:21.189739] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:37.851 [2024-07-15 11:25:21.189762] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xdcf910 00:14:37.851 [2024-07-15 11:25:21.189775] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:37.851 [2024-07-15 11:25:21.191358] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:37.851 [2024-07-15 11:25:21.191389] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:37.851 BaseBdev2 00:14:37.851 11:25:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:14:37.851 11:25:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:37.851 BaseBdev3_malloc 00:14:38.110 11:25:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev3_malloc 00:14:38.110 true 00:14:38.110 11:25:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:14:38.369 [2024-07-15 11:25:21.913436] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:14:38.369 [2024-07-15 11:25:21.913482] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:38.369 [2024-07-15 11:25:21.913505] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xdd1bd0 00:14:38.369 [2024-07-15 11:25:21.913518] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:38.369 [2024-07-15 11:25:21.915123] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:38.369 [2024-07-15 11:25:21.915152] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:38.369 BaseBdev3 00:14:38.369 11:25:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@819 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n raid_bdev1 -s 00:14:38.628 [2024-07-15 11:25:22.150098] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:38.628 [2024-07-15 11:25:22.151467] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:38.628 [2024-07-15 11:25:22.151535] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:38.628 [2024-07-15 11:25:22.151745] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xdd3280 00:14:38.628 [2024-07-15 11:25:22.151757] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:38.628 [2024-07-15 11:25:22.151969] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xdd2e20 00:14:38.628 [2024-07-15 11:25:22.152116] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xdd3280 00:14:38.628 [2024-07-15 11:25:22.152126] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0xdd3280 00:14:38.628 [2024-07-15 11:25:22.152230] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:38.628 11:25:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@820 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:14:38.628 11:25:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:14:38.628 11:25:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:14:38.628 11:25:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:14:38.628 11:25:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:38.628 11:25:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:38.628 11:25:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:38.628 11:25:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:38.628 11:25:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:38.628 11:25:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:38.628 11:25:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:38.628 11:25:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:38.887 11:25:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:38.887 "name": "raid_bdev1", 00:14:38.887 "uuid": "a845e186-7746-4cfe-bdee-10c56080f71d", 00:14:38.887 "strip_size_kb": 64, 00:14:38.887 "state": "online", 00:14:38.887 "raid_level": "raid0", 00:14:38.887 "superblock": true, 00:14:38.887 "num_base_bdevs": 3, 00:14:38.887 "num_base_bdevs_discovered": 3, 00:14:38.887 "num_base_bdevs_operational": 3, 00:14:38.887 "base_bdevs_list": [ 00:14:38.887 { 00:14:38.887 "name": "BaseBdev1", 00:14:38.887 "uuid": "a823762d-f2a3-5813-a1f9-7860b9a073a7", 00:14:38.887 "is_configured": true, 00:14:38.887 "data_offset": 2048, 00:14:38.887 "data_size": 63488 00:14:38.887 }, 00:14:38.887 { 00:14:38.887 "name": "BaseBdev2", 00:14:38.887 "uuid": "59221401-a1d6-5b67-ba07-955bab3fe21f", 00:14:38.887 "is_configured": true, 00:14:38.887 "data_offset": 2048, 00:14:38.887 "data_size": 63488 00:14:38.887 }, 00:14:38.887 { 00:14:38.887 "name": "BaseBdev3", 00:14:38.887 "uuid": "1deab32a-1c46-5f0f-b33a-f7bccb07418b", 00:14:38.887 "is_configured": true, 00:14:38.887 "data_offset": 2048, 00:14:38.887 "data_size": 63488 00:14:38.887 } 00:14:38.887 ] 00:14:38.887 }' 00:14:38.887 11:25:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:38.887 11:25:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.454 11:25:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@824 -- # sleep 1 00:14:39.454 11:25:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@823 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:14:39.712 [2024-07-15 11:25:23.092863] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xc215b0 00:14:40.645 11:25:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@827 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:14:40.903 11:25:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # local expected_num_base_bdevs 00:14:40.903 11:25:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@830 -- # [[ raid0 = \r\a\i\d\1 ]] 00:14:40.903 11:25:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=3 00:14:40.903 11:25:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:14:40.903 11:25:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:14:40.903 11:25:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:14:40.903 11:25:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:14:40.903 11:25:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:40.903 11:25:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:40.903 11:25:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:40.903 11:25:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:40.903 11:25:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:40.903 11:25:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:40.903 11:25:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:40.903 11:25:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:40.903 11:25:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:40.903 "name": "raid_bdev1", 00:14:40.903 "uuid": "a845e186-7746-4cfe-bdee-10c56080f71d", 00:14:40.903 "strip_size_kb": 64, 00:14:40.903 "state": "online", 00:14:40.903 "raid_level": "raid0", 00:14:40.903 "superblock": true, 00:14:40.903 "num_base_bdevs": 3, 00:14:40.903 "num_base_bdevs_discovered": 3, 00:14:40.903 "num_base_bdevs_operational": 3, 00:14:40.903 "base_bdevs_list": [ 00:14:40.903 { 00:14:40.903 "name": "BaseBdev1", 00:14:40.903 "uuid": "a823762d-f2a3-5813-a1f9-7860b9a073a7", 00:14:40.903 "is_configured": true, 00:14:40.903 "data_offset": 2048, 00:14:40.903 "data_size": 63488 00:14:40.903 }, 00:14:40.903 { 00:14:40.903 "name": "BaseBdev2", 00:14:40.903 "uuid": "59221401-a1d6-5b67-ba07-955bab3fe21f", 00:14:40.903 "is_configured": true, 00:14:40.903 "data_offset": 2048, 00:14:40.903 "data_size": 63488 00:14:40.903 }, 00:14:40.903 { 00:14:40.903 "name": "BaseBdev3", 00:14:40.903 "uuid": "1deab32a-1c46-5f0f-b33a-f7bccb07418b", 00:14:40.903 "is_configured": true, 00:14:40.903 "data_offset": 2048, 00:14:40.903 "data_size": 63488 00:14:40.903 } 00:14:40.903 ] 00:14:40.903 }' 00:14:40.903 11:25:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:40.903 11:25:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.505 11:25:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:14:41.786 [2024-07-15 11:25:25.241166] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:41.786 [2024-07-15 11:25:25.241191] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:41.786 [2024-07-15 11:25:25.244350] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:41.786 [2024-07-15 11:25:25.244386] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:41.786 [2024-07-15 11:25:25.244427] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:41.786 [2024-07-15 11:25:25.244438] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xdd3280 name raid_bdev1, state offline 00:14:41.786 0 00:14:41.786 11:25:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # killprocess 893564 00:14:41.786 11:25:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@948 -- # '[' -z 893564 ']' 00:14:41.786 11:25:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@952 -- # kill -0 893564 00:14:41.786 11:25:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@953 -- # uname 00:14:41.786 11:25:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:14:41.786 11:25:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 893564 00:14:41.786 11:25:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:14:41.786 11:25:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:14:41.786 11:25:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 893564' 00:14:41.786 killing process with pid 893564 00:14:41.786 11:25:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@967 -- # kill 893564 00:14:41.786 [2024-07-15 11:25:25.307498] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:41.786 11:25:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # wait 893564 00:14:41.786 [2024-07-15 11:25:25.328315] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:42.044 11:25:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # grep -v Job /raidtest/tmp.R4qeT5E5TW 00:14:42.044 11:25:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # grep raid_bdev1 00:14:42.044 11:25:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # awk '{print $6}' 00:14:42.044 11:25:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # fail_per_s=0.47 00:14:42.044 11:25:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@844 -- # has_redundancy raid0 00:14:42.044 11:25:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:14:42.044 11:25:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@215 -- # return 1 00:14:42.044 11:25:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.47 != \0\.\0\0 ]] 00:14:42.044 00:14:42.044 real 0m6.818s 00:14:42.044 user 0m10.746s 00:14:42.044 sys 0m1.204s 00:14:42.044 11:25:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:14:42.044 11:25:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.044 ************************************ 00:14:42.044 END TEST raid_read_error_test 00:14:42.044 ************************************ 00:14:42.044 11:25:25 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:14:42.044 11:25:25 bdev_raid -- bdev/bdev_raid.sh@871 -- # run_test raid_write_error_test raid_io_error_test raid0 3 write 00:14:42.044 11:25:25 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:14:42.044 11:25:25 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:14:42.044 11:25:25 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:42.303 ************************************ 00:14:42.303 START TEST raid_write_error_test 00:14:42.303 ************************************ 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1123 -- # raid_io_error_test raid0 3 write 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@788 -- # local raid_level=raid0 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@789 -- # local num_base_bdevs=3 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local error_io_type=write 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i = 1 )) 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev1 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev2 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev3 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local base_bdevs 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local raid_bdev_name=raid_bdev1 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local strip_size 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local create_arg 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local bdevperf_log 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local fail_per_s 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # '[' raid0 '!=' raid1 ']' 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@799 -- # strip_size=64 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # create_arg+=' -z 64' 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@805 -- # mktemp -p /raidtest 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@805 -- # bdevperf_log=/raidtest/tmp.4LXZCVmJBv 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@808 -- # raid_pid=894576 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # waitforlisten 894576 /var/tmp/spdk-raid.sock 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@829 -- # '[' -z 894576 ']' 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:14:42.303 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:14:42.303 11:25:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.303 [2024-07-15 11:25:25.726786] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:14:42.303 [2024-07-15 11:25:25.726843] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid894576 ] 00:14:42.303 [2024-07-15 11:25:25.840596] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:42.561 [2024-07-15 11:25:25.939408] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:14:42.561 [2024-07-15 11:25:26.003402] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:42.561 [2024-07-15 11:25:26.003449] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:43.128 11:25:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:14:43.128 11:25:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@862 -- # return 0 00:14:43.128 11:25:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:14:43.128 11:25:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:43.386 BaseBdev1_malloc 00:14:43.386 11:25:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev1_malloc 00:14:43.644 true 00:14:43.644 11:25:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:14:43.902 [2024-07-15 11:25:27.385605] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:14:43.902 [2024-07-15 11:25:27.385654] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:43.902 [2024-07-15 11:25:27.385677] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x265c0d0 00:14:43.902 [2024-07-15 11:25:27.385690] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:43.902 [2024-07-15 11:25:27.387636] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:43.902 [2024-07-15 11:25:27.387671] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:43.902 BaseBdev1 00:14:43.902 11:25:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:14:43.902 11:25:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:44.161 BaseBdev2_malloc 00:14:44.161 11:25:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev2_malloc 00:14:44.419 true 00:14:44.419 11:25:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:14:44.678 [2024-07-15 11:25:28.109398] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:14:44.678 [2024-07-15 11:25:28.109452] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:44.678 [2024-07-15 11:25:28.109473] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2660910 00:14:44.678 [2024-07-15 11:25:28.109486] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:44.678 [2024-07-15 11:25:28.111138] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:44.678 [2024-07-15 11:25:28.111172] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:44.678 BaseBdev2 00:14:44.678 11:25:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:14:44.678 11:25:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:44.937 BaseBdev3_malloc 00:14:44.937 11:25:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev3_malloc 00:14:45.196 true 00:14:45.196 11:25:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:14:45.455 [2024-07-15 11:25:28.848185] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:14:45.455 [2024-07-15 11:25:28.848235] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:45.455 [2024-07-15 11:25:28.848257] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2662bd0 00:14:45.455 [2024-07-15 11:25:28.848269] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:45.455 [2024-07-15 11:25:28.849936] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:45.455 [2024-07-15 11:25:28.849969] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:45.455 BaseBdev3 00:14:45.455 11:25:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@819 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n raid_bdev1 -s 00:14:45.713 [2024-07-15 11:25:29.088856] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:45.713 [2024-07-15 11:25:29.090244] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:45.713 [2024-07-15 11:25:29.090315] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:45.713 [2024-07-15 11:25:29.090530] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x2664280 00:14:45.713 [2024-07-15 11:25:29.090541] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:45.713 [2024-07-15 11:25:29.090753] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x2663e20 00:14:45.713 [2024-07-15 11:25:29.090906] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x2664280 00:14:45.713 [2024-07-15 11:25:29.090916] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x2664280 00:14:45.713 [2024-07-15 11:25:29.091037] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:45.713 11:25:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@820 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:14:45.713 11:25:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:14:45.713 11:25:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:14:45.713 11:25:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:14:45.713 11:25:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:45.713 11:25:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:45.713 11:25:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:45.713 11:25:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:45.713 11:25:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:45.713 11:25:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:45.714 11:25:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:45.714 11:25:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:45.972 11:25:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:45.972 "name": "raid_bdev1", 00:14:45.972 "uuid": "805ef99d-cc3d-4e5a-8704-fc9a6de40155", 00:14:45.972 "strip_size_kb": 64, 00:14:45.972 "state": "online", 00:14:45.972 "raid_level": "raid0", 00:14:45.972 "superblock": true, 00:14:45.972 "num_base_bdevs": 3, 00:14:45.972 "num_base_bdevs_discovered": 3, 00:14:45.972 "num_base_bdevs_operational": 3, 00:14:45.972 "base_bdevs_list": [ 00:14:45.972 { 00:14:45.972 "name": "BaseBdev1", 00:14:45.972 "uuid": "70ebc494-dcd8-5e73-89b1-9d0bd52178b5", 00:14:45.972 "is_configured": true, 00:14:45.972 "data_offset": 2048, 00:14:45.972 "data_size": 63488 00:14:45.972 }, 00:14:45.972 { 00:14:45.972 "name": "BaseBdev2", 00:14:45.972 "uuid": "8321098d-d374-52d5-92cd-93839eafe671", 00:14:45.972 "is_configured": true, 00:14:45.972 "data_offset": 2048, 00:14:45.972 "data_size": 63488 00:14:45.972 }, 00:14:45.972 { 00:14:45.972 "name": "BaseBdev3", 00:14:45.972 "uuid": "4e28841f-8657-520b-ac8c-5ac96fb218a7", 00:14:45.972 "is_configured": true, 00:14:45.972 "data_offset": 2048, 00:14:45.972 "data_size": 63488 00:14:45.972 } 00:14:45.972 ] 00:14:45.972 }' 00:14:45.972 11:25:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:45.972 11:25:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.539 11:25:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@824 -- # sleep 1 00:14:46.539 11:25:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@823 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:14:46.539 [2024-07-15 11:25:30.051684] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x24b25b0 00:14:47.475 11:25:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@827 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:14:47.734 11:25:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # local expected_num_base_bdevs 00:14:47.734 11:25:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@830 -- # [[ raid0 = \r\a\i\d\1 ]] 00:14:47.734 11:25:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=3 00:14:47.734 11:25:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:14:47.734 11:25:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:14:47.734 11:25:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:14:47.734 11:25:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:14:47.734 11:25:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:47.734 11:25:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:47.734 11:25:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:47.734 11:25:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:47.734 11:25:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:47.734 11:25:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:47.734 11:25:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:47.734 11:25:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:47.993 11:25:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:47.994 "name": "raid_bdev1", 00:14:47.994 "uuid": "805ef99d-cc3d-4e5a-8704-fc9a6de40155", 00:14:47.994 "strip_size_kb": 64, 00:14:47.994 "state": "online", 00:14:47.994 "raid_level": "raid0", 00:14:47.994 "superblock": true, 00:14:47.994 "num_base_bdevs": 3, 00:14:47.994 "num_base_bdevs_discovered": 3, 00:14:47.994 "num_base_bdevs_operational": 3, 00:14:47.994 "base_bdevs_list": [ 00:14:47.994 { 00:14:47.994 "name": "BaseBdev1", 00:14:47.994 "uuid": "70ebc494-dcd8-5e73-89b1-9d0bd52178b5", 00:14:47.994 "is_configured": true, 00:14:47.994 "data_offset": 2048, 00:14:47.994 "data_size": 63488 00:14:47.994 }, 00:14:47.994 { 00:14:47.994 "name": "BaseBdev2", 00:14:47.994 "uuid": "8321098d-d374-52d5-92cd-93839eafe671", 00:14:47.994 "is_configured": true, 00:14:47.994 "data_offset": 2048, 00:14:47.994 "data_size": 63488 00:14:47.994 }, 00:14:47.994 { 00:14:47.994 "name": "BaseBdev3", 00:14:47.994 "uuid": "4e28841f-8657-520b-ac8c-5ac96fb218a7", 00:14:47.994 "is_configured": true, 00:14:47.994 "data_offset": 2048, 00:14:47.994 "data_size": 63488 00:14:47.994 } 00:14:47.994 ] 00:14:47.994 }' 00:14:47.994 11:25:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:47.994 11:25:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.562 11:25:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:14:48.820 [2024-07-15 11:25:32.269189] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:48.820 [2024-07-15 11:25:32.269235] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:48.820 [2024-07-15 11:25:32.272411] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:48.820 [2024-07-15 11:25:32.272450] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:48.820 [2024-07-15 11:25:32.272486] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:48.820 [2024-07-15 11:25:32.272499] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x2664280 name raid_bdev1, state offline 00:14:48.820 0 00:14:48.820 11:25:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # killprocess 894576 00:14:48.820 11:25:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@948 -- # '[' -z 894576 ']' 00:14:48.820 11:25:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@952 -- # kill -0 894576 00:14:48.820 11:25:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@953 -- # uname 00:14:48.820 11:25:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:14:48.820 11:25:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 894576 00:14:48.820 11:25:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:14:48.820 11:25:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:14:48.820 11:25:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 894576' 00:14:48.820 killing process with pid 894576 00:14:48.820 11:25:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@967 -- # kill 894576 00:14:48.820 [2024-07-15 11:25:32.337512] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:48.820 11:25:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # wait 894576 00:14:48.820 [2024-07-15 11:25:32.358477] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:49.077 11:25:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # grep -v Job /raidtest/tmp.4LXZCVmJBv 00:14:49.077 11:25:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # grep raid_bdev1 00:14:49.077 11:25:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # awk '{print $6}' 00:14:49.077 11:25:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # fail_per_s=0.45 00:14:49.077 11:25:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@844 -- # has_redundancy raid0 00:14:49.077 11:25:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:14:49.077 11:25:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@215 -- # return 1 00:14:49.077 11:25:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.45 != \0\.\0\0 ]] 00:14:49.077 00:14:49.077 real 0m6.945s 00:14:49.077 user 0m11.013s 00:14:49.077 sys 0m1.207s 00:14:49.077 11:25:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:14:49.077 11:25:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.077 ************************************ 00:14:49.077 END TEST raid_write_error_test 00:14:49.077 ************************************ 00:14:49.077 11:25:32 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:14:49.077 11:25:32 bdev_raid -- bdev/bdev_raid.sh@866 -- # for level in raid0 concat raid1 00:14:49.077 11:25:32 bdev_raid -- bdev/bdev_raid.sh@867 -- # run_test raid_state_function_test raid_state_function_test concat 3 false 00:14:49.077 11:25:32 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:14:49.077 11:25:32 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:14:49.077 11:25:32 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:49.334 ************************************ 00:14:49.334 START TEST raid_state_function_test 00:14:49.334 ************************************ 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1123 -- # raid_state_function_test concat 3 false 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@220 -- # local raid_level=concat 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=3 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # local superblock=false 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev3 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # local strip_size 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # '[' concat '!=' raid1 ']' 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # strip_size=64 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@232 -- # strip_size_create_arg='-z 64' 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # '[' false = true ']' 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@240 -- # superblock_create_arg= 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # raid_pid=895556 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 895556' 00:14:49.334 Process raid pid: 895556 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@246 -- # waitforlisten 895556 /var/tmp/spdk-raid.sock 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@829 -- # '[' -z 895556 ']' 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:14:49.334 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:14:49.334 11:25:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.334 [2024-07-15 11:25:32.751884] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:14:49.334 [2024-07-15 11:25:32.751956] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:49.334 [2024-07-15 11:25:32.883260] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:49.592 [2024-07-15 11:25:32.990305] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:14:49.592 [2024-07-15 11:25:33.055400] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:49.592 [2024-07-15 11:25:33.055425] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:50.156 11:25:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:14:50.156 11:25:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@862 -- # return 0 00:14:50.156 11:25:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:14:50.413 [2024-07-15 11:25:33.905977] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:50.413 [2024-07-15 11:25:33.906020] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:50.413 [2024-07-15 11:25:33.906031] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:50.413 [2024-07-15 11:25:33.906042] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:50.413 [2024-07-15 11:25:33.906051] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:50.413 [2024-07-15 11:25:33.906067] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:50.413 11:25:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:14:50.413 11:25:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:14:50.413 11:25:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:14:50.413 11:25:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:14:50.413 11:25:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:50.413 11:25:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:50.413 11:25:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:50.413 11:25:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:50.413 11:25:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:50.413 11:25:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:50.414 11:25:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:50.414 11:25:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:50.671 11:25:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:50.671 "name": "Existed_Raid", 00:14:50.671 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.671 "strip_size_kb": 64, 00:14:50.671 "state": "configuring", 00:14:50.671 "raid_level": "concat", 00:14:50.671 "superblock": false, 00:14:50.671 "num_base_bdevs": 3, 00:14:50.671 "num_base_bdevs_discovered": 0, 00:14:50.671 "num_base_bdevs_operational": 3, 00:14:50.671 "base_bdevs_list": [ 00:14:50.671 { 00:14:50.671 "name": "BaseBdev1", 00:14:50.671 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.671 "is_configured": false, 00:14:50.671 "data_offset": 0, 00:14:50.671 "data_size": 0 00:14:50.671 }, 00:14:50.671 { 00:14:50.671 "name": "BaseBdev2", 00:14:50.671 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.671 "is_configured": false, 00:14:50.671 "data_offset": 0, 00:14:50.671 "data_size": 0 00:14:50.671 }, 00:14:50.671 { 00:14:50.671 "name": "BaseBdev3", 00:14:50.671 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.671 "is_configured": false, 00:14:50.671 "data_offset": 0, 00:14:50.671 "data_size": 0 00:14:50.671 } 00:14:50.671 ] 00:14:50.671 }' 00:14:50.671 11:25:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:50.671 11:25:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.235 11:25:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:14:51.799 [2024-07-15 11:25:35.237345] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:51.799 [2024-07-15 11:25:35.237376] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x10caa80 name Existed_Raid, state configuring 00:14:51.799 11:25:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:14:52.057 [2024-07-15 11:25:35.494043] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:52.057 [2024-07-15 11:25:35.494073] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:52.057 [2024-07-15 11:25:35.494082] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:52.057 [2024-07-15 11:25:35.494094] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:52.057 [2024-07-15 11:25:35.494102] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:52.057 [2024-07-15 11:25:35.494113] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:52.057 11:25:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:14:52.314 [2024-07-15 11:25:35.741725] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:52.314 BaseBdev1 00:14:52.314 11:25:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:14:52.314 11:25:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:14:52.314 11:25:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:14:52.314 11:25:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:14:52.314 11:25:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:14:52.314 11:25:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:14:52.314 11:25:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:14:52.572 11:25:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:52.830 [ 00:14:52.830 { 00:14:52.830 "name": "BaseBdev1", 00:14:52.830 "aliases": [ 00:14:52.830 "c0681378-efe7-4a97-b07c-dcbb46129f5a" 00:14:52.830 ], 00:14:52.830 "product_name": "Malloc disk", 00:14:52.830 "block_size": 512, 00:14:52.830 "num_blocks": 65536, 00:14:52.830 "uuid": "c0681378-efe7-4a97-b07c-dcbb46129f5a", 00:14:52.830 "assigned_rate_limits": { 00:14:52.830 "rw_ios_per_sec": 0, 00:14:52.830 "rw_mbytes_per_sec": 0, 00:14:52.830 "r_mbytes_per_sec": 0, 00:14:52.830 "w_mbytes_per_sec": 0 00:14:52.830 }, 00:14:52.830 "claimed": true, 00:14:52.830 "claim_type": "exclusive_write", 00:14:52.830 "zoned": false, 00:14:52.830 "supported_io_types": { 00:14:52.830 "read": true, 00:14:52.830 "write": true, 00:14:52.830 "unmap": true, 00:14:52.830 "flush": true, 00:14:52.830 "reset": true, 00:14:52.830 "nvme_admin": false, 00:14:52.830 "nvme_io": false, 00:14:52.830 "nvme_io_md": false, 00:14:52.830 "write_zeroes": true, 00:14:52.830 "zcopy": true, 00:14:52.830 "get_zone_info": false, 00:14:52.830 "zone_management": false, 00:14:52.830 "zone_append": false, 00:14:52.830 "compare": false, 00:14:52.830 "compare_and_write": false, 00:14:52.830 "abort": true, 00:14:52.830 "seek_hole": false, 00:14:52.830 "seek_data": false, 00:14:52.830 "copy": true, 00:14:52.830 "nvme_iov_md": false 00:14:52.830 }, 00:14:52.830 "memory_domains": [ 00:14:52.830 { 00:14:52.830 "dma_device_id": "system", 00:14:52.830 "dma_device_type": 1 00:14:52.830 }, 00:14:52.830 { 00:14:52.830 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:52.830 "dma_device_type": 2 00:14:52.830 } 00:14:52.830 ], 00:14:52.830 "driver_specific": {} 00:14:52.830 } 00:14:52.830 ] 00:14:52.830 11:25:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:14:52.830 11:25:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:14:52.830 11:25:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:14:52.830 11:25:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:14:52.830 11:25:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:14:52.830 11:25:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:52.830 11:25:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:52.830 11:25:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:52.830 11:25:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:52.830 11:25:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:52.830 11:25:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:52.830 11:25:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:52.830 11:25:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:53.086 11:25:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:53.086 "name": "Existed_Raid", 00:14:53.086 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:53.086 "strip_size_kb": 64, 00:14:53.086 "state": "configuring", 00:14:53.086 "raid_level": "concat", 00:14:53.086 "superblock": false, 00:14:53.086 "num_base_bdevs": 3, 00:14:53.086 "num_base_bdevs_discovered": 1, 00:14:53.086 "num_base_bdevs_operational": 3, 00:14:53.086 "base_bdevs_list": [ 00:14:53.086 { 00:14:53.086 "name": "BaseBdev1", 00:14:53.086 "uuid": "c0681378-efe7-4a97-b07c-dcbb46129f5a", 00:14:53.086 "is_configured": true, 00:14:53.086 "data_offset": 0, 00:14:53.086 "data_size": 65536 00:14:53.086 }, 00:14:53.086 { 00:14:53.086 "name": "BaseBdev2", 00:14:53.087 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:53.087 "is_configured": false, 00:14:53.087 "data_offset": 0, 00:14:53.087 "data_size": 0 00:14:53.087 }, 00:14:53.087 { 00:14:53.087 "name": "BaseBdev3", 00:14:53.087 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:53.087 "is_configured": false, 00:14:53.087 "data_offset": 0, 00:14:53.087 "data_size": 0 00:14:53.087 } 00:14:53.087 ] 00:14:53.087 }' 00:14:53.087 11:25:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:53.087 11:25:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.649 11:25:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:14:53.649 [2024-07-15 11:25:37.241720] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:53.649 [2024-07-15 11:25:37.241757] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x10ca310 name Existed_Raid, state configuring 00:14:53.907 11:25:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:14:53.907 [2024-07-15 11:25:37.414213] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:53.907 [2024-07-15 11:25:37.415639] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:53.907 [2024-07-15 11:25:37.415672] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:53.907 [2024-07-15 11:25:37.415682] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:53.907 [2024-07-15 11:25:37.415693] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:53.907 11:25:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:14:53.907 11:25:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:14:53.907 11:25:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:14:53.907 11:25:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:14:53.907 11:25:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:14:53.907 11:25:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:14:53.907 11:25:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:53.907 11:25:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:53.907 11:25:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:53.907 11:25:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:53.907 11:25:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:53.907 11:25:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:53.907 11:25:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:53.907 11:25:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:54.164 11:25:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:54.164 "name": "Existed_Raid", 00:14:54.164 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:54.164 "strip_size_kb": 64, 00:14:54.164 "state": "configuring", 00:14:54.164 "raid_level": "concat", 00:14:54.164 "superblock": false, 00:14:54.164 "num_base_bdevs": 3, 00:14:54.164 "num_base_bdevs_discovered": 1, 00:14:54.164 "num_base_bdevs_operational": 3, 00:14:54.164 "base_bdevs_list": [ 00:14:54.164 { 00:14:54.164 "name": "BaseBdev1", 00:14:54.164 "uuid": "c0681378-efe7-4a97-b07c-dcbb46129f5a", 00:14:54.164 "is_configured": true, 00:14:54.164 "data_offset": 0, 00:14:54.164 "data_size": 65536 00:14:54.164 }, 00:14:54.164 { 00:14:54.164 "name": "BaseBdev2", 00:14:54.164 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:54.164 "is_configured": false, 00:14:54.164 "data_offset": 0, 00:14:54.164 "data_size": 0 00:14:54.164 }, 00:14:54.164 { 00:14:54.164 "name": "BaseBdev3", 00:14:54.164 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:54.164 "is_configured": false, 00:14:54.164 "data_offset": 0, 00:14:54.164 "data_size": 0 00:14:54.164 } 00:14:54.164 ] 00:14:54.164 }' 00:14:54.164 11:25:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:54.164 11:25:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.727 11:25:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:14:54.983 [2024-07-15 11:25:38.452340] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:54.983 BaseBdev2 00:14:54.983 11:25:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:14:54.983 11:25:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:14:54.983 11:25:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:14:54.983 11:25:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:14:54.983 11:25:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:14:54.983 11:25:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:14:54.983 11:25:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:14:55.240 11:25:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:55.500 [ 00:14:55.500 { 00:14:55.500 "name": "BaseBdev2", 00:14:55.500 "aliases": [ 00:14:55.500 "f6085d77-0f5e-4695-aae0-1d0357e99da6" 00:14:55.500 ], 00:14:55.500 "product_name": "Malloc disk", 00:14:55.500 "block_size": 512, 00:14:55.500 "num_blocks": 65536, 00:14:55.500 "uuid": "f6085d77-0f5e-4695-aae0-1d0357e99da6", 00:14:55.500 "assigned_rate_limits": { 00:14:55.500 "rw_ios_per_sec": 0, 00:14:55.500 "rw_mbytes_per_sec": 0, 00:14:55.500 "r_mbytes_per_sec": 0, 00:14:55.500 "w_mbytes_per_sec": 0 00:14:55.500 }, 00:14:55.500 "claimed": true, 00:14:55.500 "claim_type": "exclusive_write", 00:14:55.500 "zoned": false, 00:14:55.500 "supported_io_types": { 00:14:55.500 "read": true, 00:14:55.500 "write": true, 00:14:55.500 "unmap": true, 00:14:55.500 "flush": true, 00:14:55.500 "reset": true, 00:14:55.500 "nvme_admin": false, 00:14:55.500 "nvme_io": false, 00:14:55.500 "nvme_io_md": false, 00:14:55.500 "write_zeroes": true, 00:14:55.500 "zcopy": true, 00:14:55.500 "get_zone_info": false, 00:14:55.500 "zone_management": false, 00:14:55.500 "zone_append": false, 00:14:55.500 "compare": false, 00:14:55.500 "compare_and_write": false, 00:14:55.500 "abort": true, 00:14:55.500 "seek_hole": false, 00:14:55.500 "seek_data": false, 00:14:55.500 "copy": true, 00:14:55.500 "nvme_iov_md": false 00:14:55.500 }, 00:14:55.500 "memory_domains": [ 00:14:55.500 { 00:14:55.500 "dma_device_id": "system", 00:14:55.500 "dma_device_type": 1 00:14:55.500 }, 00:14:55.500 { 00:14:55.500 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:55.500 "dma_device_type": 2 00:14:55.500 } 00:14:55.500 ], 00:14:55.500 "driver_specific": {} 00:14:55.500 } 00:14:55.500 ] 00:14:55.500 11:25:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:14:55.500 11:25:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:14:55.500 11:25:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:14:55.500 11:25:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:14:55.500 11:25:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:14:55.500 11:25:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:14:55.500 11:25:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:14:55.500 11:25:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:55.500 11:25:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:55.500 11:25:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:55.500 11:25:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:55.500 11:25:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:55.500 11:25:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:55.500 11:25:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:55.500 11:25:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:55.500 11:25:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:55.500 "name": "Existed_Raid", 00:14:55.500 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:55.500 "strip_size_kb": 64, 00:14:55.500 "state": "configuring", 00:14:55.500 "raid_level": "concat", 00:14:55.500 "superblock": false, 00:14:55.500 "num_base_bdevs": 3, 00:14:55.500 "num_base_bdevs_discovered": 2, 00:14:55.500 "num_base_bdevs_operational": 3, 00:14:55.500 "base_bdevs_list": [ 00:14:55.500 { 00:14:55.500 "name": "BaseBdev1", 00:14:55.500 "uuid": "c0681378-efe7-4a97-b07c-dcbb46129f5a", 00:14:55.500 "is_configured": true, 00:14:55.500 "data_offset": 0, 00:14:55.500 "data_size": 65536 00:14:55.500 }, 00:14:55.500 { 00:14:55.500 "name": "BaseBdev2", 00:14:55.500 "uuid": "f6085d77-0f5e-4695-aae0-1d0357e99da6", 00:14:55.500 "is_configured": true, 00:14:55.500 "data_offset": 0, 00:14:55.500 "data_size": 65536 00:14:55.500 }, 00:14:55.500 { 00:14:55.500 "name": "BaseBdev3", 00:14:55.500 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:55.500 "is_configured": false, 00:14:55.500 "data_offset": 0, 00:14:55.500 "data_size": 0 00:14:55.500 } 00:14:55.500 ] 00:14:55.500 }' 00:14:55.500 11:25:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:55.500 11:25:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.110 11:25:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:14:56.368 [2024-07-15 11:25:39.891532] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:56.368 [2024-07-15 11:25:39.891568] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x10cb400 00:14:56.368 [2024-07-15 11:25:39.891576] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:14:56.368 [2024-07-15 11:25:39.891824] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x10caef0 00:14:56.368 [2024-07-15 11:25:39.891956] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x10cb400 00:14:56.368 [2024-07-15 11:25:39.891966] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x10cb400 00:14:56.368 [2024-07-15 11:25:39.892126] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:56.368 BaseBdev3 00:14:56.368 11:25:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev3 00:14:56.368 11:25:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:14:56.368 11:25:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:14:56.368 11:25:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:14:56.368 11:25:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:14:56.368 11:25:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:14:56.368 11:25:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:14:56.626 11:25:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:56.885 [ 00:14:56.885 { 00:14:56.885 "name": "BaseBdev3", 00:14:56.885 "aliases": [ 00:14:56.885 "56bdf796-f92a-4a82-a4a7-4af9159d4ec3" 00:14:56.885 ], 00:14:56.885 "product_name": "Malloc disk", 00:14:56.885 "block_size": 512, 00:14:56.885 "num_blocks": 65536, 00:14:56.885 "uuid": "56bdf796-f92a-4a82-a4a7-4af9159d4ec3", 00:14:56.885 "assigned_rate_limits": { 00:14:56.885 "rw_ios_per_sec": 0, 00:14:56.885 "rw_mbytes_per_sec": 0, 00:14:56.885 "r_mbytes_per_sec": 0, 00:14:56.885 "w_mbytes_per_sec": 0 00:14:56.885 }, 00:14:56.885 "claimed": true, 00:14:56.885 "claim_type": "exclusive_write", 00:14:56.885 "zoned": false, 00:14:56.885 "supported_io_types": { 00:14:56.885 "read": true, 00:14:56.885 "write": true, 00:14:56.885 "unmap": true, 00:14:56.885 "flush": true, 00:14:56.885 "reset": true, 00:14:56.885 "nvme_admin": false, 00:14:56.885 "nvme_io": false, 00:14:56.885 "nvme_io_md": false, 00:14:56.885 "write_zeroes": true, 00:14:56.885 "zcopy": true, 00:14:56.885 "get_zone_info": false, 00:14:56.885 "zone_management": false, 00:14:56.885 "zone_append": false, 00:14:56.885 "compare": false, 00:14:56.885 "compare_and_write": false, 00:14:56.885 "abort": true, 00:14:56.885 "seek_hole": false, 00:14:56.885 "seek_data": false, 00:14:56.885 "copy": true, 00:14:56.885 "nvme_iov_md": false 00:14:56.885 }, 00:14:56.885 "memory_domains": [ 00:14:56.885 { 00:14:56.885 "dma_device_id": "system", 00:14:56.885 "dma_device_type": 1 00:14:56.885 }, 00:14:56.885 { 00:14:56.885 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:56.885 "dma_device_type": 2 00:14:56.885 } 00:14:56.885 ], 00:14:56.885 "driver_specific": {} 00:14:56.885 } 00:14:56.885 ] 00:14:56.885 11:25:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:14:56.885 11:25:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:14:56.885 11:25:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:14:56.885 11:25:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:14:56.885 11:25:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:14:56.885 11:25:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:14:56.885 11:25:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:14:56.885 11:25:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:14:56.885 11:25:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:14:56.885 11:25:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:14:56.885 11:25:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:14:56.885 11:25:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:14:56.885 11:25:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:14:56.885 11:25:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:14:56.885 11:25:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:57.143 11:25:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:14:57.143 "name": "Existed_Raid", 00:14:57.143 "uuid": "9241a50a-4a89-4609-8546-0b2181312aa0", 00:14:57.143 "strip_size_kb": 64, 00:14:57.143 "state": "online", 00:14:57.143 "raid_level": "concat", 00:14:57.143 "superblock": false, 00:14:57.143 "num_base_bdevs": 3, 00:14:57.143 "num_base_bdevs_discovered": 3, 00:14:57.143 "num_base_bdevs_operational": 3, 00:14:57.143 "base_bdevs_list": [ 00:14:57.143 { 00:14:57.143 "name": "BaseBdev1", 00:14:57.143 "uuid": "c0681378-efe7-4a97-b07c-dcbb46129f5a", 00:14:57.143 "is_configured": true, 00:14:57.143 "data_offset": 0, 00:14:57.143 "data_size": 65536 00:14:57.143 }, 00:14:57.144 { 00:14:57.144 "name": "BaseBdev2", 00:14:57.144 "uuid": "f6085d77-0f5e-4695-aae0-1d0357e99da6", 00:14:57.144 "is_configured": true, 00:14:57.144 "data_offset": 0, 00:14:57.144 "data_size": 65536 00:14:57.144 }, 00:14:57.144 { 00:14:57.144 "name": "BaseBdev3", 00:14:57.144 "uuid": "56bdf796-f92a-4a82-a4a7-4af9159d4ec3", 00:14:57.144 "is_configured": true, 00:14:57.144 "data_offset": 0, 00:14:57.144 "data_size": 65536 00:14:57.144 } 00:14:57.144 ] 00:14:57.144 }' 00:14:57.144 11:25:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:14:57.144 11:25:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.710 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:14:57.710 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:14:57.710 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:14:57.710 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:14:57.710 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:14:57.710 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # local name 00:14:57.710 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:14:57.710 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:14:57.969 [2024-07-15 11:25:41.484088] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:57.969 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:14:57.969 "name": "Existed_Raid", 00:14:57.969 "aliases": [ 00:14:57.969 "9241a50a-4a89-4609-8546-0b2181312aa0" 00:14:57.969 ], 00:14:57.969 "product_name": "Raid Volume", 00:14:57.969 "block_size": 512, 00:14:57.969 "num_blocks": 196608, 00:14:57.969 "uuid": "9241a50a-4a89-4609-8546-0b2181312aa0", 00:14:57.969 "assigned_rate_limits": { 00:14:57.969 "rw_ios_per_sec": 0, 00:14:57.969 "rw_mbytes_per_sec": 0, 00:14:57.969 "r_mbytes_per_sec": 0, 00:14:57.969 "w_mbytes_per_sec": 0 00:14:57.969 }, 00:14:57.969 "claimed": false, 00:14:57.969 "zoned": false, 00:14:57.969 "supported_io_types": { 00:14:57.969 "read": true, 00:14:57.969 "write": true, 00:14:57.969 "unmap": true, 00:14:57.969 "flush": true, 00:14:57.969 "reset": true, 00:14:57.969 "nvme_admin": false, 00:14:57.969 "nvme_io": false, 00:14:57.969 "nvme_io_md": false, 00:14:57.969 "write_zeroes": true, 00:14:57.969 "zcopy": false, 00:14:57.969 "get_zone_info": false, 00:14:57.969 "zone_management": false, 00:14:57.969 "zone_append": false, 00:14:57.969 "compare": false, 00:14:57.969 "compare_and_write": false, 00:14:57.969 "abort": false, 00:14:57.969 "seek_hole": false, 00:14:57.969 "seek_data": false, 00:14:57.969 "copy": false, 00:14:57.969 "nvme_iov_md": false 00:14:57.969 }, 00:14:57.969 "memory_domains": [ 00:14:57.969 { 00:14:57.969 "dma_device_id": "system", 00:14:57.969 "dma_device_type": 1 00:14:57.969 }, 00:14:57.969 { 00:14:57.969 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:57.969 "dma_device_type": 2 00:14:57.969 }, 00:14:57.969 { 00:14:57.969 "dma_device_id": "system", 00:14:57.969 "dma_device_type": 1 00:14:57.969 }, 00:14:57.969 { 00:14:57.969 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:57.969 "dma_device_type": 2 00:14:57.969 }, 00:14:57.969 { 00:14:57.969 "dma_device_id": "system", 00:14:57.969 "dma_device_type": 1 00:14:57.969 }, 00:14:57.969 { 00:14:57.969 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:57.969 "dma_device_type": 2 00:14:57.969 } 00:14:57.969 ], 00:14:57.969 "driver_specific": { 00:14:57.969 "raid": { 00:14:57.969 "uuid": "9241a50a-4a89-4609-8546-0b2181312aa0", 00:14:57.969 "strip_size_kb": 64, 00:14:57.969 "state": "online", 00:14:57.969 "raid_level": "concat", 00:14:57.969 "superblock": false, 00:14:57.969 "num_base_bdevs": 3, 00:14:57.969 "num_base_bdevs_discovered": 3, 00:14:57.969 "num_base_bdevs_operational": 3, 00:14:57.969 "base_bdevs_list": [ 00:14:57.969 { 00:14:57.969 "name": "BaseBdev1", 00:14:57.969 "uuid": "c0681378-efe7-4a97-b07c-dcbb46129f5a", 00:14:57.969 "is_configured": true, 00:14:57.969 "data_offset": 0, 00:14:57.969 "data_size": 65536 00:14:57.969 }, 00:14:57.969 { 00:14:57.969 "name": "BaseBdev2", 00:14:57.969 "uuid": "f6085d77-0f5e-4695-aae0-1d0357e99da6", 00:14:57.969 "is_configured": true, 00:14:57.969 "data_offset": 0, 00:14:57.969 "data_size": 65536 00:14:57.969 }, 00:14:57.969 { 00:14:57.969 "name": "BaseBdev3", 00:14:57.969 "uuid": "56bdf796-f92a-4a82-a4a7-4af9159d4ec3", 00:14:57.969 "is_configured": true, 00:14:57.969 "data_offset": 0, 00:14:57.969 "data_size": 65536 00:14:57.969 } 00:14:57.969 ] 00:14:57.969 } 00:14:57.969 } 00:14:57.969 }' 00:14:57.969 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:57.969 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:14:57.969 BaseBdev2 00:14:57.969 BaseBdev3' 00:14:57.969 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:14:57.969 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:14:57.969 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:14:58.228 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:14:58.228 "name": "BaseBdev1", 00:14:58.228 "aliases": [ 00:14:58.228 "c0681378-efe7-4a97-b07c-dcbb46129f5a" 00:14:58.228 ], 00:14:58.228 "product_name": "Malloc disk", 00:14:58.228 "block_size": 512, 00:14:58.228 "num_blocks": 65536, 00:14:58.228 "uuid": "c0681378-efe7-4a97-b07c-dcbb46129f5a", 00:14:58.228 "assigned_rate_limits": { 00:14:58.228 "rw_ios_per_sec": 0, 00:14:58.228 "rw_mbytes_per_sec": 0, 00:14:58.228 "r_mbytes_per_sec": 0, 00:14:58.228 "w_mbytes_per_sec": 0 00:14:58.228 }, 00:14:58.228 "claimed": true, 00:14:58.228 "claim_type": "exclusive_write", 00:14:58.228 "zoned": false, 00:14:58.228 "supported_io_types": { 00:14:58.228 "read": true, 00:14:58.228 "write": true, 00:14:58.228 "unmap": true, 00:14:58.228 "flush": true, 00:14:58.228 "reset": true, 00:14:58.228 "nvme_admin": false, 00:14:58.228 "nvme_io": false, 00:14:58.228 "nvme_io_md": false, 00:14:58.228 "write_zeroes": true, 00:14:58.228 "zcopy": true, 00:14:58.228 "get_zone_info": false, 00:14:58.228 "zone_management": false, 00:14:58.228 "zone_append": false, 00:14:58.228 "compare": false, 00:14:58.228 "compare_and_write": false, 00:14:58.228 "abort": true, 00:14:58.228 "seek_hole": false, 00:14:58.228 "seek_data": false, 00:14:58.228 "copy": true, 00:14:58.228 "nvme_iov_md": false 00:14:58.228 }, 00:14:58.228 "memory_domains": [ 00:14:58.228 { 00:14:58.228 "dma_device_id": "system", 00:14:58.228 "dma_device_type": 1 00:14:58.228 }, 00:14:58.228 { 00:14:58.228 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:58.228 "dma_device_type": 2 00:14:58.228 } 00:14:58.228 ], 00:14:58.228 "driver_specific": {} 00:14:58.228 }' 00:14:58.228 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:58.487 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:58.487 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:14:58.487 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:58.487 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:58.487 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:14:58.487 11:25:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:58.487 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:58.487 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:14:58.487 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:58.747 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:58.747 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:14:58.747 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:14:58.747 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:14:58.747 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:14:59.005 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:14:59.005 "name": "BaseBdev2", 00:14:59.005 "aliases": [ 00:14:59.005 "f6085d77-0f5e-4695-aae0-1d0357e99da6" 00:14:59.005 ], 00:14:59.005 "product_name": "Malloc disk", 00:14:59.005 "block_size": 512, 00:14:59.005 "num_blocks": 65536, 00:14:59.005 "uuid": "f6085d77-0f5e-4695-aae0-1d0357e99da6", 00:14:59.005 "assigned_rate_limits": { 00:14:59.005 "rw_ios_per_sec": 0, 00:14:59.005 "rw_mbytes_per_sec": 0, 00:14:59.005 "r_mbytes_per_sec": 0, 00:14:59.005 "w_mbytes_per_sec": 0 00:14:59.005 }, 00:14:59.005 "claimed": true, 00:14:59.005 "claim_type": "exclusive_write", 00:14:59.005 "zoned": false, 00:14:59.005 "supported_io_types": { 00:14:59.005 "read": true, 00:14:59.005 "write": true, 00:14:59.005 "unmap": true, 00:14:59.005 "flush": true, 00:14:59.005 "reset": true, 00:14:59.005 "nvme_admin": false, 00:14:59.005 "nvme_io": false, 00:14:59.005 "nvme_io_md": false, 00:14:59.005 "write_zeroes": true, 00:14:59.005 "zcopy": true, 00:14:59.005 "get_zone_info": false, 00:14:59.005 "zone_management": false, 00:14:59.005 "zone_append": false, 00:14:59.005 "compare": false, 00:14:59.005 "compare_and_write": false, 00:14:59.005 "abort": true, 00:14:59.005 "seek_hole": false, 00:14:59.005 "seek_data": false, 00:14:59.005 "copy": true, 00:14:59.005 "nvme_iov_md": false 00:14:59.005 }, 00:14:59.005 "memory_domains": [ 00:14:59.005 { 00:14:59.005 "dma_device_id": "system", 00:14:59.005 "dma_device_type": 1 00:14:59.005 }, 00:14:59.005 { 00:14:59.005 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:59.005 "dma_device_type": 2 00:14:59.005 } 00:14:59.005 ], 00:14:59.005 "driver_specific": {} 00:14:59.005 }' 00:14:59.005 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:59.005 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:59.005 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:14:59.005 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:59.005 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:59.005 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:14:59.005 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:59.005 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:59.264 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:14:59.264 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:59.264 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:59.264 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:14:59.264 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:14:59.264 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:14:59.264 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:14:59.523 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:14:59.523 "name": "BaseBdev3", 00:14:59.523 "aliases": [ 00:14:59.523 "56bdf796-f92a-4a82-a4a7-4af9159d4ec3" 00:14:59.523 ], 00:14:59.523 "product_name": "Malloc disk", 00:14:59.523 "block_size": 512, 00:14:59.523 "num_blocks": 65536, 00:14:59.523 "uuid": "56bdf796-f92a-4a82-a4a7-4af9159d4ec3", 00:14:59.523 "assigned_rate_limits": { 00:14:59.523 "rw_ios_per_sec": 0, 00:14:59.523 "rw_mbytes_per_sec": 0, 00:14:59.523 "r_mbytes_per_sec": 0, 00:14:59.523 "w_mbytes_per_sec": 0 00:14:59.523 }, 00:14:59.523 "claimed": true, 00:14:59.523 "claim_type": "exclusive_write", 00:14:59.523 "zoned": false, 00:14:59.523 "supported_io_types": { 00:14:59.523 "read": true, 00:14:59.523 "write": true, 00:14:59.523 "unmap": true, 00:14:59.523 "flush": true, 00:14:59.523 "reset": true, 00:14:59.523 "nvme_admin": false, 00:14:59.523 "nvme_io": false, 00:14:59.523 "nvme_io_md": false, 00:14:59.523 "write_zeroes": true, 00:14:59.523 "zcopy": true, 00:14:59.523 "get_zone_info": false, 00:14:59.523 "zone_management": false, 00:14:59.523 "zone_append": false, 00:14:59.523 "compare": false, 00:14:59.523 "compare_and_write": false, 00:14:59.523 "abort": true, 00:14:59.523 "seek_hole": false, 00:14:59.523 "seek_data": false, 00:14:59.523 "copy": true, 00:14:59.523 "nvme_iov_md": false 00:14:59.523 }, 00:14:59.523 "memory_domains": [ 00:14:59.523 { 00:14:59.523 "dma_device_id": "system", 00:14:59.523 "dma_device_type": 1 00:14:59.523 }, 00:14:59.523 { 00:14:59.523 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:59.523 "dma_device_type": 2 00:14:59.523 } 00:14:59.523 ], 00:14:59.523 "driver_specific": {} 00:14:59.523 }' 00:14:59.523 11:25:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:59.523 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:14:59.523 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:14:59.523 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:59.523 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:14:59.797 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:14:59.797 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:59.797 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:14:59.797 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:14:59.797 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:59.797 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:14:59.797 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:14:59.797 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:15:00.368 [2024-07-15 11:25:43.810026] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:00.368 [2024-07-15 11:25:43.810054] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:00.368 [2024-07-15 11:25:43.810097] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:00.368 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@275 -- # local expected_state 00:15:00.368 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # has_redundancy concat 00:15:00.368 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:15:00.368 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # return 1 00:15:00.368 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@277 -- # expected_state=offline 00:15:00.368 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:15:00.368 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:00.368 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=offline 00:15:00.368 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:00.368 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:00.368 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:15:00.368 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:00.368 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:00.368 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:00.368 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:00.368 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:00.368 11:25:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:00.626 11:25:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:00.626 "name": "Existed_Raid", 00:15:00.626 "uuid": "9241a50a-4a89-4609-8546-0b2181312aa0", 00:15:00.626 "strip_size_kb": 64, 00:15:00.626 "state": "offline", 00:15:00.626 "raid_level": "concat", 00:15:00.626 "superblock": false, 00:15:00.626 "num_base_bdevs": 3, 00:15:00.626 "num_base_bdevs_discovered": 2, 00:15:00.626 "num_base_bdevs_operational": 2, 00:15:00.626 "base_bdevs_list": [ 00:15:00.626 { 00:15:00.626 "name": null, 00:15:00.626 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:00.626 "is_configured": false, 00:15:00.626 "data_offset": 0, 00:15:00.626 "data_size": 65536 00:15:00.626 }, 00:15:00.626 { 00:15:00.626 "name": "BaseBdev2", 00:15:00.626 "uuid": "f6085d77-0f5e-4695-aae0-1d0357e99da6", 00:15:00.626 "is_configured": true, 00:15:00.626 "data_offset": 0, 00:15:00.626 "data_size": 65536 00:15:00.626 }, 00:15:00.626 { 00:15:00.626 "name": "BaseBdev3", 00:15:00.626 "uuid": "56bdf796-f92a-4a82-a4a7-4af9159d4ec3", 00:15:00.626 "is_configured": true, 00:15:00.626 "data_offset": 0, 00:15:00.626 "data_size": 65536 00:15:00.626 } 00:15:00.626 ] 00:15:00.626 }' 00:15:00.626 11:25:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:00.626 11:25:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:01.193 11:25:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:15:01.193 11:25:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:15:01.193 11:25:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:01.193 11:25:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:15:01.451 11:25:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:15:01.451 11:25:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:01.451 11:25:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:15:01.710 [2024-07-15 11:25:45.071279] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:01.710 11:25:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:15:01.710 11:25:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:15:01.710 11:25:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:01.710 11:25:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:15:01.969 11:25:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:15:01.969 11:25:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:01.969 11:25:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev3 00:15:02.228 [2024-07-15 11:25:45.567561] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:15:02.228 [2024-07-15 11:25:45.567601] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x10cb400 name Existed_Raid, state offline 00:15:02.228 11:25:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:15:02.228 11:25:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:15:02.228 11:25:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:02.228 11:25:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:15:02.487 11:25:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:15:02.487 11:25:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:15:02.487 11:25:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # '[' 3 -gt 2 ']' 00:15:02.487 11:25:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i = 1 )) 00:15:02.487 11:25:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:15:02.487 11:25:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:15:02.487 BaseBdev2 00:15:02.487 11:25:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev2 00:15:02.487 11:25:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:15:02.487 11:25:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:15:02.487 11:25:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:15:02.487 11:25:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:15:02.487 11:25:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:15:02.487 11:25:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:15:02.745 11:25:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:03.004 [ 00:15:03.004 { 00:15:03.004 "name": "BaseBdev2", 00:15:03.005 "aliases": [ 00:15:03.005 "1ddd7008-4f0d-47a8-87e4-535f57e249cf" 00:15:03.005 ], 00:15:03.005 "product_name": "Malloc disk", 00:15:03.005 "block_size": 512, 00:15:03.005 "num_blocks": 65536, 00:15:03.005 "uuid": "1ddd7008-4f0d-47a8-87e4-535f57e249cf", 00:15:03.005 "assigned_rate_limits": { 00:15:03.005 "rw_ios_per_sec": 0, 00:15:03.005 "rw_mbytes_per_sec": 0, 00:15:03.005 "r_mbytes_per_sec": 0, 00:15:03.005 "w_mbytes_per_sec": 0 00:15:03.005 }, 00:15:03.005 "claimed": false, 00:15:03.005 "zoned": false, 00:15:03.005 "supported_io_types": { 00:15:03.005 "read": true, 00:15:03.005 "write": true, 00:15:03.005 "unmap": true, 00:15:03.005 "flush": true, 00:15:03.005 "reset": true, 00:15:03.005 "nvme_admin": false, 00:15:03.005 "nvme_io": false, 00:15:03.005 "nvme_io_md": false, 00:15:03.005 "write_zeroes": true, 00:15:03.005 "zcopy": true, 00:15:03.005 "get_zone_info": false, 00:15:03.005 "zone_management": false, 00:15:03.005 "zone_append": false, 00:15:03.005 "compare": false, 00:15:03.005 "compare_and_write": false, 00:15:03.005 "abort": true, 00:15:03.005 "seek_hole": false, 00:15:03.005 "seek_data": false, 00:15:03.005 "copy": true, 00:15:03.005 "nvme_iov_md": false 00:15:03.005 }, 00:15:03.005 "memory_domains": [ 00:15:03.005 { 00:15:03.005 "dma_device_id": "system", 00:15:03.005 "dma_device_type": 1 00:15:03.005 }, 00:15:03.005 { 00:15:03.005 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:03.005 "dma_device_type": 2 00:15:03.005 } 00:15:03.005 ], 00:15:03.005 "driver_specific": {} 00:15:03.005 } 00:15:03.005 ] 00:15:03.005 11:25:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:15:03.005 11:25:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:15:03.005 11:25:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:15:03.005 11:25:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:15:03.571 BaseBdev3 00:15:03.571 11:25:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev3 00:15:03.571 11:25:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:15:03.571 11:25:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:15:03.571 11:25:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:15:03.571 11:25:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:15:03.571 11:25:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:15:03.571 11:25:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:15:03.828 11:25:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:15:04.087 [ 00:15:04.087 { 00:15:04.087 "name": "BaseBdev3", 00:15:04.087 "aliases": [ 00:15:04.087 "ffb4e9cf-b6c7-40f9-a2d6-5e5ecbef5504" 00:15:04.087 ], 00:15:04.087 "product_name": "Malloc disk", 00:15:04.087 "block_size": 512, 00:15:04.087 "num_blocks": 65536, 00:15:04.087 "uuid": "ffb4e9cf-b6c7-40f9-a2d6-5e5ecbef5504", 00:15:04.087 "assigned_rate_limits": { 00:15:04.087 "rw_ios_per_sec": 0, 00:15:04.087 "rw_mbytes_per_sec": 0, 00:15:04.087 "r_mbytes_per_sec": 0, 00:15:04.087 "w_mbytes_per_sec": 0 00:15:04.087 }, 00:15:04.087 "claimed": false, 00:15:04.087 "zoned": false, 00:15:04.087 "supported_io_types": { 00:15:04.087 "read": true, 00:15:04.087 "write": true, 00:15:04.087 "unmap": true, 00:15:04.087 "flush": true, 00:15:04.087 "reset": true, 00:15:04.087 "nvme_admin": false, 00:15:04.087 "nvme_io": false, 00:15:04.087 "nvme_io_md": false, 00:15:04.087 "write_zeroes": true, 00:15:04.087 "zcopy": true, 00:15:04.087 "get_zone_info": false, 00:15:04.087 "zone_management": false, 00:15:04.087 "zone_append": false, 00:15:04.087 "compare": false, 00:15:04.087 "compare_and_write": false, 00:15:04.087 "abort": true, 00:15:04.087 "seek_hole": false, 00:15:04.087 "seek_data": false, 00:15:04.087 "copy": true, 00:15:04.087 "nvme_iov_md": false 00:15:04.087 }, 00:15:04.087 "memory_domains": [ 00:15:04.087 { 00:15:04.087 "dma_device_id": "system", 00:15:04.087 "dma_device_type": 1 00:15:04.087 }, 00:15:04.087 { 00:15:04.087 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:04.087 "dma_device_type": 2 00:15:04.087 } 00:15:04.087 ], 00:15:04.087 "driver_specific": {} 00:15:04.087 } 00:15:04.087 ] 00:15:04.087 11:25:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:15:04.087 11:25:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:15:04.087 11:25:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:15:04.087 11:25:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@305 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:15:04.345 [2024-07-15 11:25:47.787410] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:04.345 [2024-07-15 11:25:47.787450] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:04.345 [2024-07-15 11:25:47.787468] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:04.345 [2024-07-15 11:25:47.788810] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:04.345 11:25:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:15:04.345 11:25:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:04.345 11:25:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:15:04.345 11:25:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:04.345 11:25:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:04.345 11:25:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:04.345 11:25:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:04.345 11:25:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:04.345 11:25:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:04.345 11:25:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:04.345 11:25:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:04.345 11:25:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:04.604 11:25:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:04.604 "name": "Existed_Raid", 00:15:04.604 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:04.604 "strip_size_kb": 64, 00:15:04.604 "state": "configuring", 00:15:04.604 "raid_level": "concat", 00:15:04.604 "superblock": false, 00:15:04.604 "num_base_bdevs": 3, 00:15:04.604 "num_base_bdevs_discovered": 2, 00:15:04.604 "num_base_bdevs_operational": 3, 00:15:04.604 "base_bdevs_list": [ 00:15:04.604 { 00:15:04.604 "name": "BaseBdev1", 00:15:04.604 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:04.604 "is_configured": false, 00:15:04.604 "data_offset": 0, 00:15:04.604 "data_size": 0 00:15:04.604 }, 00:15:04.604 { 00:15:04.604 "name": "BaseBdev2", 00:15:04.604 "uuid": "1ddd7008-4f0d-47a8-87e4-535f57e249cf", 00:15:04.604 "is_configured": true, 00:15:04.604 "data_offset": 0, 00:15:04.604 "data_size": 65536 00:15:04.604 }, 00:15:04.604 { 00:15:04.604 "name": "BaseBdev3", 00:15:04.604 "uuid": "ffb4e9cf-b6c7-40f9-a2d6-5e5ecbef5504", 00:15:04.604 "is_configured": true, 00:15:04.604 "data_offset": 0, 00:15:04.604 "data_size": 65536 00:15:04.604 } 00:15:04.604 ] 00:15:04.604 }' 00:15:04.604 11:25:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:04.604 11:25:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.169 11:25:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev2 00:15:05.428 [2024-07-15 11:25:48.818131] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:05.428 11:25:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@309 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:15:05.428 11:25:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:05.428 11:25:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:15:05.428 11:25:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:05.428 11:25:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:05.428 11:25:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:05.428 11:25:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:05.428 11:25:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:05.428 11:25:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:05.428 11:25:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:05.428 11:25:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:05.428 11:25:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:05.428 11:25:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:05.428 "name": "Existed_Raid", 00:15:05.428 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:05.428 "strip_size_kb": 64, 00:15:05.428 "state": "configuring", 00:15:05.428 "raid_level": "concat", 00:15:05.428 "superblock": false, 00:15:05.428 "num_base_bdevs": 3, 00:15:05.428 "num_base_bdevs_discovered": 1, 00:15:05.428 "num_base_bdevs_operational": 3, 00:15:05.428 "base_bdevs_list": [ 00:15:05.428 { 00:15:05.428 "name": "BaseBdev1", 00:15:05.428 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:05.428 "is_configured": false, 00:15:05.428 "data_offset": 0, 00:15:05.428 "data_size": 0 00:15:05.428 }, 00:15:05.428 { 00:15:05.428 "name": null, 00:15:05.428 "uuid": "1ddd7008-4f0d-47a8-87e4-535f57e249cf", 00:15:05.428 "is_configured": false, 00:15:05.428 "data_offset": 0, 00:15:05.428 "data_size": 65536 00:15:05.428 }, 00:15:05.428 { 00:15:05.428 "name": "BaseBdev3", 00:15:05.428 "uuid": "ffb4e9cf-b6c7-40f9-a2d6-5e5ecbef5504", 00:15:05.428 "is_configured": true, 00:15:05.428 "data_offset": 0, 00:15:05.428 "data_size": 65536 00:15:05.428 } 00:15:05.428 ] 00:15:05.428 }' 00:15:05.428 11:25:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:05.428 11:25:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.995 11:25:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:05.995 11:25:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:15:06.253 11:25:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # [[ false == \f\a\l\s\e ]] 00:15:06.253 11:25:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:15:06.511 [2024-07-15 11:25:49.933653] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:06.511 BaseBdev1 00:15:06.511 11:25:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@313 -- # waitforbdev BaseBdev1 00:15:06.511 11:25:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:15:06.511 11:25:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:15:06.511 11:25:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:15:06.511 11:25:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:15:06.511 11:25:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:15:06.511 11:25:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:15:06.769 11:25:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:07.028 [ 00:15:07.028 { 00:15:07.028 "name": "BaseBdev1", 00:15:07.028 "aliases": [ 00:15:07.028 "335a3305-c3bb-4b14-bbdf-a98f723bc2c9" 00:15:07.028 ], 00:15:07.028 "product_name": "Malloc disk", 00:15:07.028 "block_size": 512, 00:15:07.028 "num_blocks": 65536, 00:15:07.028 "uuid": "335a3305-c3bb-4b14-bbdf-a98f723bc2c9", 00:15:07.028 "assigned_rate_limits": { 00:15:07.028 "rw_ios_per_sec": 0, 00:15:07.028 "rw_mbytes_per_sec": 0, 00:15:07.028 "r_mbytes_per_sec": 0, 00:15:07.028 "w_mbytes_per_sec": 0 00:15:07.028 }, 00:15:07.028 "claimed": true, 00:15:07.028 "claim_type": "exclusive_write", 00:15:07.028 "zoned": false, 00:15:07.028 "supported_io_types": { 00:15:07.028 "read": true, 00:15:07.028 "write": true, 00:15:07.028 "unmap": true, 00:15:07.028 "flush": true, 00:15:07.028 "reset": true, 00:15:07.028 "nvme_admin": false, 00:15:07.028 "nvme_io": false, 00:15:07.028 "nvme_io_md": false, 00:15:07.028 "write_zeroes": true, 00:15:07.028 "zcopy": true, 00:15:07.028 "get_zone_info": false, 00:15:07.028 "zone_management": false, 00:15:07.028 "zone_append": false, 00:15:07.028 "compare": false, 00:15:07.028 "compare_and_write": false, 00:15:07.028 "abort": true, 00:15:07.028 "seek_hole": false, 00:15:07.028 "seek_data": false, 00:15:07.028 "copy": true, 00:15:07.028 "nvme_iov_md": false 00:15:07.028 }, 00:15:07.028 "memory_domains": [ 00:15:07.028 { 00:15:07.028 "dma_device_id": "system", 00:15:07.028 "dma_device_type": 1 00:15:07.028 }, 00:15:07.028 { 00:15:07.028 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:07.028 "dma_device_type": 2 00:15:07.028 } 00:15:07.028 ], 00:15:07.028 "driver_specific": {} 00:15:07.028 } 00:15:07.028 ] 00:15:07.028 11:25:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:15:07.028 11:25:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:15:07.028 11:25:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:07.028 11:25:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:15:07.028 11:25:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:07.028 11:25:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:07.028 11:25:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:07.028 11:25:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:07.028 11:25:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:07.028 11:25:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:07.028 11:25:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:07.028 11:25:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:07.028 11:25:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:07.287 11:25:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:07.287 "name": "Existed_Raid", 00:15:07.287 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:07.287 "strip_size_kb": 64, 00:15:07.287 "state": "configuring", 00:15:07.287 "raid_level": "concat", 00:15:07.287 "superblock": false, 00:15:07.287 "num_base_bdevs": 3, 00:15:07.287 "num_base_bdevs_discovered": 2, 00:15:07.287 "num_base_bdevs_operational": 3, 00:15:07.287 "base_bdevs_list": [ 00:15:07.287 { 00:15:07.287 "name": "BaseBdev1", 00:15:07.287 "uuid": "335a3305-c3bb-4b14-bbdf-a98f723bc2c9", 00:15:07.287 "is_configured": true, 00:15:07.287 "data_offset": 0, 00:15:07.287 "data_size": 65536 00:15:07.287 }, 00:15:07.287 { 00:15:07.287 "name": null, 00:15:07.287 "uuid": "1ddd7008-4f0d-47a8-87e4-535f57e249cf", 00:15:07.287 "is_configured": false, 00:15:07.287 "data_offset": 0, 00:15:07.287 "data_size": 65536 00:15:07.287 }, 00:15:07.287 { 00:15:07.287 "name": "BaseBdev3", 00:15:07.287 "uuid": "ffb4e9cf-b6c7-40f9-a2d6-5e5ecbef5504", 00:15:07.287 "is_configured": true, 00:15:07.287 "data_offset": 0, 00:15:07.287 "data_size": 65536 00:15:07.287 } 00:15:07.287 ] 00:15:07.287 }' 00:15:07.287 11:25:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:07.287 11:25:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.853 11:25:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:15:07.853 11:25:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:08.111 11:25:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # [[ true == \t\r\u\e ]] 00:15:08.111 11:25:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@317 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev3 00:15:08.370 [2024-07-15 11:25:51.790620] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:15:08.370 11:25:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:15:08.370 11:25:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:08.370 11:25:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:15:08.370 11:25:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:08.370 11:25:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:08.370 11:25:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:08.370 11:25:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:08.370 11:25:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:08.370 11:25:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:08.370 11:25:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:08.371 11:25:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:08.371 11:25:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:08.629 11:25:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:08.629 "name": "Existed_Raid", 00:15:08.629 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:08.629 "strip_size_kb": 64, 00:15:08.629 "state": "configuring", 00:15:08.629 "raid_level": "concat", 00:15:08.629 "superblock": false, 00:15:08.629 "num_base_bdevs": 3, 00:15:08.629 "num_base_bdevs_discovered": 1, 00:15:08.629 "num_base_bdevs_operational": 3, 00:15:08.629 "base_bdevs_list": [ 00:15:08.629 { 00:15:08.629 "name": "BaseBdev1", 00:15:08.629 "uuid": "335a3305-c3bb-4b14-bbdf-a98f723bc2c9", 00:15:08.629 "is_configured": true, 00:15:08.629 "data_offset": 0, 00:15:08.629 "data_size": 65536 00:15:08.629 }, 00:15:08.629 { 00:15:08.629 "name": null, 00:15:08.629 "uuid": "1ddd7008-4f0d-47a8-87e4-535f57e249cf", 00:15:08.629 "is_configured": false, 00:15:08.629 "data_offset": 0, 00:15:08.629 "data_size": 65536 00:15:08.629 }, 00:15:08.629 { 00:15:08.629 "name": null, 00:15:08.629 "uuid": "ffb4e9cf-b6c7-40f9-a2d6-5e5ecbef5504", 00:15:08.629 "is_configured": false, 00:15:08.629 "data_offset": 0, 00:15:08.629 "data_size": 65536 00:15:08.629 } 00:15:08.629 ] 00:15:08.629 }' 00:15:08.629 11:25:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:08.629 11:25:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:09.195 11:25:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:09.195 11:25:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:15:09.453 11:25:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # [[ false == \f\a\l\s\e ]] 00:15:09.453 11:25:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:15:09.720 [2024-07-15 11:25:53.074055] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:09.720 11:25:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@322 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:15:09.720 11:25:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:09.720 11:25:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:15:09.720 11:25:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:09.720 11:25:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:09.720 11:25:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:09.720 11:25:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:09.720 11:25:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:09.720 11:25:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:09.720 11:25:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:09.720 11:25:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:09.720 11:25:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:09.720 11:25:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:09.720 "name": "Existed_Raid", 00:15:09.720 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:09.720 "strip_size_kb": 64, 00:15:09.720 "state": "configuring", 00:15:09.720 "raid_level": "concat", 00:15:09.720 "superblock": false, 00:15:09.720 "num_base_bdevs": 3, 00:15:09.720 "num_base_bdevs_discovered": 2, 00:15:09.720 "num_base_bdevs_operational": 3, 00:15:09.720 "base_bdevs_list": [ 00:15:09.720 { 00:15:09.720 "name": "BaseBdev1", 00:15:09.720 "uuid": "335a3305-c3bb-4b14-bbdf-a98f723bc2c9", 00:15:09.720 "is_configured": true, 00:15:09.720 "data_offset": 0, 00:15:09.720 "data_size": 65536 00:15:09.720 }, 00:15:09.720 { 00:15:09.720 "name": null, 00:15:09.720 "uuid": "1ddd7008-4f0d-47a8-87e4-535f57e249cf", 00:15:09.720 "is_configured": false, 00:15:09.720 "data_offset": 0, 00:15:09.720 "data_size": 65536 00:15:09.720 }, 00:15:09.720 { 00:15:09.720 "name": "BaseBdev3", 00:15:09.720 "uuid": "ffb4e9cf-b6c7-40f9-a2d6-5e5ecbef5504", 00:15:09.720 "is_configured": true, 00:15:09.720 "data_offset": 0, 00:15:09.720 "data_size": 65536 00:15:09.720 } 00:15:09.720 ] 00:15:09.720 }' 00:15:09.720 11:25:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:09.720 11:25:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.371 11:25:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:10.371 11:25:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:15:10.628 11:25:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # [[ true == \t\r\u\e ]] 00:15:10.628 11:25:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@325 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:15:10.885 [2024-07-15 11:25:54.297314] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:10.885 11:25:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:15:10.885 11:25:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:10.885 11:25:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:15:10.885 11:25:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:10.885 11:25:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:10.885 11:25:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:10.885 11:25:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:10.885 11:25:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:10.885 11:25:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:10.885 11:25:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:10.885 11:25:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:10.885 11:25:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:11.140 11:25:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:11.140 "name": "Existed_Raid", 00:15:11.140 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:11.140 "strip_size_kb": 64, 00:15:11.140 "state": "configuring", 00:15:11.140 "raid_level": "concat", 00:15:11.140 "superblock": false, 00:15:11.140 "num_base_bdevs": 3, 00:15:11.140 "num_base_bdevs_discovered": 1, 00:15:11.140 "num_base_bdevs_operational": 3, 00:15:11.140 "base_bdevs_list": [ 00:15:11.140 { 00:15:11.140 "name": null, 00:15:11.140 "uuid": "335a3305-c3bb-4b14-bbdf-a98f723bc2c9", 00:15:11.140 "is_configured": false, 00:15:11.140 "data_offset": 0, 00:15:11.140 "data_size": 65536 00:15:11.140 }, 00:15:11.140 { 00:15:11.140 "name": null, 00:15:11.140 "uuid": "1ddd7008-4f0d-47a8-87e4-535f57e249cf", 00:15:11.140 "is_configured": false, 00:15:11.140 "data_offset": 0, 00:15:11.140 "data_size": 65536 00:15:11.140 }, 00:15:11.140 { 00:15:11.140 "name": "BaseBdev3", 00:15:11.140 "uuid": "ffb4e9cf-b6c7-40f9-a2d6-5e5ecbef5504", 00:15:11.140 "is_configured": true, 00:15:11.140 "data_offset": 0, 00:15:11.140 "data_size": 65536 00:15:11.140 } 00:15:11.140 ] 00:15:11.140 }' 00:15:11.140 11:25:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:11.140 11:25:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.703 11:25:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@327 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:11.703 11:25:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@327 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:15:11.961 11:25:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@327 -- # [[ false == \f\a\l\s\e ]] 00:15:11.961 11:25:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@329 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:15:12.219 [2024-07-15 11:25:55.576656] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:12.219 11:25:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@330 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:15:12.219 11:25:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:12.219 11:25:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:15:12.219 11:25:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:12.219 11:25:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:12.219 11:25:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:12.219 11:25:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:12.219 11:25:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:12.219 11:25:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:12.219 11:25:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:12.219 11:25:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:12.219 11:25:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:12.476 11:25:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:12.476 "name": "Existed_Raid", 00:15:12.476 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:12.476 "strip_size_kb": 64, 00:15:12.476 "state": "configuring", 00:15:12.476 "raid_level": "concat", 00:15:12.476 "superblock": false, 00:15:12.476 "num_base_bdevs": 3, 00:15:12.476 "num_base_bdevs_discovered": 2, 00:15:12.476 "num_base_bdevs_operational": 3, 00:15:12.476 "base_bdevs_list": [ 00:15:12.476 { 00:15:12.476 "name": null, 00:15:12.476 "uuid": "335a3305-c3bb-4b14-bbdf-a98f723bc2c9", 00:15:12.476 "is_configured": false, 00:15:12.476 "data_offset": 0, 00:15:12.476 "data_size": 65536 00:15:12.476 }, 00:15:12.476 { 00:15:12.476 "name": "BaseBdev2", 00:15:12.476 "uuid": "1ddd7008-4f0d-47a8-87e4-535f57e249cf", 00:15:12.476 "is_configured": true, 00:15:12.476 "data_offset": 0, 00:15:12.476 "data_size": 65536 00:15:12.476 }, 00:15:12.476 { 00:15:12.476 "name": "BaseBdev3", 00:15:12.476 "uuid": "ffb4e9cf-b6c7-40f9-a2d6-5e5ecbef5504", 00:15:12.476 "is_configured": true, 00:15:12.476 "data_offset": 0, 00:15:12.476 "data_size": 65536 00:15:12.476 } 00:15:12.476 ] 00:15:12.476 }' 00:15:12.476 11:25:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:12.476 11:25:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.042 11:25:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@331 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:13.042 11:25:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@331 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:15:13.325 11:25:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@331 -- # [[ true == \t\r\u\e ]] 00:15:13.325 11:25:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:13.325 11:25:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@333 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:15:13.325 11:25:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b NewBaseBdev -u 335a3305-c3bb-4b14-bbdf-a98f723bc2c9 00:15:13.581 [2024-07-15 11:25:57.144242] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:15:13.582 [2024-07-15 11:25:57.144279] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x10c9450 00:15:13.582 [2024-07-15 11:25:57.144288] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:15:13.582 [2024-07-15 11:25:57.144476] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x10caed0 00:15:13.582 [2024-07-15 11:25:57.144589] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x10c9450 00:15:13.582 [2024-07-15 11:25:57.144599] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x10c9450 00:15:13.582 [2024-07-15 11:25:57.144762] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:13.582 NewBaseBdev 00:15:13.582 11:25:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@334 -- # waitforbdev NewBaseBdev 00:15:13.582 11:25:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=NewBaseBdev 00:15:13.582 11:25:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:15:13.582 11:25:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:15:13.582 11:25:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:15:13.582 11:25:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:15:13.582 11:25:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:15:13.837 11:25:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev -t 2000 00:15:14.094 [ 00:15:14.094 { 00:15:14.094 "name": "NewBaseBdev", 00:15:14.094 "aliases": [ 00:15:14.094 "335a3305-c3bb-4b14-bbdf-a98f723bc2c9" 00:15:14.094 ], 00:15:14.094 "product_name": "Malloc disk", 00:15:14.095 "block_size": 512, 00:15:14.095 "num_blocks": 65536, 00:15:14.095 "uuid": "335a3305-c3bb-4b14-bbdf-a98f723bc2c9", 00:15:14.095 "assigned_rate_limits": { 00:15:14.095 "rw_ios_per_sec": 0, 00:15:14.095 "rw_mbytes_per_sec": 0, 00:15:14.095 "r_mbytes_per_sec": 0, 00:15:14.095 "w_mbytes_per_sec": 0 00:15:14.095 }, 00:15:14.095 "claimed": true, 00:15:14.095 "claim_type": "exclusive_write", 00:15:14.095 "zoned": false, 00:15:14.095 "supported_io_types": { 00:15:14.095 "read": true, 00:15:14.095 "write": true, 00:15:14.095 "unmap": true, 00:15:14.095 "flush": true, 00:15:14.095 "reset": true, 00:15:14.095 "nvme_admin": false, 00:15:14.095 "nvme_io": false, 00:15:14.095 "nvme_io_md": false, 00:15:14.095 "write_zeroes": true, 00:15:14.095 "zcopy": true, 00:15:14.095 "get_zone_info": false, 00:15:14.095 "zone_management": false, 00:15:14.095 "zone_append": false, 00:15:14.095 "compare": false, 00:15:14.095 "compare_and_write": false, 00:15:14.095 "abort": true, 00:15:14.095 "seek_hole": false, 00:15:14.095 "seek_data": false, 00:15:14.095 "copy": true, 00:15:14.095 "nvme_iov_md": false 00:15:14.095 }, 00:15:14.095 "memory_domains": [ 00:15:14.095 { 00:15:14.095 "dma_device_id": "system", 00:15:14.095 "dma_device_type": 1 00:15:14.095 }, 00:15:14.095 { 00:15:14.095 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:14.095 "dma_device_type": 2 00:15:14.095 } 00:15:14.095 ], 00:15:14.095 "driver_specific": {} 00:15:14.095 } 00:15:14.095 ] 00:15:14.095 11:25:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:15:14.095 11:25:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@335 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:15:14.095 11:25:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:14.095 11:25:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:15:14.095 11:25:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:14.095 11:25:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:14.095 11:25:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:14.095 11:25:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:14.095 11:25:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:14.095 11:25:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:14.095 11:25:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:14.095 11:25:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:14.095 11:25:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:14.351 11:25:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:14.351 "name": "Existed_Raid", 00:15:14.351 "uuid": "b24f045b-ab18-4e18-900e-b19aa29b6f8f", 00:15:14.351 "strip_size_kb": 64, 00:15:14.351 "state": "online", 00:15:14.351 "raid_level": "concat", 00:15:14.351 "superblock": false, 00:15:14.351 "num_base_bdevs": 3, 00:15:14.351 "num_base_bdevs_discovered": 3, 00:15:14.351 "num_base_bdevs_operational": 3, 00:15:14.351 "base_bdevs_list": [ 00:15:14.351 { 00:15:14.351 "name": "NewBaseBdev", 00:15:14.351 "uuid": "335a3305-c3bb-4b14-bbdf-a98f723bc2c9", 00:15:14.351 "is_configured": true, 00:15:14.351 "data_offset": 0, 00:15:14.351 "data_size": 65536 00:15:14.351 }, 00:15:14.351 { 00:15:14.351 "name": "BaseBdev2", 00:15:14.351 "uuid": "1ddd7008-4f0d-47a8-87e4-535f57e249cf", 00:15:14.351 "is_configured": true, 00:15:14.351 "data_offset": 0, 00:15:14.351 "data_size": 65536 00:15:14.351 }, 00:15:14.351 { 00:15:14.351 "name": "BaseBdev3", 00:15:14.351 "uuid": "ffb4e9cf-b6c7-40f9-a2d6-5e5ecbef5504", 00:15:14.351 "is_configured": true, 00:15:14.352 "data_offset": 0, 00:15:14.352 "data_size": 65536 00:15:14.352 } 00:15:14.352 ] 00:15:14.352 }' 00:15:14.352 11:25:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:14.352 11:25:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:14.914 11:25:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@336 -- # verify_raid_bdev_properties Existed_Raid 00:15:14.914 11:25:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:15:14.914 11:25:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:15:14.914 11:25:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:15:14.914 11:25:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:15:14.914 11:25:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # local name 00:15:14.914 11:25:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:15:14.914 11:25:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:15:15.171 [2024-07-15 11:25:58.600399] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:15.171 11:25:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:15:15.171 "name": "Existed_Raid", 00:15:15.171 "aliases": [ 00:15:15.171 "b24f045b-ab18-4e18-900e-b19aa29b6f8f" 00:15:15.171 ], 00:15:15.171 "product_name": "Raid Volume", 00:15:15.171 "block_size": 512, 00:15:15.171 "num_blocks": 196608, 00:15:15.171 "uuid": "b24f045b-ab18-4e18-900e-b19aa29b6f8f", 00:15:15.171 "assigned_rate_limits": { 00:15:15.171 "rw_ios_per_sec": 0, 00:15:15.171 "rw_mbytes_per_sec": 0, 00:15:15.171 "r_mbytes_per_sec": 0, 00:15:15.171 "w_mbytes_per_sec": 0 00:15:15.171 }, 00:15:15.171 "claimed": false, 00:15:15.171 "zoned": false, 00:15:15.171 "supported_io_types": { 00:15:15.171 "read": true, 00:15:15.171 "write": true, 00:15:15.171 "unmap": true, 00:15:15.171 "flush": true, 00:15:15.171 "reset": true, 00:15:15.171 "nvme_admin": false, 00:15:15.171 "nvme_io": false, 00:15:15.171 "nvme_io_md": false, 00:15:15.171 "write_zeroes": true, 00:15:15.171 "zcopy": false, 00:15:15.171 "get_zone_info": false, 00:15:15.171 "zone_management": false, 00:15:15.171 "zone_append": false, 00:15:15.171 "compare": false, 00:15:15.171 "compare_and_write": false, 00:15:15.171 "abort": false, 00:15:15.171 "seek_hole": false, 00:15:15.171 "seek_data": false, 00:15:15.171 "copy": false, 00:15:15.171 "nvme_iov_md": false 00:15:15.171 }, 00:15:15.171 "memory_domains": [ 00:15:15.171 { 00:15:15.171 "dma_device_id": "system", 00:15:15.171 "dma_device_type": 1 00:15:15.171 }, 00:15:15.171 { 00:15:15.171 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:15.171 "dma_device_type": 2 00:15:15.171 }, 00:15:15.171 { 00:15:15.171 "dma_device_id": "system", 00:15:15.171 "dma_device_type": 1 00:15:15.171 }, 00:15:15.171 { 00:15:15.171 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:15.171 "dma_device_type": 2 00:15:15.171 }, 00:15:15.171 { 00:15:15.171 "dma_device_id": "system", 00:15:15.171 "dma_device_type": 1 00:15:15.171 }, 00:15:15.171 { 00:15:15.171 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:15.171 "dma_device_type": 2 00:15:15.171 } 00:15:15.171 ], 00:15:15.171 "driver_specific": { 00:15:15.171 "raid": { 00:15:15.171 "uuid": "b24f045b-ab18-4e18-900e-b19aa29b6f8f", 00:15:15.171 "strip_size_kb": 64, 00:15:15.171 "state": "online", 00:15:15.171 "raid_level": "concat", 00:15:15.171 "superblock": false, 00:15:15.171 "num_base_bdevs": 3, 00:15:15.171 "num_base_bdevs_discovered": 3, 00:15:15.171 "num_base_bdevs_operational": 3, 00:15:15.171 "base_bdevs_list": [ 00:15:15.171 { 00:15:15.171 "name": "NewBaseBdev", 00:15:15.171 "uuid": "335a3305-c3bb-4b14-bbdf-a98f723bc2c9", 00:15:15.171 "is_configured": true, 00:15:15.171 "data_offset": 0, 00:15:15.171 "data_size": 65536 00:15:15.171 }, 00:15:15.171 { 00:15:15.171 "name": "BaseBdev2", 00:15:15.171 "uuid": "1ddd7008-4f0d-47a8-87e4-535f57e249cf", 00:15:15.171 "is_configured": true, 00:15:15.171 "data_offset": 0, 00:15:15.171 "data_size": 65536 00:15:15.171 }, 00:15:15.171 { 00:15:15.171 "name": "BaseBdev3", 00:15:15.171 "uuid": "ffb4e9cf-b6c7-40f9-a2d6-5e5ecbef5504", 00:15:15.171 "is_configured": true, 00:15:15.171 "data_offset": 0, 00:15:15.171 "data_size": 65536 00:15:15.171 } 00:15:15.171 ] 00:15:15.171 } 00:15:15.171 } 00:15:15.171 }' 00:15:15.171 11:25:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:15.171 11:25:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='NewBaseBdev 00:15:15.171 BaseBdev2 00:15:15.171 BaseBdev3' 00:15:15.171 11:25:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:15:15.171 11:25:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev 00:15:15.171 11:25:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:15:15.428 11:25:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:15:15.428 "name": "NewBaseBdev", 00:15:15.428 "aliases": [ 00:15:15.428 "335a3305-c3bb-4b14-bbdf-a98f723bc2c9" 00:15:15.428 ], 00:15:15.428 "product_name": "Malloc disk", 00:15:15.428 "block_size": 512, 00:15:15.428 "num_blocks": 65536, 00:15:15.428 "uuid": "335a3305-c3bb-4b14-bbdf-a98f723bc2c9", 00:15:15.428 "assigned_rate_limits": { 00:15:15.428 "rw_ios_per_sec": 0, 00:15:15.428 "rw_mbytes_per_sec": 0, 00:15:15.428 "r_mbytes_per_sec": 0, 00:15:15.428 "w_mbytes_per_sec": 0 00:15:15.428 }, 00:15:15.428 "claimed": true, 00:15:15.428 "claim_type": "exclusive_write", 00:15:15.428 "zoned": false, 00:15:15.428 "supported_io_types": { 00:15:15.428 "read": true, 00:15:15.428 "write": true, 00:15:15.428 "unmap": true, 00:15:15.428 "flush": true, 00:15:15.428 "reset": true, 00:15:15.428 "nvme_admin": false, 00:15:15.428 "nvme_io": false, 00:15:15.428 "nvme_io_md": false, 00:15:15.428 "write_zeroes": true, 00:15:15.428 "zcopy": true, 00:15:15.428 "get_zone_info": false, 00:15:15.428 "zone_management": false, 00:15:15.428 "zone_append": false, 00:15:15.428 "compare": false, 00:15:15.428 "compare_and_write": false, 00:15:15.428 "abort": true, 00:15:15.428 "seek_hole": false, 00:15:15.428 "seek_data": false, 00:15:15.428 "copy": true, 00:15:15.428 "nvme_iov_md": false 00:15:15.428 }, 00:15:15.428 "memory_domains": [ 00:15:15.428 { 00:15:15.428 "dma_device_id": "system", 00:15:15.428 "dma_device_type": 1 00:15:15.428 }, 00:15:15.428 { 00:15:15.428 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:15.428 "dma_device_type": 2 00:15:15.428 } 00:15:15.428 ], 00:15:15.428 "driver_specific": {} 00:15:15.428 }' 00:15:15.428 11:25:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:15.428 11:25:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:15.428 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:15:15.428 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:15.685 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:15.685 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:15:15.685 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:15.685 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:15.685 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:15:15.685 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:15.685 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:15.685 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:15:15.685 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:15:15.685 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:15:15.685 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:15:15.942 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:15:15.942 "name": "BaseBdev2", 00:15:15.942 "aliases": [ 00:15:15.942 "1ddd7008-4f0d-47a8-87e4-535f57e249cf" 00:15:15.942 ], 00:15:15.942 "product_name": "Malloc disk", 00:15:15.942 "block_size": 512, 00:15:15.942 "num_blocks": 65536, 00:15:15.942 "uuid": "1ddd7008-4f0d-47a8-87e4-535f57e249cf", 00:15:15.942 "assigned_rate_limits": { 00:15:15.942 "rw_ios_per_sec": 0, 00:15:15.942 "rw_mbytes_per_sec": 0, 00:15:15.942 "r_mbytes_per_sec": 0, 00:15:15.942 "w_mbytes_per_sec": 0 00:15:15.942 }, 00:15:15.942 "claimed": true, 00:15:15.942 "claim_type": "exclusive_write", 00:15:15.942 "zoned": false, 00:15:15.942 "supported_io_types": { 00:15:15.942 "read": true, 00:15:15.942 "write": true, 00:15:15.942 "unmap": true, 00:15:15.942 "flush": true, 00:15:15.942 "reset": true, 00:15:15.942 "nvme_admin": false, 00:15:15.942 "nvme_io": false, 00:15:15.942 "nvme_io_md": false, 00:15:15.942 "write_zeroes": true, 00:15:15.942 "zcopy": true, 00:15:15.942 "get_zone_info": false, 00:15:15.942 "zone_management": false, 00:15:15.942 "zone_append": false, 00:15:15.942 "compare": false, 00:15:15.942 "compare_and_write": false, 00:15:15.942 "abort": true, 00:15:15.942 "seek_hole": false, 00:15:15.942 "seek_data": false, 00:15:15.942 "copy": true, 00:15:15.942 "nvme_iov_md": false 00:15:15.942 }, 00:15:15.942 "memory_domains": [ 00:15:15.942 { 00:15:15.942 "dma_device_id": "system", 00:15:15.942 "dma_device_type": 1 00:15:15.942 }, 00:15:15.942 { 00:15:15.942 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:15.942 "dma_device_type": 2 00:15:15.942 } 00:15:15.942 ], 00:15:15.942 "driver_specific": {} 00:15:15.942 }' 00:15:15.942 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:16.199 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:16.199 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:15:16.199 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:16.199 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:16.199 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:15:16.199 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:16.199 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:16.199 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:15:16.199 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:16.455 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:16.455 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:15:16.455 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:15:16.455 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:15:16.455 11:25:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:15:16.711 11:26:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:15:16.711 "name": "BaseBdev3", 00:15:16.711 "aliases": [ 00:15:16.711 "ffb4e9cf-b6c7-40f9-a2d6-5e5ecbef5504" 00:15:16.711 ], 00:15:16.711 "product_name": "Malloc disk", 00:15:16.711 "block_size": 512, 00:15:16.711 "num_blocks": 65536, 00:15:16.711 "uuid": "ffb4e9cf-b6c7-40f9-a2d6-5e5ecbef5504", 00:15:16.711 "assigned_rate_limits": { 00:15:16.711 "rw_ios_per_sec": 0, 00:15:16.711 "rw_mbytes_per_sec": 0, 00:15:16.711 "r_mbytes_per_sec": 0, 00:15:16.711 "w_mbytes_per_sec": 0 00:15:16.711 }, 00:15:16.711 "claimed": true, 00:15:16.711 "claim_type": "exclusive_write", 00:15:16.711 "zoned": false, 00:15:16.711 "supported_io_types": { 00:15:16.711 "read": true, 00:15:16.711 "write": true, 00:15:16.711 "unmap": true, 00:15:16.711 "flush": true, 00:15:16.711 "reset": true, 00:15:16.711 "nvme_admin": false, 00:15:16.711 "nvme_io": false, 00:15:16.711 "nvme_io_md": false, 00:15:16.712 "write_zeroes": true, 00:15:16.712 "zcopy": true, 00:15:16.712 "get_zone_info": false, 00:15:16.712 "zone_management": false, 00:15:16.712 "zone_append": false, 00:15:16.712 "compare": false, 00:15:16.712 "compare_and_write": false, 00:15:16.712 "abort": true, 00:15:16.712 "seek_hole": false, 00:15:16.712 "seek_data": false, 00:15:16.712 "copy": true, 00:15:16.712 "nvme_iov_md": false 00:15:16.712 }, 00:15:16.712 "memory_domains": [ 00:15:16.712 { 00:15:16.712 "dma_device_id": "system", 00:15:16.712 "dma_device_type": 1 00:15:16.712 }, 00:15:16.712 { 00:15:16.712 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:16.712 "dma_device_type": 2 00:15:16.712 } 00:15:16.712 ], 00:15:16.712 "driver_specific": {} 00:15:16.712 }' 00:15:16.712 11:26:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:16.712 11:26:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:16.712 11:26:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:15:16.712 11:26:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:16.712 11:26:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:16.712 11:26:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:15:16.712 11:26:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:16.967 11:26:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:16.967 11:26:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:15:16.967 11:26:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:16.967 11:26:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:16.967 11:26:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:15:16.967 11:26:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@338 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:15:17.224 [2024-07-15 11:26:00.629473] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:17.224 [2024-07-15 11:26:00.629498] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:17.224 [2024-07-15 11:26:00.629548] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:17.224 [2024-07-15 11:26:00.629596] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:17.224 [2024-07-15 11:26:00.629608] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x10c9450 name Existed_Raid, state offline 00:15:17.224 11:26:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@341 -- # killprocess 895556 00:15:17.224 11:26:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@948 -- # '[' -z 895556 ']' 00:15:17.224 11:26:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@952 -- # kill -0 895556 00:15:17.224 11:26:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@953 -- # uname 00:15:17.224 11:26:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:15:17.224 11:26:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 895556 00:15:17.224 11:26:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:15:17.224 11:26:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:15:17.224 11:26:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 895556' 00:15:17.224 killing process with pid 895556 00:15:17.224 11:26:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@967 -- # kill 895556 00:15:17.224 [2024-07-15 11:26:00.695519] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:17.224 11:26:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # wait 895556 00:15:17.224 [2024-07-15 11:26:00.722778] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:17.481 11:26:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@343 -- # return 0 00:15:17.481 00:15:17.481 real 0m28.255s 00:15:17.481 user 0m51.859s 00:15:17.481 sys 0m5.032s 00:15:17.481 11:26:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:15:17.481 11:26:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.481 ************************************ 00:15:17.481 END TEST raid_state_function_test 00:15:17.481 ************************************ 00:15:17.481 11:26:00 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:15:17.481 11:26:00 bdev_raid -- bdev/bdev_raid.sh@868 -- # run_test raid_state_function_test_sb raid_state_function_test concat 3 true 00:15:17.481 11:26:00 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:15:17.481 11:26:00 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:15:17.481 11:26:00 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:17.481 ************************************ 00:15:17.481 START TEST raid_state_function_test_sb 00:15:17.481 ************************************ 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1123 -- # raid_state_function_test concat 3 true 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@220 -- # local raid_level=concat 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=3 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # local superblock=true 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev3 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # local strip_size 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # '[' concat '!=' raid1 ']' 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # strip_size=64 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@232 -- # strip_size_create_arg='-z 64' 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # '[' true = true ']' 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@238 -- # superblock_create_arg=-s 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # raid_pid=899850 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 899850' 00:15:17.481 Process raid pid: 899850 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@246 -- # waitforlisten 899850 /var/tmp/spdk-raid.sock 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@829 -- # '[' -z 899850 ']' 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@834 -- # local max_retries=100 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:15:17.481 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # xtrace_disable 00:15:17.481 11:26:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:17.738 [2024-07-15 11:26:01.087907] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:15:17.738 [2024-07-15 11:26:01.087964] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:17.738 [2024-07-15 11:26:01.200617] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:17.738 [2024-07-15 11:26:01.297057] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:15:17.995 [2024-07-15 11:26:01.360015] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:17.995 [2024-07-15 11:26:01.360053] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:18.559 11:26:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:15:18.559 11:26:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@862 -- # return 0 00:15:18.559 11:26:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r concat -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:15:18.816 [2024-07-15 11:26:02.185170] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:18.816 [2024-07-15 11:26:02.185213] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:18.816 [2024-07-15 11:26:02.185224] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:18.816 [2024-07-15 11:26:02.185236] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:18.816 [2024-07-15 11:26:02.185245] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:15:18.816 [2024-07-15 11:26:02.185256] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:15:18.816 11:26:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:15:18.816 11:26:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:18.816 11:26:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:15:18.816 11:26:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:18.816 11:26:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:18.816 11:26:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:18.816 11:26:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:18.816 11:26:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:18.816 11:26:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:18.816 11:26:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:18.816 11:26:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:18.816 11:26:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:18.816 11:26:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:18.816 "name": "Existed_Raid", 00:15:18.816 "uuid": "ad96f640-8316-48ec-aef8-b8095605ec95", 00:15:18.816 "strip_size_kb": 64, 00:15:18.816 "state": "configuring", 00:15:18.816 "raid_level": "concat", 00:15:18.816 "superblock": true, 00:15:18.816 "num_base_bdevs": 3, 00:15:18.816 "num_base_bdevs_discovered": 0, 00:15:18.816 "num_base_bdevs_operational": 3, 00:15:18.816 "base_bdevs_list": [ 00:15:18.816 { 00:15:18.816 "name": "BaseBdev1", 00:15:18.816 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:18.816 "is_configured": false, 00:15:18.816 "data_offset": 0, 00:15:18.816 "data_size": 0 00:15:18.816 }, 00:15:18.816 { 00:15:18.816 "name": "BaseBdev2", 00:15:18.816 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:18.816 "is_configured": false, 00:15:18.816 "data_offset": 0, 00:15:18.816 "data_size": 0 00:15:18.816 }, 00:15:18.816 { 00:15:18.816 "name": "BaseBdev3", 00:15:18.816 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:18.816 "is_configured": false, 00:15:18.816 "data_offset": 0, 00:15:18.816 "data_size": 0 00:15:18.816 } 00:15:18.816 ] 00:15:18.816 }' 00:15:18.816 11:26:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:18.816 11:26:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:19.379 11:26:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:15:19.637 [2024-07-15 11:26:03.187665] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:19.637 [2024-07-15 11:26:03.187697] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x139da80 name Existed_Raid, state configuring 00:15:19.637 11:26:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r concat -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:15:19.895 [2024-07-15 11:26:03.436347] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:19.895 [2024-07-15 11:26:03.436379] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:19.895 [2024-07-15 11:26:03.436389] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:19.895 [2024-07-15 11:26:03.436400] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:19.895 [2024-07-15 11:26:03.436408] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:15:19.895 [2024-07-15 11:26:03.436420] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:15:19.895 11:26:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:15:20.152 [2024-07-15 11:26:03.690874] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:20.152 BaseBdev1 00:15:20.152 11:26:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:15:20.152 11:26:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:15:20.152 11:26:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:15:20.152 11:26:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:15:20.152 11:26:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:15:20.152 11:26:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:15:20.152 11:26:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:15:20.410 11:26:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:20.696 [ 00:15:20.696 { 00:15:20.696 "name": "BaseBdev1", 00:15:20.696 "aliases": [ 00:15:20.696 "fca9bc5c-de1b-4bb3-b3f0-0699ff6c2e9b" 00:15:20.696 ], 00:15:20.696 "product_name": "Malloc disk", 00:15:20.696 "block_size": 512, 00:15:20.696 "num_blocks": 65536, 00:15:20.696 "uuid": "fca9bc5c-de1b-4bb3-b3f0-0699ff6c2e9b", 00:15:20.696 "assigned_rate_limits": { 00:15:20.696 "rw_ios_per_sec": 0, 00:15:20.696 "rw_mbytes_per_sec": 0, 00:15:20.696 "r_mbytes_per_sec": 0, 00:15:20.696 "w_mbytes_per_sec": 0 00:15:20.696 }, 00:15:20.696 "claimed": true, 00:15:20.696 "claim_type": "exclusive_write", 00:15:20.696 "zoned": false, 00:15:20.696 "supported_io_types": { 00:15:20.696 "read": true, 00:15:20.696 "write": true, 00:15:20.696 "unmap": true, 00:15:20.696 "flush": true, 00:15:20.696 "reset": true, 00:15:20.696 "nvme_admin": false, 00:15:20.696 "nvme_io": false, 00:15:20.696 "nvme_io_md": false, 00:15:20.696 "write_zeroes": true, 00:15:20.696 "zcopy": true, 00:15:20.696 "get_zone_info": false, 00:15:20.696 "zone_management": false, 00:15:20.696 "zone_append": false, 00:15:20.696 "compare": false, 00:15:20.696 "compare_and_write": false, 00:15:20.696 "abort": true, 00:15:20.696 "seek_hole": false, 00:15:20.696 "seek_data": false, 00:15:20.696 "copy": true, 00:15:20.696 "nvme_iov_md": false 00:15:20.696 }, 00:15:20.696 "memory_domains": [ 00:15:20.696 { 00:15:20.696 "dma_device_id": "system", 00:15:20.696 "dma_device_type": 1 00:15:20.696 }, 00:15:20.696 { 00:15:20.696 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:20.696 "dma_device_type": 2 00:15:20.696 } 00:15:20.696 ], 00:15:20.696 "driver_specific": {} 00:15:20.696 } 00:15:20.696 ] 00:15:20.696 11:26:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:15:20.696 11:26:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:15:20.696 11:26:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:20.696 11:26:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:15:20.696 11:26:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:20.696 11:26:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:20.696 11:26:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:20.696 11:26:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:20.696 11:26:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:20.696 11:26:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:20.696 11:26:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:20.696 11:26:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:20.696 11:26:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:20.955 11:26:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:20.955 "name": "Existed_Raid", 00:15:20.955 "uuid": "ef3e7a18-5064-4dae-acb4-331a1ae193d6", 00:15:20.955 "strip_size_kb": 64, 00:15:20.955 "state": "configuring", 00:15:20.955 "raid_level": "concat", 00:15:20.955 "superblock": true, 00:15:20.955 "num_base_bdevs": 3, 00:15:20.955 "num_base_bdevs_discovered": 1, 00:15:20.955 "num_base_bdevs_operational": 3, 00:15:20.955 "base_bdevs_list": [ 00:15:20.955 { 00:15:20.955 "name": "BaseBdev1", 00:15:20.955 "uuid": "fca9bc5c-de1b-4bb3-b3f0-0699ff6c2e9b", 00:15:20.955 "is_configured": true, 00:15:20.955 "data_offset": 2048, 00:15:20.955 "data_size": 63488 00:15:20.955 }, 00:15:20.955 { 00:15:20.955 "name": "BaseBdev2", 00:15:20.955 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:20.955 "is_configured": false, 00:15:20.955 "data_offset": 0, 00:15:20.955 "data_size": 0 00:15:20.955 }, 00:15:20.955 { 00:15:20.955 "name": "BaseBdev3", 00:15:20.955 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:20.955 "is_configured": false, 00:15:20.955 "data_offset": 0, 00:15:20.955 "data_size": 0 00:15:20.955 } 00:15:20.955 ] 00:15:20.955 }' 00:15:20.955 11:26:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:20.955 11:26:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.521 11:26:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:15:21.779 [2024-07-15 11:26:05.210906] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:21.779 [2024-07-15 11:26:05.210951] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x139d310 name Existed_Raid, state configuring 00:15:21.779 11:26:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r concat -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:15:22.065 [2024-07-15 11:26:05.455604] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:22.065 [2024-07-15 11:26:05.457048] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:22.065 [2024-07-15 11:26:05.457083] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:22.065 [2024-07-15 11:26:05.457094] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:15:22.065 [2024-07-15 11:26:05.457105] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:15:22.065 11:26:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:15:22.065 11:26:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:15:22.065 11:26:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:15:22.065 11:26:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:22.065 11:26:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:15:22.065 11:26:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:22.065 11:26:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:22.065 11:26:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:22.065 11:26:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:22.065 11:26:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:22.065 11:26:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:22.065 11:26:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:22.065 11:26:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:22.065 11:26:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:22.323 11:26:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:22.323 "name": "Existed_Raid", 00:15:22.323 "uuid": "d62129ec-33b3-4d41-9611-2ce060790a4f", 00:15:22.323 "strip_size_kb": 64, 00:15:22.323 "state": "configuring", 00:15:22.323 "raid_level": "concat", 00:15:22.323 "superblock": true, 00:15:22.323 "num_base_bdevs": 3, 00:15:22.323 "num_base_bdevs_discovered": 1, 00:15:22.323 "num_base_bdevs_operational": 3, 00:15:22.323 "base_bdevs_list": [ 00:15:22.323 { 00:15:22.323 "name": "BaseBdev1", 00:15:22.323 "uuid": "fca9bc5c-de1b-4bb3-b3f0-0699ff6c2e9b", 00:15:22.323 "is_configured": true, 00:15:22.323 "data_offset": 2048, 00:15:22.323 "data_size": 63488 00:15:22.323 }, 00:15:22.323 { 00:15:22.323 "name": "BaseBdev2", 00:15:22.323 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:22.323 "is_configured": false, 00:15:22.323 "data_offset": 0, 00:15:22.323 "data_size": 0 00:15:22.323 }, 00:15:22.323 { 00:15:22.323 "name": "BaseBdev3", 00:15:22.323 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:22.323 "is_configured": false, 00:15:22.323 "data_offset": 0, 00:15:22.323 "data_size": 0 00:15:22.323 } 00:15:22.323 ] 00:15:22.323 }' 00:15:22.323 11:26:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:22.323 11:26:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:22.890 11:26:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:15:23.148 [2024-07-15 11:26:06.561871] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:23.148 BaseBdev2 00:15:23.148 11:26:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:15:23.148 11:26:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:15:23.148 11:26:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:15:23.148 11:26:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:15:23.148 11:26:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:15:23.148 11:26:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:15:23.148 11:26:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:15:23.406 11:26:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:23.406 [ 00:15:23.406 { 00:15:23.406 "name": "BaseBdev2", 00:15:23.406 "aliases": [ 00:15:23.406 "e66250a7-f1bd-496c-8413-3e698fe359c8" 00:15:23.406 ], 00:15:23.406 "product_name": "Malloc disk", 00:15:23.406 "block_size": 512, 00:15:23.406 "num_blocks": 65536, 00:15:23.406 "uuid": "e66250a7-f1bd-496c-8413-3e698fe359c8", 00:15:23.406 "assigned_rate_limits": { 00:15:23.406 "rw_ios_per_sec": 0, 00:15:23.406 "rw_mbytes_per_sec": 0, 00:15:23.406 "r_mbytes_per_sec": 0, 00:15:23.406 "w_mbytes_per_sec": 0 00:15:23.406 }, 00:15:23.406 "claimed": true, 00:15:23.406 "claim_type": "exclusive_write", 00:15:23.406 "zoned": false, 00:15:23.406 "supported_io_types": { 00:15:23.406 "read": true, 00:15:23.406 "write": true, 00:15:23.406 "unmap": true, 00:15:23.406 "flush": true, 00:15:23.406 "reset": true, 00:15:23.406 "nvme_admin": false, 00:15:23.406 "nvme_io": false, 00:15:23.406 "nvme_io_md": false, 00:15:23.406 "write_zeroes": true, 00:15:23.406 "zcopy": true, 00:15:23.406 "get_zone_info": false, 00:15:23.406 "zone_management": false, 00:15:23.406 "zone_append": false, 00:15:23.406 "compare": false, 00:15:23.406 "compare_and_write": false, 00:15:23.406 "abort": true, 00:15:23.406 "seek_hole": false, 00:15:23.406 "seek_data": false, 00:15:23.406 "copy": true, 00:15:23.406 "nvme_iov_md": false 00:15:23.406 }, 00:15:23.406 "memory_domains": [ 00:15:23.406 { 00:15:23.406 "dma_device_id": "system", 00:15:23.406 "dma_device_type": 1 00:15:23.406 }, 00:15:23.406 { 00:15:23.406 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:23.406 "dma_device_type": 2 00:15:23.406 } 00:15:23.406 ], 00:15:23.406 "driver_specific": {} 00:15:23.406 } 00:15:23.406 ] 00:15:23.664 11:26:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:15:23.664 11:26:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:15:23.664 11:26:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:15:23.664 11:26:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:15:23.664 11:26:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:23.664 11:26:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:15:23.664 11:26:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:23.664 11:26:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:23.664 11:26:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:23.664 11:26:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:23.664 11:26:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:23.664 11:26:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:23.664 11:26:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:23.664 11:26:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:23.664 11:26:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:23.922 11:26:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:23.922 "name": "Existed_Raid", 00:15:23.922 "uuid": "d62129ec-33b3-4d41-9611-2ce060790a4f", 00:15:23.922 "strip_size_kb": 64, 00:15:23.922 "state": "configuring", 00:15:23.922 "raid_level": "concat", 00:15:23.922 "superblock": true, 00:15:23.922 "num_base_bdevs": 3, 00:15:23.922 "num_base_bdevs_discovered": 2, 00:15:23.922 "num_base_bdevs_operational": 3, 00:15:23.922 "base_bdevs_list": [ 00:15:23.922 { 00:15:23.922 "name": "BaseBdev1", 00:15:23.922 "uuid": "fca9bc5c-de1b-4bb3-b3f0-0699ff6c2e9b", 00:15:23.922 "is_configured": true, 00:15:23.922 "data_offset": 2048, 00:15:23.922 "data_size": 63488 00:15:23.922 }, 00:15:23.922 { 00:15:23.922 "name": "BaseBdev2", 00:15:23.922 "uuid": "e66250a7-f1bd-496c-8413-3e698fe359c8", 00:15:23.922 "is_configured": true, 00:15:23.922 "data_offset": 2048, 00:15:23.922 "data_size": 63488 00:15:23.922 }, 00:15:23.922 { 00:15:23.922 "name": "BaseBdev3", 00:15:23.922 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:23.922 "is_configured": false, 00:15:23.922 "data_offset": 0, 00:15:23.922 "data_size": 0 00:15:23.922 } 00:15:23.922 ] 00:15:23.922 }' 00:15:23.922 11:26:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:23.922 11:26:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:24.516 11:26:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:15:24.516 [2024-07-15 11:26:08.089402] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:24.516 [2024-07-15 11:26:08.089571] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x139e400 00:15:24.516 [2024-07-15 11:26:08.089585] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:24.516 [2024-07-15 11:26:08.089761] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x139def0 00:15:24.516 [2024-07-15 11:26:08.089875] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x139e400 00:15:24.516 [2024-07-15 11:26:08.089884] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x139e400 00:15:24.516 [2024-07-15 11:26:08.089982] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:24.516 BaseBdev3 00:15:24.516 11:26:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev3 00:15:24.516 11:26:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:15:24.516 11:26:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:15:24.516 11:26:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:15:24.516 11:26:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:15:24.516 11:26:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:15:24.516 11:26:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:15:24.772 11:26:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:15:25.029 [ 00:15:25.029 { 00:15:25.029 "name": "BaseBdev3", 00:15:25.029 "aliases": [ 00:15:25.029 "9c6f987e-a54a-4586-ac74-95f75fb6e0a3" 00:15:25.029 ], 00:15:25.029 "product_name": "Malloc disk", 00:15:25.029 "block_size": 512, 00:15:25.029 "num_blocks": 65536, 00:15:25.029 "uuid": "9c6f987e-a54a-4586-ac74-95f75fb6e0a3", 00:15:25.029 "assigned_rate_limits": { 00:15:25.029 "rw_ios_per_sec": 0, 00:15:25.029 "rw_mbytes_per_sec": 0, 00:15:25.029 "r_mbytes_per_sec": 0, 00:15:25.029 "w_mbytes_per_sec": 0 00:15:25.029 }, 00:15:25.029 "claimed": true, 00:15:25.029 "claim_type": "exclusive_write", 00:15:25.029 "zoned": false, 00:15:25.029 "supported_io_types": { 00:15:25.029 "read": true, 00:15:25.029 "write": true, 00:15:25.029 "unmap": true, 00:15:25.029 "flush": true, 00:15:25.029 "reset": true, 00:15:25.029 "nvme_admin": false, 00:15:25.029 "nvme_io": false, 00:15:25.029 "nvme_io_md": false, 00:15:25.029 "write_zeroes": true, 00:15:25.029 "zcopy": true, 00:15:25.029 "get_zone_info": false, 00:15:25.029 "zone_management": false, 00:15:25.029 "zone_append": false, 00:15:25.029 "compare": false, 00:15:25.029 "compare_and_write": false, 00:15:25.029 "abort": true, 00:15:25.029 "seek_hole": false, 00:15:25.029 "seek_data": false, 00:15:25.029 "copy": true, 00:15:25.029 "nvme_iov_md": false 00:15:25.029 }, 00:15:25.029 "memory_domains": [ 00:15:25.029 { 00:15:25.029 "dma_device_id": "system", 00:15:25.029 "dma_device_type": 1 00:15:25.029 }, 00:15:25.029 { 00:15:25.029 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:25.029 "dma_device_type": 2 00:15:25.029 } 00:15:25.029 ], 00:15:25.029 "driver_specific": {} 00:15:25.029 } 00:15:25.029 ] 00:15:25.029 11:26:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:15:25.029 11:26:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:15:25.029 11:26:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:15:25.029 11:26:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:15:25.029 11:26:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:25.029 11:26:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:15:25.029 11:26:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:25.029 11:26:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:25.029 11:26:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:25.029 11:26:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:25.029 11:26:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:25.029 11:26:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:25.029 11:26:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:25.029 11:26:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:25.029 11:26:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:25.285 11:26:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:25.285 "name": "Existed_Raid", 00:15:25.285 "uuid": "d62129ec-33b3-4d41-9611-2ce060790a4f", 00:15:25.285 "strip_size_kb": 64, 00:15:25.285 "state": "online", 00:15:25.285 "raid_level": "concat", 00:15:25.285 "superblock": true, 00:15:25.285 "num_base_bdevs": 3, 00:15:25.285 "num_base_bdevs_discovered": 3, 00:15:25.285 "num_base_bdevs_operational": 3, 00:15:25.285 "base_bdevs_list": [ 00:15:25.285 { 00:15:25.285 "name": "BaseBdev1", 00:15:25.285 "uuid": "fca9bc5c-de1b-4bb3-b3f0-0699ff6c2e9b", 00:15:25.285 "is_configured": true, 00:15:25.285 "data_offset": 2048, 00:15:25.285 "data_size": 63488 00:15:25.285 }, 00:15:25.285 { 00:15:25.285 "name": "BaseBdev2", 00:15:25.285 "uuid": "e66250a7-f1bd-496c-8413-3e698fe359c8", 00:15:25.285 "is_configured": true, 00:15:25.285 "data_offset": 2048, 00:15:25.285 "data_size": 63488 00:15:25.285 }, 00:15:25.285 { 00:15:25.285 "name": "BaseBdev3", 00:15:25.285 "uuid": "9c6f987e-a54a-4586-ac74-95f75fb6e0a3", 00:15:25.285 "is_configured": true, 00:15:25.285 "data_offset": 2048, 00:15:25.285 "data_size": 63488 00:15:25.285 } 00:15:25.285 ] 00:15:25.285 }' 00:15:25.285 11:26:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:25.285 11:26:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:25.851 11:26:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:15:25.851 11:26:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:15:25.851 11:26:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:15:25.851 11:26:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:15:25.851 11:26:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:15:25.851 11:26:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # local name 00:15:25.851 11:26:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:15:25.852 11:26:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:15:26.109 [2024-07-15 11:26:09.637812] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:26.109 11:26:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:15:26.109 "name": "Existed_Raid", 00:15:26.109 "aliases": [ 00:15:26.109 "d62129ec-33b3-4d41-9611-2ce060790a4f" 00:15:26.109 ], 00:15:26.109 "product_name": "Raid Volume", 00:15:26.109 "block_size": 512, 00:15:26.109 "num_blocks": 190464, 00:15:26.109 "uuid": "d62129ec-33b3-4d41-9611-2ce060790a4f", 00:15:26.109 "assigned_rate_limits": { 00:15:26.109 "rw_ios_per_sec": 0, 00:15:26.109 "rw_mbytes_per_sec": 0, 00:15:26.109 "r_mbytes_per_sec": 0, 00:15:26.109 "w_mbytes_per_sec": 0 00:15:26.109 }, 00:15:26.109 "claimed": false, 00:15:26.109 "zoned": false, 00:15:26.109 "supported_io_types": { 00:15:26.109 "read": true, 00:15:26.109 "write": true, 00:15:26.109 "unmap": true, 00:15:26.109 "flush": true, 00:15:26.109 "reset": true, 00:15:26.109 "nvme_admin": false, 00:15:26.109 "nvme_io": false, 00:15:26.109 "nvme_io_md": false, 00:15:26.109 "write_zeroes": true, 00:15:26.109 "zcopy": false, 00:15:26.109 "get_zone_info": false, 00:15:26.110 "zone_management": false, 00:15:26.110 "zone_append": false, 00:15:26.110 "compare": false, 00:15:26.110 "compare_and_write": false, 00:15:26.110 "abort": false, 00:15:26.110 "seek_hole": false, 00:15:26.110 "seek_data": false, 00:15:26.110 "copy": false, 00:15:26.110 "nvme_iov_md": false 00:15:26.110 }, 00:15:26.110 "memory_domains": [ 00:15:26.110 { 00:15:26.110 "dma_device_id": "system", 00:15:26.110 "dma_device_type": 1 00:15:26.110 }, 00:15:26.110 { 00:15:26.110 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:26.110 "dma_device_type": 2 00:15:26.110 }, 00:15:26.110 { 00:15:26.110 "dma_device_id": "system", 00:15:26.110 "dma_device_type": 1 00:15:26.110 }, 00:15:26.110 { 00:15:26.110 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:26.110 "dma_device_type": 2 00:15:26.110 }, 00:15:26.110 { 00:15:26.110 "dma_device_id": "system", 00:15:26.110 "dma_device_type": 1 00:15:26.110 }, 00:15:26.110 { 00:15:26.110 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:26.110 "dma_device_type": 2 00:15:26.110 } 00:15:26.110 ], 00:15:26.110 "driver_specific": { 00:15:26.110 "raid": { 00:15:26.110 "uuid": "d62129ec-33b3-4d41-9611-2ce060790a4f", 00:15:26.110 "strip_size_kb": 64, 00:15:26.110 "state": "online", 00:15:26.110 "raid_level": "concat", 00:15:26.110 "superblock": true, 00:15:26.110 "num_base_bdevs": 3, 00:15:26.110 "num_base_bdevs_discovered": 3, 00:15:26.110 "num_base_bdevs_operational": 3, 00:15:26.110 "base_bdevs_list": [ 00:15:26.110 { 00:15:26.110 "name": "BaseBdev1", 00:15:26.110 "uuid": "fca9bc5c-de1b-4bb3-b3f0-0699ff6c2e9b", 00:15:26.110 "is_configured": true, 00:15:26.110 "data_offset": 2048, 00:15:26.110 "data_size": 63488 00:15:26.110 }, 00:15:26.110 { 00:15:26.110 "name": "BaseBdev2", 00:15:26.110 "uuid": "e66250a7-f1bd-496c-8413-3e698fe359c8", 00:15:26.110 "is_configured": true, 00:15:26.110 "data_offset": 2048, 00:15:26.110 "data_size": 63488 00:15:26.110 }, 00:15:26.110 { 00:15:26.110 "name": "BaseBdev3", 00:15:26.110 "uuid": "9c6f987e-a54a-4586-ac74-95f75fb6e0a3", 00:15:26.110 "is_configured": true, 00:15:26.110 "data_offset": 2048, 00:15:26.110 "data_size": 63488 00:15:26.110 } 00:15:26.110 ] 00:15:26.110 } 00:15:26.110 } 00:15:26.110 }' 00:15:26.110 11:26:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:26.368 11:26:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:15:26.368 BaseBdev2 00:15:26.368 BaseBdev3' 00:15:26.368 11:26:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:15:26.368 11:26:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:15:26.368 11:26:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:15:26.368 11:26:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:15:26.368 "name": "BaseBdev1", 00:15:26.368 "aliases": [ 00:15:26.368 "fca9bc5c-de1b-4bb3-b3f0-0699ff6c2e9b" 00:15:26.368 ], 00:15:26.368 "product_name": "Malloc disk", 00:15:26.368 "block_size": 512, 00:15:26.368 "num_blocks": 65536, 00:15:26.368 "uuid": "fca9bc5c-de1b-4bb3-b3f0-0699ff6c2e9b", 00:15:26.368 "assigned_rate_limits": { 00:15:26.368 "rw_ios_per_sec": 0, 00:15:26.368 "rw_mbytes_per_sec": 0, 00:15:26.368 "r_mbytes_per_sec": 0, 00:15:26.368 "w_mbytes_per_sec": 0 00:15:26.368 }, 00:15:26.368 "claimed": true, 00:15:26.368 "claim_type": "exclusive_write", 00:15:26.368 "zoned": false, 00:15:26.368 "supported_io_types": { 00:15:26.368 "read": true, 00:15:26.368 "write": true, 00:15:26.368 "unmap": true, 00:15:26.368 "flush": true, 00:15:26.368 "reset": true, 00:15:26.368 "nvme_admin": false, 00:15:26.368 "nvme_io": false, 00:15:26.368 "nvme_io_md": false, 00:15:26.368 "write_zeroes": true, 00:15:26.368 "zcopy": true, 00:15:26.368 "get_zone_info": false, 00:15:26.368 "zone_management": false, 00:15:26.368 "zone_append": false, 00:15:26.368 "compare": false, 00:15:26.368 "compare_and_write": false, 00:15:26.368 "abort": true, 00:15:26.368 "seek_hole": false, 00:15:26.368 "seek_data": false, 00:15:26.368 "copy": true, 00:15:26.368 "nvme_iov_md": false 00:15:26.368 }, 00:15:26.368 "memory_domains": [ 00:15:26.368 { 00:15:26.368 "dma_device_id": "system", 00:15:26.368 "dma_device_type": 1 00:15:26.368 }, 00:15:26.368 { 00:15:26.368 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:26.368 "dma_device_type": 2 00:15:26.368 } 00:15:26.368 ], 00:15:26.368 "driver_specific": {} 00:15:26.368 }' 00:15:26.368 11:26:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:26.627 11:26:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:26.627 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:15:26.627 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:26.627 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:26.627 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:15:26.627 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:26.884 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:26.884 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:15:26.884 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:26.884 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:26.884 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:15:26.884 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:15:26.884 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:15:26.884 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:15:27.142 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:15:27.142 "name": "BaseBdev2", 00:15:27.142 "aliases": [ 00:15:27.142 "e66250a7-f1bd-496c-8413-3e698fe359c8" 00:15:27.142 ], 00:15:27.142 "product_name": "Malloc disk", 00:15:27.142 "block_size": 512, 00:15:27.142 "num_blocks": 65536, 00:15:27.142 "uuid": "e66250a7-f1bd-496c-8413-3e698fe359c8", 00:15:27.142 "assigned_rate_limits": { 00:15:27.142 "rw_ios_per_sec": 0, 00:15:27.142 "rw_mbytes_per_sec": 0, 00:15:27.142 "r_mbytes_per_sec": 0, 00:15:27.142 "w_mbytes_per_sec": 0 00:15:27.142 }, 00:15:27.142 "claimed": true, 00:15:27.142 "claim_type": "exclusive_write", 00:15:27.142 "zoned": false, 00:15:27.142 "supported_io_types": { 00:15:27.142 "read": true, 00:15:27.142 "write": true, 00:15:27.142 "unmap": true, 00:15:27.142 "flush": true, 00:15:27.142 "reset": true, 00:15:27.142 "nvme_admin": false, 00:15:27.142 "nvme_io": false, 00:15:27.142 "nvme_io_md": false, 00:15:27.142 "write_zeroes": true, 00:15:27.142 "zcopy": true, 00:15:27.142 "get_zone_info": false, 00:15:27.142 "zone_management": false, 00:15:27.142 "zone_append": false, 00:15:27.142 "compare": false, 00:15:27.142 "compare_and_write": false, 00:15:27.142 "abort": true, 00:15:27.142 "seek_hole": false, 00:15:27.142 "seek_data": false, 00:15:27.142 "copy": true, 00:15:27.142 "nvme_iov_md": false 00:15:27.142 }, 00:15:27.142 "memory_domains": [ 00:15:27.142 { 00:15:27.142 "dma_device_id": "system", 00:15:27.142 "dma_device_type": 1 00:15:27.142 }, 00:15:27.142 { 00:15:27.142 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:27.142 "dma_device_type": 2 00:15:27.142 } 00:15:27.142 ], 00:15:27.142 "driver_specific": {} 00:15:27.142 }' 00:15:27.142 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:27.142 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:27.142 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:15:27.142 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:27.413 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:27.413 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:15:27.413 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:27.413 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:27.413 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:15:27.413 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:27.413 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:27.413 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:15:27.413 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:15:27.413 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:15:27.413 11:26:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:15:27.673 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:15:27.673 "name": "BaseBdev3", 00:15:27.673 "aliases": [ 00:15:27.673 "9c6f987e-a54a-4586-ac74-95f75fb6e0a3" 00:15:27.673 ], 00:15:27.673 "product_name": "Malloc disk", 00:15:27.673 "block_size": 512, 00:15:27.673 "num_blocks": 65536, 00:15:27.673 "uuid": "9c6f987e-a54a-4586-ac74-95f75fb6e0a3", 00:15:27.673 "assigned_rate_limits": { 00:15:27.673 "rw_ios_per_sec": 0, 00:15:27.673 "rw_mbytes_per_sec": 0, 00:15:27.673 "r_mbytes_per_sec": 0, 00:15:27.673 "w_mbytes_per_sec": 0 00:15:27.673 }, 00:15:27.673 "claimed": true, 00:15:27.673 "claim_type": "exclusive_write", 00:15:27.673 "zoned": false, 00:15:27.673 "supported_io_types": { 00:15:27.673 "read": true, 00:15:27.673 "write": true, 00:15:27.673 "unmap": true, 00:15:27.673 "flush": true, 00:15:27.673 "reset": true, 00:15:27.673 "nvme_admin": false, 00:15:27.673 "nvme_io": false, 00:15:27.673 "nvme_io_md": false, 00:15:27.673 "write_zeroes": true, 00:15:27.673 "zcopy": true, 00:15:27.673 "get_zone_info": false, 00:15:27.673 "zone_management": false, 00:15:27.673 "zone_append": false, 00:15:27.673 "compare": false, 00:15:27.673 "compare_and_write": false, 00:15:27.673 "abort": true, 00:15:27.673 "seek_hole": false, 00:15:27.673 "seek_data": false, 00:15:27.673 "copy": true, 00:15:27.673 "nvme_iov_md": false 00:15:27.673 }, 00:15:27.673 "memory_domains": [ 00:15:27.673 { 00:15:27.673 "dma_device_id": "system", 00:15:27.673 "dma_device_type": 1 00:15:27.673 }, 00:15:27.673 { 00:15:27.673 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:27.673 "dma_device_type": 2 00:15:27.673 } 00:15:27.673 ], 00:15:27.673 "driver_specific": {} 00:15:27.673 }' 00:15:27.673 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:27.673 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:27.673 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:15:27.673 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:27.931 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:27.931 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:15:27.931 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:27.931 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:27.931 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:15:27.931 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:27.931 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:27.931 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:15:27.931 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:15:28.189 [2024-07-15 11:26:11.703055] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:28.189 [2024-07-15 11:26:11.703085] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:28.189 [2024-07-15 11:26:11.703128] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:28.189 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@275 -- # local expected_state 00:15:28.189 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # has_redundancy concat 00:15:28.189 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # case $1 in 00:15:28.189 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # return 1 00:15:28.189 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@277 -- # expected_state=offline 00:15:28.189 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:15:28.189 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:28.189 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=offline 00:15:28.189 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:28.189 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:28.189 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:15:28.189 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:28.189 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:28.189 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:28.189 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:28.189 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:28.189 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:28.446 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:28.446 "name": "Existed_Raid", 00:15:28.446 "uuid": "d62129ec-33b3-4d41-9611-2ce060790a4f", 00:15:28.446 "strip_size_kb": 64, 00:15:28.446 "state": "offline", 00:15:28.446 "raid_level": "concat", 00:15:28.446 "superblock": true, 00:15:28.446 "num_base_bdevs": 3, 00:15:28.446 "num_base_bdevs_discovered": 2, 00:15:28.446 "num_base_bdevs_operational": 2, 00:15:28.446 "base_bdevs_list": [ 00:15:28.446 { 00:15:28.446 "name": null, 00:15:28.446 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:28.446 "is_configured": false, 00:15:28.446 "data_offset": 2048, 00:15:28.446 "data_size": 63488 00:15:28.446 }, 00:15:28.446 { 00:15:28.446 "name": "BaseBdev2", 00:15:28.446 "uuid": "e66250a7-f1bd-496c-8413-3e698fe359c8", 00:15:28.446 "is_configured": true, 00:15:28.446 "data_offset": 2048, 00:15:28.446 "data_size": 63488 00:15:28.446 }, 00:15:28.446 { 00:15:28.446 "name": "BaseBdev3", 00:15:28.446 "uuid": "9c6f987e-a54a-4586-ac74-95f75fb6e0a3", 00:15:28.446 "is_configured": true, 00:15:28.446 "data_offset": 2048, 00:15:28.446 "data_size": 63488 00:15:28.446 } 00:15:28.446 ] 00:15:28.446 }' 00:15:28.446 11:26:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:28.446 11:26:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:29.014 11:26:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:15:29.014 11:26:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:15:29.014 11:26:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:29.014 11:26:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:15:29.273 11:26:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:15:29.273 11:26:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:29.273 11:26:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:15:29.531 [2024-07-15 11:26:13.040693] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:29.531 11:26:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:15:29.531 11:26:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:15:29.531 11:26:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:29.531 11:26:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:15:29.803 11:26:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:15:29.803 11:26:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:29.803 11:26:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev3 00:15:30.071 [2024-07-15 11:26:13.534379] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:15:30.071 [2024-07-15 11:26:13.534429] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x139e400 name Existed_Raid, state offline 00:15:30.071 11:26:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:15:30.071 11:26:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:15:30.071 11:26:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:15:30.071 11:26:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:30.329 11:26:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:15:30.329 11:26:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:15:30.329 11:26:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # '[' 3 -gt 2 ']' 00:15:30.329 11:26:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i = 1 )) 00:15:30.329 11:26:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:15:30.329 11:26:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:15:30.587 BaseBdev2 00:15:30.587 11:26:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev2 00:15:30.587 11:26:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:15:30.587 11:26:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:15:30.587 11:26:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:15:30.587 11:26:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:15:30.588 11:26:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:15:30.588 11:26:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:15:30.846 11:26:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:31.110 [ 00:15:31.110 { 00:15:31.110 "name": "BaseBdev2", 00:15:31.110 "aliases": [ 00:15:31.110 "b2417fa7-1966-42f0-9987-b6d00431bd93" 00:15:31.110 ], 00:15:31.110 "product_name": "Malloc disk", 00:15:31.110 "block_size": 512, 00:15:31.110 "num_blocks": 65536, 00:15:31.110 "uuid": "b2417fa7-1966-42f0-9987-b6d00431bd93", 00:15:31.110 "assigned_rate_limits": { 00:15:31.110 "rw_ios_per_sec": 0, 00:15:31.110 "rw_mbytes_per_sec": 0, 00:15:31.110 "r_mbytes_per_sec": 0, 00:15:31.110 "w_mbytes_per_sec": 0 00:15:31.110 }, 00:15:31.110 "claimed": false, 00:15:31.110 "zoned": false, 00:15:31.110 "supported_io_types": { 00:15:31.110 "read": true, 00:15:31.110 "write": true, 00:15:31.110 "unmap": true, 00:15:31.110 "flush": true, 00:15:31.110 "reset": true, 00:15:31.110 "nvme_admin": false, 00:15:31.110 "nvme_io": false, 00:15:31.110 "nvme_io_md": false, 00:15:31.110 "write_zeroes": true, 00:15:31.110 "zcopy": true, 00:15:31.110 "get_zone_info": false, 00:15:31.110 "zone_management": false, 00:15:31.110 "zone_append": false, 00:15:31.110 "compare": false, 00:15:31.110 "compare_and_write": false, 00:15:31.110 "abort": true, 00:15:31.110 "seek_hole": false, 00:15:31.110 "seek_data": false, 00:15:31.110 "copy": true, 00:15:31.110 "nvme_iov_md": false 00:15:31.110 }, 00:15:31.110 "memory_domains": [ 00:15:31.110 { 00:15:31.110 "dma_device_id": "system", 00:15:31.110 "dma_device_type": 1 00:15:31.110 }, 00:15:31.110 { 00:15:31.110 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:31.110 "dma_device_type": 2 00:15:31.110 } 00:15:31.110 ], 00:15:31.110 "driver_specific": {} 00:15:31.110 } 00:15:31.110 ] 00:15:31.110 11:26:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:15:31.110 11:26:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:15:31.110 11:26:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:15:31.110 11:26:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:15:31.368 BaseBdev3 00:15:31.368 11:26:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev3 00:15:31.368 11:26:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:15:31.368 11:26:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:15:31.368 11:26:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:15:31.368 11:26:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:15:31.368 11:26:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:15:31.368 11:26:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:15:31.627 11:26:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:15:31.885 [ 00:15:31.885 { 00:15:31.885 "name": "BaseBdev3", 00:15:31.885 "aliases": [ 00:15:31.885 "fa343c90-ced5-49bd-ab13-fe669d6c7579" 00:15:31.885 ], 00:15:31.885 "product_name": "Malloc disk", 00:15:31.885 "block_size": 512, 00:15:31.885 "num_blocks": 65536, 00:15:31.885 "uuid": "fa343c90-ced5-49bd-ab13-fe669d6c7579", 00:15:31.885 "assigned_rate_limits": { 00:15:31.885 "rw_ios_per_sec": 0, 00:15:31.885 "rw_mbytes_per_sec": 0, 00:15:31.885 "r_mbytes_per_sec": 0, 00:15:31.885 "w_mbytes_per_sec": 0 00:15:31.885 }, 00:15:31.885 "claimed": false, 00:15:31.885 "zoned": false, 00:15:31.885 "supported_io_types": { 00:15:31.885 "read": true, 00:15:31.885 "write": true, 00:15:31.885 "unmap": true, 00:15:31.885 "flush": true, 00:15:31.885 "reset": true, 00:15:31.885 "nvme_admin": false, 00:15:31.885 "nvme_io": false, 00:15:31.885 "nvme_io_md": false, 00:15:31.885 "write_zeroes": true, 00:15:31.885 "zcopy": true, 00:15:31.885 "get_zone_info": false, 00:15:31.885 "zone_management": false, 00:15:31.885 "zone_append": false, 00:15:31.885 "compare": false, 00:15:31.885 "compare_and_write": false, 00:15:31.885 "abort": true, 00:15:31.885 "seek_hole": false, 00:15:31.885 "seek_data": false, 00:15:31.885 "copy": true, 00:15:31.885 "nvme_iov_md": false 00:15:31.885 }, 00:15:31.885 "memory_domains": [ 00:15:31.885 { 00:15:31.885 "dma_device_id": "system", 00:15:31.885 "dma_device_type": 1 00:15:31.885 }, 00:15:31.885 { 00:15:31.885 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:31.885 "dma_device_type": 2 00:15:31.885 } 00:15:31.885 ], 00:15:31.885 "driver_specific": {} 00:15:31.885 } 00:15:31.885 ] 00:15:31.885 11:26:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:15:31.885 11:26:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:15:31.885 11:26:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:15:31.885 11:26:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@305 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r concat -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:15:31.885 [2024-07-15 11:26:15.478128] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:31.885 [2024-07-15 11:26:15.478173] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:31.885 [2024-07-15 11:26:15.478191] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:31.885 [2024-07-15 11:26:15.479530] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:32.144 11:26:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:15:32.144 11:26:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:32.144 11:26:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:15:32.144 11:26:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:32.144 11:26:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:32.144 11:26:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:32.144 11:26:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:32.144 11:26:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:32.144 11:26:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:32.144 11:26:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:32.144 11:26:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:32.144 11:26:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:32.402 11:26:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:32.402 "name": "Existed_Raid", 00:15:32.402 "uuid": "82c5ad70-b465-4435-8c2c-c3ae777041b9", 00:15:32.402 "strip_size_kb": 64, 00:15:32.402 "state": "configuring", 00:15:32.402 "raid_level": "concat", 00:15:32.402 "superblock": true, 00:15:32.402 "num_base_bdevs": 3, 00:15:32.402 "num_base_bdevs_discovered": 2, 00:15:32.402 "num_base_bdevs_operational": 3, 00:15:32.402 "base_bdevs_list": [ 00:15:32.402 { 00:15:32.402 "name": "BaseBdev1", 00:15:32.402 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:32.402 "is_configured": false, 00:15:32.402 "data_offset": 0, 00:15:32.402 "data_size": 0 00:15:32.402 }, 00:15:32.402 { 00:15:32.402 "name": "BaseBdev2", 00:15:32.402 "uuid": "b2417fa7-1966-42f0-9987-b6d00431bd93", 00:15:32.402 "is_configured": true, 00:15:32.402 "data_offset": 2048, 00:15:32.402 "data_size": 63488 00:15:32.402 }, 00:15:32.402 { 00:15:32.402 "name": "BaseBdev3", 00:15:32.402 "uuid": "fa343c90-ced5-49bd-ab13-fe669d6c7579", 00:15:32.402 "is_configured": true, 00:15:32.402 "data_offset": 2048, 00:15:32.402 "data_size": 63488 00:15:32.402 } 00:15:32.402 ] 00:15:32.402 }' 00:15:32.402 11:26:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:32.402 11:26:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:32.966 11:26:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev2 00:15:32.966 [2024-07-15 11:26:16.468725] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:32.966 11:26:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@309 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:15:32.966 11:26:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:32.966 11:26:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:15:32.966 11:26:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:32.966 11:26:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:32.966 11:26:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:32.966 11:26:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:32.966 11:26:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:32.966 11:26:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:32.966 11:26:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:32.966 11:26:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:32.967 11:26:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:33.225 11:26:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:33.225 "name": "Existed_Raid", 00:15:33.225 "uuid": "82c5ad70-b465-4435-8c2c-c3ae777041b9", 00:15:33.225 "strip_size_kb": 64, 00:15:33.225 "state": "configuring", 00:15:33.225 "raid_level": "concat", 00:15:33.225 "superblock": true, 00:15:33.225 "num_base_bdevs": 3, 00:15:33.225 "num_base_bdevs_discovered": 1, 00:15:33.225 "num_base_bdevs_operational": 3, 00:15:33.225 "base_bdevs_list": [ 00:15:33.225 { 00:15:33.225 "name": "BaseBdev1", 00:15:33.225 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:33.225 "is_configured": false, 00:15:33.225 "data_offset": 0, 00:15:33.225 "data_size": 0 00:15:33.225 }, 00:15:33.225 { 00:15:33.225 "name": null, 00:15:33.225 "uuid": "b2417fa7-1966-42f0-9987-b6d00431bd93", 00:15:33.225 "is_configured": false, 00:15:33.225 "data_offset": 2048, 00:15:33.225 "data_size": 63488 00:15:33.225 }, 00:15:33.225 { 00:15:33.225 "name": "BaseBdev3", 00:15:33.225 "uuid": "fa343c90-ced5-49bd-ab13-fe669d6c7579", 00:15:33.225 "is_configured": true, 00:15:33.225 "data_offset": 2048, 00:15:33.225 "data_size": 63488 00:15:33.225 } 00:15:33.225 ] 00:15:33.225 }' 00:15:33.225 11:26:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:33.225 11:26:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:33.792 11:26:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:33.792 11:26:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:15:34.077 11:26:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # [[ false == \f\a\l\s\e ]] 00:15:34.077 11:26:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:15:34.335 [2024-07-15 11:26:17.759472] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:34.335 BaseBdev1 00:15:34.335 11:26:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@313 -- # waitforbdev BaseBdev1 00:15:34.335 11:26:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:15:34.335 11:26:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:15:34.335 11:26:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:15:34.335 11:26:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:15:34.335 11:26:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:15:34.335 11:26:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:15:34.593 11:26:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:34.852 [ 00:15:34.852 { 00:15:34.852 "name": "BaseBdev1", 00:15:34.852 "aliases": [ 00:15:34.852 "ae763fb0-db7e-4831-a3f3-491303dec5a8" 00:15:34.852 ], 00:15:34.852 "product_name": "Malloc disk", 00:15:34.852 "block_size": 512, 00:15:34.852 "num_blocks": 65536, 00:15:34.852 "uuid": "ae763fb0-db7e-4831-a3f3-491303dec5a8", 00:15:34.852 "assigned_rate_limits": { 00:15:34.852 "rw_ios_per_sec": 0, 00:15:34.852 "rw_mbytes_per_sec": 0, 00:15:34.852 "r_mbytes_per_sec": 0, 00:15:34.852 "w_mbytes_per_sec": 0 00:15:34.852 }, 00:15:34.852 "claimed": true, 00:15:34.852 "claim_type": "exclusive_write", 00:15:34.852 "zoned": false, 00:15:34.852 "supported_io_types": { 00:15:34.852 "read": true, 00:15:34.852 "write": true, 00:15:34.852 "unmap": true, 00:15:34.852 "flush": true, 00:15:34.852 "reset": true, 00:15:34.852 "nvme_admin": false, 00:15:34.852 "nvme_io": false, 00:15:34.852 "nvme_io_md": false, 00:15:34.852 "write_zeroes": true, 00:15:34.852 "zcopy": true, 00:15:34.852 "get_zone_info": false, 00:15:34.852 "zone_management": false, 00:15:34.852 "zone_append": false, 00:15:34.852 "compare": false, 00:15:34.852 "compare_and_write": false, 00:15:34.852 "abort": true, 00:15:34.852 "seek_hole": false, 00:15:34.852 "seek_data": false, 00:15:34.852 "copy": true, 00:15:34.852 "nvme_iov_md": false 00:15:34.852 }, 00:15:34.852 "memory_domains": [ 00:15:34.852 { 00:15:34.852 "dma_device_id": "system", 00:15:34.852 "dma_device_type": 1 00:15:34.852 }, 00:15:34.852 { 00:15:34.852 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:34.852 "dma_device_type": 2 00:15:34.852 } 00:15:34.852 ], 00:15:34.852 "driver_specific": {} 00:15:34.852 } 00:15:34.852 ] 00:15:34.852 11:26:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:15:34.852 11:26:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:15:34.852 11:26:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:34.852 11:26:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:15:34.852 11:26:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:34.852 11:26:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:34.852 11:26:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:34.852 11:26:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:34.852 11:26:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:34.852 11:26:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:34.852 11:26:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:34.852 11:26:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:34.852 11:26:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:35.111 11:26:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:35.111 "name": "Existed_Raid", 00:15:35.111 "uuid": "82c5ad70-b465-4435-8c2c-c3ae777041b9", 00:15:35.111 "strip_size_kb": 64, 00:15:35.111 "state": "configuring", 00:15:35.111 "raid_level": "concat", 00:15:35.111 "superblock": true, 00:15:35.111 "num_base_bdevs": 3, 00:15:35.111 "num_base_bdevs_discovered": 2, 00:15:35.111 "num_base_bdevs_operational": 3, 00:15:35.111 "base_bdevs_list": [ 00:15:35.111 { 00:15:35.111 "name": "BaseBdev1", 00:15:35.111 "uuid": "ae763fb0-db7e-4831-a3f3-491303dec5a8", 00:15:35.111 "is_configured": true, 00:15:35.111 "data_offset": 2048, 00:15:35.111 "data_size": 63488 00:15:35.111 }, 00:15:35.111 { 00:15:35.111 "name": null, 00:15:35.111 "uuid": "b2417fa7-1966-42f0-9987-b6d00431bd93", 00:15:35.111 "is_configured": false, 00:15:35.111 "data_offset": 2048, 00:15:35.111 "data_size": 63488 00:15:35.111 }, 00:15:35.111 { 00:15:35.111 "name": "BaseBdev3", 00:15:35.111 "uuid": "fa343c90-ced5-49bd-ab13-fe669d6c7579", 00:15:35.111 "is_configured": true, 00:15:35.111 "data_offset": 2048, 00:15:35.111 "data_size": 63488 00:15:35.111 } 00:15:35.111 ] 00:15:35.112 }' 00:15:35.112 11:26:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:35.112 11:26:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.678 11:26:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:35.678 11:26:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:15:35.678 11:26:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # [[ true == \t\r\u\e ]] 00:15:35.678 11:26:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@317 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev3 00:15:35.938 [2024-07-15 11:26:19.472050] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:15:35.938 11:26:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:15:35.938 11:26:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:35.938 11:26:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:15:35.938 11:26:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:35.938 11:26:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:35.938 11:26:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:35.938 11:26:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:35.938 11:26:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:35.938 11:26:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:35.938 11:26:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:35.938 11:26:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:35.938 11:26:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:36.196 11:26:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:36.196 "name": "Existed_Raid", 00:15:36.197 "uuid": "82c5ad70-b465-4435-8c2c-c3ae777041b9", 00:15:36.197 "strip_size_kb": 64, 00:15:36.197 "state": "configuring", 00:15:36.197 "raid_level": "concat", 00:15:36.197 "superblock": true, 00:15:36.197 "num_base_bdevs": 3, 00:15:36.197 "num_base_bdevs_discovered": 1, 00:15:36.197 "num_base_bdevs_operational": 3, 00:15:36.197 "base_bdevs_list": [ 00:15:36.197 { 00:15:36.197 "name": "BaseBdev1", 00:15:36.197 "uuid": "ae763fb0-db7e-4831-a3f3-491303dec5a8", 00:15:36.197 "is_configured": true, 00:15:36.197 "data_offset": 2048, 00:15:36.197 "data_size": 63488 00:15:36.197 }, 00:15:36.197 { 00:15:36.197 "name": null, 00:15:36.197 "uuid": "b2417fa7-1966-42f0-9987-b6d00431bd93", 00:15:36.197 "is_configured": false, 00:15:36.197 "data_offset": 2048, 00:15:36.197 "data_size": 63488 00:15:36.197 }, 00:15:36.197 { 00:15:36.197 "name": null, 00:15:36.197 "uuid": "fa343c90-ced5-49bd-ab13-fe669d6c7579", 00:15:36.197 "is_configured": false, 00:15:36.197 "data_offset": 2048, 00:15:36.197 "data_size": 63488 00:15:36.197 } 00:15:36.197 ] 00:15:36.197 }' 00:15:36.197 11:26:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:36.197 11:26:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:37.130 11:26:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:37.130 11:26:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:15:37.389 11:26:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # [[ false == \f\a\l\s\e ]] 00:15:37.389 11:26:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:15:37.389 [2024-07-15 11:26:20.943987] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:37.389 11:26:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@322 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:15:37.389 11:26:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:37.389 11:26:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:15:37.389 11:26:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:37.389 11:26:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:37.389 11:26:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:37.389 11:26:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:37.389 11:26:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:37.389 11:26:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:37.389 11:26:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:37.389 11:26:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:37.389 11:26:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:37.650 11:26:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:37.650 "name": "Existed_Raid", 00:15:37.650 "uuid": "82c5ad70-b465-4435-8c2c-c3ae777041b9", 00:15:37.650 "strip_size_kb": 64, 00:15:37.650 "state": "configuring", 00:15:37.650 "raid_level": "concat", 00:15:37.650 "superblock": true, 00:15:37.650 "num_base_bdevs": 3, 00:15:37.650 "num_base_bdevs_discovered": 2, 00:15:37.650 "num_base_bdevs_operational": 3, 00:15:37.650 "base_bdevs_list": [ 00:15:37.650 { 00:15:37.650 "name": "BaseBdev1", 00:15:37.650 "uuid": "ae763fb0-db7e-4831-a3f3-491303dec5a8", 00:15:37.650 "is_configured": true, 00:15:37.650 "data_offset": 2048, 00:15:37.650 "data_size": 63488 00:15:37.650 }, 00:15:37.650 { 00:15:37.650 "name": null, 00:15:37.650 "uuid": "b2417fa7-1966-42f0-9987-b6d00431bd93", 00:15:37.650 "is_configured": false, 00:15:37.650 "data_offset": 2048, 00:15:37.650 "data_size": 63488 00:15:37.650 }, 00:15:37.650 { 00:15:37.650 "name": "BaseBdev3", 00:15:37.650 "uuid": "fa343c90-ced5-49bd-ab13-fe669d6c7579", 00:15:37.650 "is_configured": true, 00:15:37.650 "data_offset": 2048, 00:15:37.650 "data_size": 63488 00:15:37.650 } 00:15:37.650 ] 00:15:37.650 }' 00:15:37.650 11:26:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:37.650 11:26:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:38.261 11:26:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:38.261 11:26:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:15:38.520 11:26:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # [[ true == \t\r\u\e ]] 00:15:38.520 11:26:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@325 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:15:39.086 [2024-07-15 11:26:22.532200] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:39.086 11:26:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:15:39.086 11:26:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:39.086 11:26:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:15:39.086 11:26:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:39.086 11:26:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:39.086 11:26:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:39.086 11:26:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:39.086 11:26:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:39.086 11:26:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:39.086 11:26:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:39.086 11:26:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:39.086 11:26:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:39.344 11:26:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:39.344 "name": "Existed_Raid", 00:15:39.344 "uuid": "82c5ad70-b465-4435-8c2c-c3ae777041b9", 00:15:39.344 "strip_size_kb": 64, 00:15:39.344 "state": "configuring", 00:15:39.344 "raid_level": "concat", 00:15:39.344 "superblock": true, 00:15:39.344 "num_base_bdevs": 3, 00:15:39.344 "num_base_bdevs_discovered": 1, 00:15:39.344 "num_base_bdevs_operational": 3, 00:15:39.344 "base_bdevs_list": [ 00:15:39.344 { 00:15:39.344 "name": null, 00:15:39.344 "uuid": "ae763fb0-db7e-4831-a3f3-491303dec5a8", 00:15:39.344 "is_configured": false, 00:15:39.344 "data_offset": 2048, 00:15:39.344 "data_size": 63488 00:15:39.344 }, 00:15:39.344 { 00:15:39.344 "name": null, 00:15:39.344 "uuid": "b2417fa7-1966-42f0-9987-b6d00431bd93", 00:15:39.344 "is_configured": false, 00:15:39.344 "data_offset": 2048, 00:15:39.344 "data_size": 63488 00:15:39.344 }, 00:15:39.344 { 00:15:39.344 "name": "BaseBdev3", 00:15:39.344 "uuid": "fa343c90-ced5-49bd-ab13-fe669d6c7579", 00:15:39.344 "is_configured": true, 00:15:39.344 "data_offset": 2048, 00:15:39.344 "data_size": 63488 00:15:39.344 } 00:15:39.344 ] 00:15:39.344 }' 00:15:39.344 11:26:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:39.344 11:26:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:39.910 11:26:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@327 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:39.910 11:26:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@327 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:15:40.168 11:26:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@327 -- # [[ false == \f\a\l\s\e ]] 00:15:40.168 11:26:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@329 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:15:40.735 [2024-07-15 11:26:24.159029] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:40.735 11:26:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@330 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:15:40.735 11:26:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:40.735 11:26:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:15:40.735 11:26:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:40.735 11:26:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:40.735 11:26:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:40.735 11:26:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:40.735 11:26:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:40.735 11:26:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:40.735 11:26:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:40.735 11:26:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:40.735 11:26:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:40.994 11:26:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:40.994 "name": "Existed_Raid", 00:15:40.994 "uuid": "82c5ad70-b465-4435-8c2c-c3ae777041b9", 00:15:40.994 "strip_size_kb": 64, 00:15:40.994 "state": "configuring", 00:15:40.994 "raid_level": "concat", 00:15:40.994 "superblock": true, 00:15:40.994 "num_base_bdevs": 3, 00:15:40.994 "num_base_bdevs_discovered": 2, 00:15:40.994 "num_base_bdevs_operational": 3, 00:15:40.994 "base_bdevs_list": [ 00:15:40.994 { 00:15:40.994 "name": null, 00:15:40.994 "uuid": "ae763fb0-db7e-4831-a3f3-491303dec5a8", 00:15:40.994 "is_configured": false, 00:15:40.994 "data_offset": 2048, 00:15:40.994 "data_size": 63488 00:15:40.994 }, 00:15:40.994 { 00:15:40.994 "name": "BaseBdev2", 00:15:40.994 "uuid": "b2417fa7-1966-42f0-9987-b6d00431bd93", 00:15:40.994 "is_configured": true, 00:15:40.994 "data_offset": 2048, 00:15:40.994 "data_size": 63488 00:15:40.994 }, 00:15:40.994 { 00:15:40.994 "name": "BaseBdev3", 00:15:40.994 "uuid": "fa343c90-ced5-49bd-ab13-fe669d6c7579", 00:15:40.994 "is_configured": true, 00:15:40.994 "data_offset": 2048, 00:15:40.994 "data_size": 63488 00:15:40.994 } 00:15:40.994 ] 00:15:40.994 }' 00:15:40.994 11:26:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:40.994 11:26:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:41.559 11:26:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@331 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:41.559 11:26:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@331 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:15:41.816 11:26:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@331 -- # [[ true == \t\r\u\e ]] 00:15:41.816 11:26:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:41.816 11:26:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@333 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:15:42.073 11:26:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b NewBaseBdev -u ae763fb0-db7e-4831-a3f3-491303dec5a8 00:15:42.331 [2024-07-15 11:26:25.747802] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:15:42.331 [2024-07-15 11:26:25.747969] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x139cf50 00:15:42.331 [2024-07-15 11:26:25.747983] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:42.331 [2024-07-15 11:26:25.748161] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x10a3940 00:15:42.331 [2024-07-15 11:26:25.748275] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x139cf50 00:15:42.331 [2024-07-15 11:26:25.748284] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x139cf50 00:15:42.331 [2024-07-15 11:26:25.748376] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:42.331 NewBaseBdev 00:15:42.331 11:26:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@334 -- # waitforbdev NewBaseBdev 00:15:42.331 11:26:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=NewBaseBdev 00:15:42.331 11:26:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:15:42.331 11:26:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:15:42.331 11:26:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:15:42.331 11:26:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:15:42.331 11:26:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:15:42.589 11:26:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev -t 2000 00:15:42.846 [ 00:15:42.846 { 00:15:42.846 "name": "NewBaseBdev", 00:15:42.846 "aliases": [ 00:15:42.846 "ae763fb0-db7e-4831-a3f3-491303dec5a8" 00:15:42.846 ], 00:15:42.846 "product_name": "Malloc disk", 00:15:42.846 "block_size": 512, 00:15:42.846 "num_blocks": 65536, 00:15:42.846 "uuid": "ae763fb0-db7e-4831-a3f3-491303dec5a8", 00:15:42.846 "assigned_rate_limits": { 00:15:42.846 "rw_ios_per_sec": 0, 00:15:42.846 "rw_mbytes_per_sec": 0, 00:15:42.846 "r_mbytes_per_sec": 0, 00:15:42.846 "w_mbytes_per_sec": 0 00:15:42.846 }, 00:15:42.846 "claimed": true, 00:15:42.846 "claim_type": "exclusive_write", 00:15:42.846 "zoned": false, 00:15:42.846 "supported_io_types": { 00:15:42.846 "read": true, 00:15:42.846 "write": true, 00:15:42.846 "unmap": true, 00:15:42.846 "flush": true, 00:15:42.846 "reset": true, 00:15:42.846 "nvme_admin": false, 00:15:42.846 "nvme_io": false, 00:15:42.846 "nvme_io_md": false, 00:15:42.846 "write_zeroes": true, 00:15:42.846 "zcopy": true, 00:15:42.846 "get_zone_info": false, 00:15:42.846 "zone_management": false, 00:15:42.846 "zone_append": false, 00:15:42.846 "compare": false, 00:15:42.846 "compare_and_write": false, 00:15:42.846 "abort": true, 00:15:42.846 "seek_hole": false, 00:15:42.846 "seek_data": false, 00:15:42.846 "copy": true, 00:15:42.846 "nvme_iov_md": false 00:15:42.846 }, 00:15:42.846 "memory_domains": [ 00:15:42.846 { 00:15:42.846 "dma_device_id": "system", 00:15:42.846 "dma_device_type": 1 00:15:42.846 }, 00:15:42.846 { 00:15:42.846 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:42.846 "dma_device_type": 2 00:15:42.846 } 00:15:42.846 ], 00:15:42.846 "driver_specific": {} 00:15:42.846 } 00:15:42.846 ] 00:15:42.846 11:26:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:15:42.846 11:26:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@335 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:15:42.846 11:26:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:15:42.846 11:26:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:15:42.846 11:26:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:42.846 11:26:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:42.846 11:26:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:42.846 11:26:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:42.846 11:26:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:42.846 11:26:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:42.846 11:26:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:42.846 11:26:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:42.846 11:26:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:43.103 11:26:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:43.103 "name": "Existed_Raid", 00:15:43.103 "uuid": "82c5ad70-b465-4435-8c2c-c3ae777041b9", 00:15:43.104 "strip_size_kb": 64, 00:15:43.104 "state": "online", 00:15:43.104 "raid_level": "concat", 00:15:43.104 "superblock": true, 00:15:43.104 "num_base_bdevs": 3, 00:15:43.104 "num_base_bdevs_discovered": 3, 00:15:43.104 "num_base_bdevs_operational": 3, 00:15:43.104 "base_bdevs_list": [ 00:15:43.104 { 00:15:43.104 "name": "NewBaseBdev", 00:15:43.104 "uuid": "ae763fb0-db7e-4831-a3f3-491303dec5a8", 00:15:43.104 "is_configured": true, 00:15:43.104 "data_offset": 2048, 00:15:43.104 "data_size": 63488 00:15:43.104 }, 00:15:43.104 { 00:15:43.104 "name": "BaseBdev2", 00:15:43.104 "uuid": "b2417fa7-1966-42f0-9987-b6d00431bd93", 00:15:43.104 "is_configured": true, 00:15:43.104 "data_offset": 2048, 00:15:43.104 "data_size": 63488 00:15:43.104 }, 00:15:43.104 { 00:15:43.104 "name": "BaseBdev3", 00:15:43.104 "uuid": "fa343c90-ced5-49bd-ab13-fe669d6c7579", 00:15:43.104 "is_configured": true, 00:15:43.104 "data_offset": 2048, 00:15:43.104 "data_size": 63488 00:15:43.104 } 00:15:43.104 ] 00:15:43.104 }' 00:15:43.104 11:26:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:43.104 11:26:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:43.667 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@336 -- # verify_raid_bdev_properties Existed_Raid 00:15:43.667 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:15:43.667 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:15:43.667 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:15:43.667 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:15:43.667 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # local name 00:15:43.667 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:15:43.667 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:15:43.924 [2024-07-15 11:26:27.324286] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:43.924 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:15:43.924 "name": "Existed_Raid", 00:15:43.924 "aliases": [ 00:15:43.924 "82c5ad70-b465-4435-8c2c-c3ae777041b9" 00:15:43.924 ], 00:15:43.924 "product_name": "Raid Volume", 00:15:43.924 "block_size": 512, 00:15:43.924 "num_blocks": 190464, 00:15:43.924 "uuid": "82c5ad70-b465-4435-8c2c-c3ae777041b9", 00:15:43.924 "assigned_rate_limits": { 00:15:43.924 "rw_ios_per_sec": 0, 00:15:43.924 "rw_mbytes_per_sec": 0, 00:15:43.924 "r_mbytes_per_sec": 0, 00:15:43.924 "w_mbytes_per_sec": 0 00:15:43.924 }, 00:15:43.924 "claimed": false, 00:15:43.924 "zoned": false, 00:15:43.924 "supported_io_types": { 00:15:43.924 "read": true, 00:15:43.924 "write": true, 00:15:43.924 "unmap": true, 00:15:43.924 "flush": true, 00:15:43.924 "reset": true, 00:15:43.924 "nvme_admin": false, 00:15:43.924 "nvme_io": false, 00:15:43.924 "nvme_io_md": false, 00:15:43.924 "write_zeroes": true, 00:15:43.924 "zcopy": false, 00:15:43.924 "get_zone_info": false, 00:15:43.924 "zone_management": false, 00:15:43.924 "zone_append": false, 00:15:43.924 "compare": false, 00:15:43.924 "compare_and_write": false, 00:15:43.924 "abort": false, 00:15:43.924 "seek_hole": false, 00:15:43.924 "seek_data": false, 00:15:43.924 "copy": false, 00:15:43.924 "nvme_iov_md": false 00:15:43.924 }, 00:15:43.924 "memory_domains": [ 00:15:43.924 { 00:15:43.924 "dma_device_id": "system", 00:15:43.924 "dma_device_type": 1 00:15:43.924 }, 00:15:43.924 { 00:15:43.924 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:43.924 "dma_device_type": 2 00:15:43.924 }, 00:15:43.924 { 00:15:43.924 "dma_device_id": "system", 00:15:43.924 "dma_device_type": 1 00:15:43.924 }, 00:15:43.924 { 00:15:43.924 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:43.924 "dma_device_type": 2 00:15:43.924 }, 00:15:43.924 { 00:15:43.924 "dma_device_id": "system", 00:15:43.924 "dma_device_type": 1 00:15:43.924 }, 00:15:43.924 { 00:15:43.924 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:43.924 "dma_device_type": 2 00:15:43.924 } 00:15:43.924 ], 00:15:43.924 "driver_specific": { 00:15:43.924 "raid": { 00:15:43.924 "uuid": "82c5ad70-b465-4435-8c2c-c3ae777041b9", 00:15:43.924 "strip_size_kb": 64, 00:15:43.924 "state": "online", 00:15:43.924 "raid_level": "concat", 00:15:43.924 "superblock": true, 00:15:43.924 "num_base_bdevs": 3, 00:15:43.924 "num_base_bdevs_discovered": 3, 00:15:43.924 "num_base_bdevs_operational": 3, 00:15:43.924 "base_bdevs_list": [ 00:15:43.924 { 00:15:43.924 "name": "NewBaseBdev", 00:15:43.924 "uuid": "ae763fb0-db7e-4831-a3f3-491303dec5a8", 00:15:43.924 "is_configured": true, 00:15:43.924 "data_offset": 2048, 00:15:43.924 "data_size": 63488 00:15:43.924 }, 00:15:43.924 { 00:15:43.924 "name": "BaseBdev2", 00:15:43.924 "uuid": "b2417fa7-1966-42f0-9987-b6d00431bd93", 00:15:43.924 "is_configured": true, 00:15:43.924 "data_offset": 2048, 00:15:43.924 "data_size": 63488 00:15:43.924 }, 00:15:43.924 { 00:15:43.924 "name": "BaseBdev3", 00:15:43.924 "uuid": "fa343c90-ced5-49bd-ab13-fe669d6c7579", 00:15:43.924 "is_configured": true, 00:15:43.924 "data_offset": 2048, 00:15:43.924 "data_size": 63488 00:15:43.925 } 00:15:43.925 ] 00:15:43.925 } 00:15:43.925 } 00:15:43.925 }' 00:15:43.925 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:43.925 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # base_bdev_names='NewBaseBdev 00:15:43.925 BaseBdev2 00:15:43.925 BaseBdev3' 00:15:43.925 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:15:43.925 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev 00:15:43.925 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:15:44.182 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:15:44.182 "name": "NewBaseBdev", 00:15:44.182 "aliases": [ 00:15:44.182 "ae763fb0-db7e-4831-a3f3-491303dec5a8" 00:15:44.182 ], 00:15:44.182 "product_name": "Malloc disk", 00:15:44.182 "block_size": 512, 00:15:44.182 "num_blocks": 65536, 00:15:44.182 "uuid": "ae763fb0-db7e-4831-a3f3-491303dec5a8", 00:15:44.182 "assigned_rate_limits": { 00:15:44.182 "rw_ios_per_sec": 0, 00:15:44.182 "rw_mbytes_per_sec": 0, 00:15:44.182 "r_mbytes_per_sec": 0, 00:15:44.182 "w_mbytes_per_sec": 0 00:15:44.182 }, 00:15:44.182 "claimed": true, 00:15:44.182 "claim_type": "exclusive_write", 00:15:44.182 "zoned": false, 00:15:44.182 "supported_io_types": { 00:15:44.182 "read": true, 00:15:44.182 "write": true, 00:15:44.182 "unmap": true, 00:15:44.182 "flush": true, 00:15:44.182 "reset": true, 00:15:44.182 "nvme_admin": false, 00:15:44.182 "nvme_io": false, 00:15:44.182 "nvme_io_md": false, 00:15:44.182 "write_zeroes": true, 00:15:44.182 "zcopy": true, 00:15:44.182 "get_zone_info": false, 00:15:44.182 "zone_management": false, 00:15:44.182 "zone_append": false, 00:15:44.182 "compare": false, 00:15:44.182 "compare_and_write": false, 00:15:44.182 "abort": true, 00:15:44.182 "seek_hole": false, 00:15:44.182 "seek_data": false, 00:15:44.182 "copy": true, 00:15:44.182 "nvme_iov_md": false 00:15:44.182 }, 00:15:44.182 "memory_domains": [ 00:15:44.182 { 00:15:44.182 "dma_device_id": "system", 00:15:44.182 "dma_device_type": 1 00:15:44.182 }, 00:15:44.182 { 00:15:44.182 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:44.182 "dma_device_type": 2 00:15:44.182 } 00:15:44.182 ], 00:15:44.182 "driver_specific": {} 00:15:44.182 }' 00:15:44.182 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:44.182 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:44.182 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:15:44.182 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:44.182 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:44.439 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:15:44.439 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:44.439 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:44.439 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:15:44.439 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:44.439 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:44.439 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:15:44.439 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:15:44.439 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:15:44.439 11:26:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:15:44.695 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:15:44.695 "name": "BaseBdev2", 00:15:44.695 "aliases": [ 00:15:44.695 "b2417fa7-1966-42f0-9987-b6d00431bd93" 00:15:44.695 ], 00:15:44.695 "product_name": "Malloc disk", 00:15:44.695 "block_size": 512, 00:15:44.695 "num_blocks": 65536, 00:15:44.695 "uuid": "b2417fa7-1966-42f0-9987-b6d00431bd93", 00:15:44.695 "assigned_rate_limits": { 00:15:44.695 "rw_ios_per_sec": 0, 00:15:44.695 "rw_mbytes_per_sec": 0, 00:15:44.695 "r_mbytes_per_sec": 0, 00:15:44.695 "w_mbytes_per_sec": 0 00:15:44.695 }, 00:15:44.695 "claimed": true, 00:15:44.695 "claim_type": "exclusive_write", 00:15:44.695 "zoned": false, 00:15:44.695 "supported_io_types": { 00:15:44.695 "read": true, 00:15:44.695 "write": true, 00:15:44.695 "unmap": true, 00:15:44.695 "flush": true, 00:15:44.695 "reset": true, 00:15:44.695 "nvme_admin": false, 00:15:44.695 "nvme_io": false, 00:15:44.695 "nvme_io_md": false, 00:15:44.695 "write_zeroes": true, 00:15:44.695 "zcopy": true, 00:15:44.695 "get_zone_info": false, 00:15:44.695 "zone_management": false, 00:15:44.695 "zone_append": false, 00:15:44.695 "compare": false, 00:15:44.695 "compare_and_write": false, 00:15:44.695 "abort": true, 00:15:44.695 "seek_hole": false, 00:15:44.695 "seek_data": false, 00:15:44.695 "copy": true, 00:15:44.695 "nvme_iov_md": false 00:15:44.695 }, 00:15:44.695 "memory_domains": [ 00:15:44.695 { 00:15:44.695 "dma_device_id": "system", 00:15:44.695 "dma_device_type": 1 00:15:44.695 }, 00:15:44.695 { 00:15:44.695 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:44.695 "dma_device_type": 2 00:15:44.695 } 00:15:44.695 ], 00:15:44.695 "driver_specific": {} 00:15:44.695 }' 00:15:44.695 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:44.695 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:44.949 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:15:44.949 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:44.949 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:44.949 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:15:44.949 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:44.949 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:44.949 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:15:44.949 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:44.949 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:45.205 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:15:45.205 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:15:45.205 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:15:45.205 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:15:45.462 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:15:45.462 "name": "BaseBdev3", 00:15:45.462 "aliases": [ 00:15:45.462 "fa343c90-ced5-49bd-ab13-fe669d6c7579" 00:15:45.462 ], 00:15:45.462 "product_name": "Malloc disk", 00:15:45.462 "block_size": 512, 00:15:45.462 "num_blocks": 65536, 00:15:45.462 "uuid": "fa343c90-ced5-49bd-ab13-fe669d6c7579", 00:15:45.462 "assigned_rate_limits": { 00:15:45.462 "rw_ios_per_sec": 0, 00:15:45.462 "rw_mbytes_per_sec": 0, 00:15:45.462 "r_mbytes_per_sec": 0, 00:15:45.462 "w_mbytes_per_sec": 0 00:15:45.462 }, 00:15:45.462 "claimed": true, 00:15:45.462 "claim_type": "exclusive_write", 00:15:45.462 "zoned": false, 00:15:45.462 "supported_io_types": { 00:15:45.462 "read": true, 00:15:45.462 "write": true, 00:15:45.462 "unmap": true, 00:15:45.462 "flush": true, 00:15:45.462 "reset": true, 00:15:45.462 "nvme_admin": false, 00:15:45.462 "nvme_io": false, 00:15:45.462 "nvme_io_md": false, 00:15:45.462 "write_zeroes": true, 00:15:45.462 "zcopy": true, 00:15:45.462 "get_zone_info": false, 00:15:45.462 "zone_management": false, 00:15:45.462 "zone_append": false, 00:15:45.462 "compare": false, 00:15:45.462 "compare_and_write": false, 00:15:45.462 "abort": true, 00:15:45.462 "seek_hole": false, 00:15:45.462 "seek_data": false, 00:15:45.462 "copy": true, 00:15:45.462 "nvme_iov_md": false 00:15:45.462 }, 00:15:45.462 "memory_domains": [ 00:15:45.462 { 00:15:45.462 "dma_device_id": "system", 00:15:45.462 "dma_device_type": 1 00:15:45.462 }, 00:15:45.462 { 00:15:45.462 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:45.462 "dma_device_type": 2 00:15:45.462 } 00:15:45.462 ], 00:15:45.462 "driver_specific": {} 00:15:45.462 }' 00:15:45.462 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:45.462 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:45.462 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:15:45.462 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:45.462 11:26:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:45.462 11:26:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:15:45.462 11:26:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:45.462 11:26:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:45.720 11:26:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:15:45.720 11:26:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:45.720 11:26:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:45.720 11:26:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:15:45.720 11:26:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@338 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:15:45.977 [2024-07-15 11:26:29.413587] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:45.977 [2024-07-15 11:26:29.413616] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:45.977 [2024-07-15 11:26:29.413667] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:45.977 [2024-07-15 11:26:29.413715] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:45.977 [2024-07-15 11:26:29.413728] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x139cf50 name Existed_Raid, state offline 00:15:45.977 11:26:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@341 -- # killprocess 899850 00:15:45.977 11:26:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@948 -- # '[' -z 899850 ']' 00:15:45.977 11:26:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@952 -- # kill -0 899850 00:15:45.977 11:26:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@953 -- # uname 00:15:45.977 11:26:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:15:45.977 11:26:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 899850 00:15:45.977 11:26:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:15:45.977 11:26:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:15:45.977 11:26:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@966 -- # echo 'killing process with pid 899850' 00:15:45.977 killing process with pid 899850 00:15:45.977 11:26:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@967 -- # kill 899850 00:15:45.977 [2024-07-15 11:26:29.480290] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:45.977 11:26:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # wait 899850 00:15:45.977 [2024-07-15 11:26:29.507574] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:46.236 11:26:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@343 -- # return 0 00:15:46.236 00:15:46.236 real 0m28.705s 00:15:46.236 user 0m52.667s 00:15:46.236 sys 0m5.123s 00:15:46.236 11:26:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1124 -- # xtrace_disable 00:15:46.236 11:26:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:46.236 ************************************ 00:15:46.236 END TEST raid_state_function_test_sb 00:15:46.236 ************************************ 00:15:46.236 11:26:29 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:15:46.236 11:26:29 bdev_raid -- bdev/bdev_raid.sh@869 -- # run_test raid_superblock_test raid_superblock_test concat 3 00:15:46.236 11:26:29 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:15:46.236 11:26:29 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:15:46.236 11:26:29 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:46.236 ************************************ 00:15:46.236 START TEST raid_superblock_test 00:15:46.236 ************************************ 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1123 -- # raid_superblock_test concat 3 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@392 -- # local raid_level=concat 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local num_base_bdevs=3 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # base_bdevs_malloc=() 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local base_bdevs_malloc 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_pt=() 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_pt 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt_uuid=() 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt_uuid 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local raid_bdev_name=raid_bdev1 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local strip_size 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size_create_arg 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local raid_bdev_uuid 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@403 -- # '[' concat '!=' raid1 ']' 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # strip_size=64 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size_create_arg='-z 64' 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # raid_pid=904277 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # waitforlisten 904277 /var/tmp/spdk-raid.sock 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@410 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -L bdev_raid 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@829 -- # '[' -z 904277 ']' 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:15:46.236 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:15:46.236 11:26:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.494 [2024-07-15 11:26:29.885605] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:15:46.494 [2024-07-15 11:26:29.885676] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid904277 ] 00:15:46.494 [2024-07-15 11:26:30.016786] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:46.751 [2024-07-15 11:26:30.119552] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:15:46.751 [2024-07-15 11:26:30.185507] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:46.751 [2024-07-15 11:26:30.185547] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:47.317 11:26:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:15:47.317 11:26:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@862 -- # return 0 00:15:47.317 11:26:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i = 1 )) 00:15:47.317 11:26:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:15:47.317 11:26:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc1 00:15:47.317 11:26:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt1 00:15:47.317 11:26:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:15:47.317 11:26:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:47.317 11:26:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:15:47.317 11:26:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:47.318 11:26:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc1 00:15:47.574 malloc1 00:15:47.574 11:26:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:47.832 [2024-07-15 11:26:31.299120] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:47.832 [2024-07-15 11:26:31.299171] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:47.832 [2024-07-15 11:26:31.299192] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x17ee570 00:15:47.832 [2024-07-15 11:26:31.299204] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:47.832 [2024-07-15 11:26:31.300843] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:47.832 [2024-07-15 11:26:31.300873] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:47.832 pt1 00:15:47.832 11:26:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:15:47.833 11:26:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:15:47.833 11:26:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc2 00:15:47.833 11:26:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt2 00:15:47.833 11:26:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:15:47.833 11:26:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:47.833 11:26:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:15:47.833 11:26:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:47.833 11:26:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc2 00:15:48.091 malloc2 00:15:48.091 11:26:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:48.349 [2024-07-15 11:26:31.797198] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:48.349 [2024-07-15 11:26:31.797244] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:48.349 [2024-07-15 11:26:31.797266] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x17ef970 00:15:48.349 [2024-07-15 11:26:31.797279] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:48.349 [2024-07-15 11:26:31.798727] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:48.349 [2024-07-15 11:26:31.798754] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:48.349 pt2 00:15:48.349 11:26:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:15:48.349 11:26:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:15:48.349 11:26:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc3 00:15:48.349 11:26:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt3 00:15:48.349 11:26:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:15:48.349 11:26:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:48.349 11:26:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:15:48.349 11:26:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:48.349 11:26:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc3 00:15:48.606 malloc3 00:15:48.606 11:26:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:15:48.864 [2024-07-15 11:26:32.295189] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:15:48.864 [2024-07-15 11:26:32.295238] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:48.864 [2024-07-15 11:26:32.295256] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1986340 00:15:48.864 [2024-07-15 11:26:32.295268] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:48.864 [2024-07-15 11:26:32.296834] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:48.864 [2024-07-15 11:26:32.296862] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:15:48.864 pt3 00:15:48.864 11:26:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:15:48.864 11:26:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:15:48.864 11:26:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@429 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'pt1 pt2 pt3' -n raid_bdev1 -s 00:15:49.121 [2024-07-15 11:26:32.531839] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:49.121 [2024-07-15 11:26:32.533129] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:49.121 [2024-07-15 11:26:32.533186] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:15:49.121 [2024-07-15 11:26:32.533337] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x17e6ea0 00:15:49.121 [2024-07-15 11:26:32.533348] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:49.121 [2024-07-15 11:26:32.533548] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x17ee240 00:15:49.121 [2024-07-15 11:26:32.533690] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x17e6ea0 00:15:49.121 [2024-07-15 11:26:32.533700] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x17e6ea0 00:15:49.121 [2024-07-15 11:26:32.533798] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:49.121 11:26:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:15:49.121 11:26:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:15:49.121 11:26:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:15:49.121 11:26:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:49.121 11:26:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:49.121 11:26:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:49.121 11:26:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:49.121 11:26:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:49.121 11:26:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:49.121 11:26:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:49.121 11:26:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:49.121 11:26:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:49.378 11:26:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:49.378 "name": "raid_bdev1", 00:15:49.378 "uuid": "9bd35c1a-b498-4611-8f66-2ad677fd6a6e", 00:15:49.378 "strip_size_kb": 64, 00:15:49.379 "state": "online", 00:15:49.379 "raid_level": "concat", 00:15:49.379 "superblock": true, 00:15:49.379 "num_base_bdevs": 3, 00:15:49.379 "num_base_bdevs_discovered": 3, 00:15:49.379 "num_base_bdevs_operational": 3, 00:15:49.379 "base_bdevs_list": [ 00:15:49.379 { 00:15:49.379 "name": "pt1", 00:15:49.379 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:49.379 "is_configured": true, 00:15:49.379 "data_offset": 2048, 00:15:49.379 "data_size": 63488 00:15:49.379 }, 00:15:49.379 { 00:15:49.379 "name": "pt2", 00:15:49.379 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:49.379 "is_configured": true, 00:15:49.379 "data_offset": 2048, 00:15:49.379 "data_size": 63488 00:15:49.379 }, 00:15:49.379 { 00:15:49.379 "name": "pt3", 00:15:49.379 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:49.379 "is_configured": true, 00:15:49.379 "data_offset": 2048, 00:15:49.379 "data_size": 63488 00:15:49.379 } 00:15:49.379 ] 00:15:49.379 }' 00:15:49.379 11:26:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:49.379 11:26:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:49.943 11:26:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_properties raid_bdev1 00:15:49.943 11:26:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:15:49.943 11:26:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:15:49.943 11:26:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:15:49.943 11:26:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:15:49.943 11:26:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # local name 00:15:49.943 11:26:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:15:49.943 11:26:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:15:50.199 [2024-07-15 11:26:33.546774] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:50.199 11:26:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:15:50.199 "name": "raid_bdev1", 00:15:50.199 "aliases": [ 00:15:50.199 "9bd35c1a-b498-4611-8f66-2ad677fd6a6e" 00:15:50.199 ], 00:15:50.199 "product_name": "Raid Volume", 00:15:50.199 "block_size": 512, 00:15:50.199 "num_blocks": 190464, 00:15:50.199 "uuid": "9bd35c1a-b498-4611-8f66-2ad677fd6a6e", 00:15:50.199 "assigned_rate_limits": { 00:15:50.199 "rw_ios_per_sec": 0, 00:15:50.199 "rw_mbytes_per_sec": 0, 00:15:50.199 "r_mbytes_per_sec": 0, 00:15:50.199 "w_mbytes_per_sec": 0 00:15:50.199 }, 00:15:50.199 "claimed": false, 00:15:50.199 "zoned": false, 00:15:50.199 "supported_io_types": { 00:15:50.199 "read": true, 00:15:50.199 "write": true, 00:15:50.199 "unmap": true, 00:15:50.199 "flush": true, 00:15:50.199 "reset": true, 00:15:50.199 "nvme_admin": false, 00:15:50.199 "nvme_io": false, 00:15:50.199 "nvme_io_md": false, 00:15:50.199 "write_zeroes": true, 00:15:50.199 "zcopy": false, 00:15:50.199 "get_zone_info": false, 00:15:50.199 "zone_management": false, 00:15:50.199 "zone_append": false, 00:15:50.199 "compare": false, 00:15:50.199 "compare_and_write": false, 00:15:50.199 "abort": false, 00:15:50.199 "seek_hole": false, 00:15:50.199 "seek_data": false, 00:15:50.199 "copy": false, 00:15:50.199 "nvme_iov_md": false 00:15:50.199 }, 00:15:50.199 "memory_domains": [ 00:15:50.199 { 00:15:50.199 "dma_device_id": "system", 00:15:50.199 "dma_device_type": 1 00:15:50.199 }, 00:15:50.199 { 00:15:50.199 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:50.199 "dma_device_type": 2 00:15:50.199 }, 00:15:50.199 { 00:15:50.199 "dma_device_id": "system", 00:15:50.199 "dma_device_type": 1 00:15:50.199 }, 00:15:50.199 { 00:15:50.199 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:50.200 "dma_device_type": 2 00:15:50.200 }, 00:15:50.200 { 00:15:50.200 "dma_device_id": "system", 00:15:50.200 "dma_device_type": 1 00:15:50.200 }, 00:15:50.200 { 00:15:50.200 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:50.200 "dma_device_type": 2 00:15:50.200 } 00:15:50.200 ], 00:15:50.200 "driver_specific": { 00:15:50.200 "raid": { 00:15:50.200 "uuid": "9bd35c1a-b498-4611-8f66-2ad677fd6a6e", 00:15:50.200 "strip_size_kb": 64, 00:15:50.200 "state": "online", 00:15:50.200 "raid_level": "concat", 00:15:50.200 "superblock": true, 00:15:50.200 "num_base_bdevs": 3, 00:15:50.200 "num_base_bdevs_discovered": 3, 00:15:50.200 "num_base_bdevs_operational": 3, 00:15:50.200 "base_bdevs_list": [ 00:15:50.200 { 00:15:50.200 "name": "pt1", 00:15:50.200 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:50.200 "is_configured": true, 00:15:50.200 "data_offset": 2048, 00:15:50.200 "data_size": 63488 00:15:50.200 }, 00:15:50.200 { 00:15:50.200 "name": "pt2", 00:15:50.200 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:50.200 "is_configured": true, 00:15:50.200 "data_offset": 2048, 00:15:50.200 "data_size": 63488 00:15:50.200 }, 00:15:50.200 { 00:15:50.200 "name": "pt3", 00:15:50.200 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:50.200 "is_configured": true, 00:15:50.200 "data_offset": 2048, 00:15:50.200 "data_size": 63488 00:15:50.200 } 00:15:50.200 ] 00:15:50.200 } 00:15:50.200 } 00:15:50.200 }' 00:15:50.200 11:26:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:50.200 11:26:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:15:50.200 pt2 00:15:50.200 pt3' 00:15:50.200 11:26:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:15:50.200 11:26:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:15:50.200 11:26:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:15:50.457 11:26:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:15:50.457 "name": "pt1", 00:15:50.457 "aliases": [ 00:15:50.457 "00000000-0000-0000-0000-000000000001" 00:15:50.457 ], 00:15:50.457 "product_name": "passthru", 00:15:50.457 "block_size": 512, 00:15:50.457 "num_blocks": 65536, 00:15:50.457 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:50.457 "assigned_rate_limits": { 00:15:50.457 "rw_ios_per_sec": 0, 00:15:50.457 "rw_mbytes_per_sec": 0, 00:15:50.457 "r_mbytes_per_sec": 0, 00:15:50.457 "w_mbytes_per_sec": 0 00:15:50.457 }, 00:15:50.457 "claimed": true, 00:15:50.457 "claim_type": "exclusive_write", 00:15:50.457 "zoned": false, 00:15:50.457 "supported_io_types": { 00:15:50.457 "read": true, 00:15:50.457 "write": true, 00:15:50.457 "unmap": true, 00:15:50.457 "flush": true, 00:15:50.457 "reset": true, 00:15:50.457 "nvme_admin": false, 00:15:50.457 "nvme_io": false, 00:15:50.457 "nvme_io_md": false, 00:15:50.457 "write_zeroes": true, 00:15:50.457 "zcopy": true, 00:15:50.457 "get_zone_info": false, 00:15:50.457 "zone_management": false, 00:15:50.457 "zone_append": false, 00:15:50.457 "compare": false, 00:15:50.457 "compare_and_write": false, 00:15:50.457 "abort": true, 00:15:50.457 "seek_hole": false, 00:15:50.457 "seek_data": false, 00:15:50.457 "copy": true, 00:15:50.457 "nvme_iov_md": false 00:15:50.457 }, 00:15:50.457 "memory_domains": [ 00:15:50.457 { 00:15:50.457 "dma_device_id": "system", 00:15:50.457 "dma_device_type": 1 00:15:50.457 }, 00:15:50.457 { 00:15:50.457 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:50.457 "dma_device_type": 2 00:15:50.457 } 00:15:50.457 ], 00:15:50.457 "driver_specific": { 00:15:50.457 "passthru": { 00:15:50.457 "name": "pt1", 00:15:50.457 "base_bdev_name": "malloc1" 00:15:50.457 } 00:15:50.457 } 00:15:50.457 }' 00:15:50.457 11:26:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:50.457 11:26:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:50.457 11:26:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:15:50.457 11:26:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:50.457 11:26:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:50.457 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:15:50.457 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:50.714 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:50.714 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:15:50.714 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:50.714 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:50.714 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:15:50.714 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:15:50.714 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:15:50.714 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:15:50.971 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:15:50.971 "name": "pt2", 00:15:50.971 "aliases": [ 00:15:50.971 "00000000-0000-0000-0000-000000000002" 00:15:50.971 ], 00:15:50.971 "product_name": "passthru", 00:15:50.971 "block_size": 512, 00:15:50.971 "num_blocks": 65536, 00:15:50.971 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:50.971 "assigned_rate_limits": { 00:15:50.971 "rw_ios_per_sec": 0, 00:15:50.971 "rw_mbytes_per_sec": 0, 00:15:50.971 "r_mbytes_per_sec": 0, 00:15:50.971 "w_mbytes_per_sec": 0 00:15:50.971 }, 00:15:50.971 "claimed": true, 00:15:50.971 "claim_type": "exclusive_write", 00:15:50.971 "zoned": false, 00:15:50.971 "supported_io_types": { 00:15:50.971 "read": true, 00:15:50.971 "write": true, 00:15:50.971 "unmap": true, 00:15:50.971 "flush": true, 00:15:50.971 "reset": true, 00:15:50.971 "nvme_admin": false, 00:15:50.971 "nvme_io": false, 00:15:50.971 "nvme_io_md": false, 00:15:50.971 "write_zeroes": true, 00:15:50.971 "zcopy": true, 00:15:50.971 "get_zone_info": false, 00:15:50.971 "zone_management": false, 00:15:50.971 "zone_append": false, 00:15:50.971 "compare": false, 00:15:50.971 "compare_and_write": false, 00:15:50.971 "abort": true, 00:15:50.971 "seek_hole": false, 00:15:50.971 "seek_data": false, 00:15:50.971 "copy": true, 00:15:50.971 "nvme_iov_md": false 00:15:50.971 }, 00:15:50.971 "memory_domains": [ 00:15:50.971 { 00:15:50.971 "dma_device_id": "system", 00:15:50.971 "dma_device_type": 1 00:15:50.971 }, 00:15:50.971 { 00:15:50.971 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:50.971 "dma_device_type": 2 00:15:50.971 } 00:15:50.971 ], 00:15:50.971 "driver_specific": { 00:15:50.971 "passthru": { 00:15:50.971 "name": "pt2", 00:15:50.971 "base_bdev_name": "malloc2" 00:15:50.971 } 00:15:50.971 } 00:15:50.971 }' 00:15:50.971 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:50.971 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:50.971 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:15:50.971 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:51.228 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:51.228 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:15:51.228 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:51.228 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:51.228 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:15:51.228 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:51.228 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:51.228 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:15:51.228 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:15:51.228 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt3 00:15:51.228 11:26:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:15:51.524 11:26:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:15:51.524 "name": "pt3", 00:15:51.524 "aliases": [ 00:15:51.524 "00000000-0000-0000-0000-000000000003" 00:15:51.524 ], 00:15:51.524 "product_name": "passthru", 00:15:51.524 "block_size": 512, 00:15:51.524 "num_blocks": 65536, 00:15:51.524 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:51.524 "assigned_rate_limits": { 00:15:51.524 "rw_ios_per_sec": 0, 00:15:51.524 "rw_mbytes_per_sec": 0, 00:15:51.524 "r_mbytes_per_sec": 0, 00:15:51.524 "w_mbytes_per_sec": 0 00:15:51.524 }, 00:15:51.524 "claimed": true, 00:15:51.524 "claim_type": "exclusive_write", 00:15:51.524 "zoned": false, 00:15:51.524 "supported_io_types": { 00:15:51.524 "read": true, 00:15:51.524 "write": true, 00:15:51.524 "unmap": true, 00:15:51.524 "flush": true, 00:15:51.524 "reset": true, 00:15:51.524 "nvme_admin": false, 00:15:51.524 "nvme_io": false, 00:15:51.524 "nvme_io_md": false, 00:15:51.524 "write_zeroes": true, 00:15:51.524 "zcopy": true, 00:15:51.524 "get_zone_info": false, 00:15:51.524 "zone_management": false, 00:15:51.524 "zone_append": false, 00:15:51.524 "compare": false, 00:15:51.524 "compare_and_write": false, 00:15:51.524 "abort": true, 00:15:51.524 "seek_hole": false, 00:15:51.524 "seek_data": false, 00:15:51.524 "copy": true, 00:15:51.524 "nvme_iov_md": false 00:15:51.524 }, 00:15:51.524 "memory_domains": [ 00:15:51.524 { 00:15:51.524 "dma_device_id": "system", 00:15:51.524 "dma_device_type": 1 00:15:51.524 }, 00:15:51.524 { 00:15:51.524 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:51.524 "dma_device_type": 2 00:15:51.524 } 00:15:51.524 ], 00:15:51.524 "driver_specific": { 00:15:51.524 "passthru": { 00:15:51.524 "name": "pt3", 00:15:51.524 "base_bdev_name": "malloc3" 00:15:51.524 } 00:15:51.524 } 00:15:51.524 }' 00:15:51.524 11:26:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:51.524 11:26:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:51.802 11:26:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:15:51.802 11:26:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:51.802 11:26:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:51.802 11:26:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:15:51.802 11:26:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:51.802 11:26:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:51.802 11:26:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:15:51.802 11:26:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:51.802 11:26:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:51.802 11:26:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:15:51.802 11:26:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:15:51.802 11:26:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # jq -r '.[] | .uuid' 00:15:52.059 [2024-07-15 11:26:35.620244] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:52.059 11:26:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # raid_bdev_uuid=9bd35c1a-b498-4611-8f66-2ad677fd6a6e 00:15:52.059 11:26:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # '[' -z 9bd35c1a-b498-4611-8f66-2ad677fd6a6e ']' 00:15:52.059 11:26:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@440 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:15:52.316 [2024-07-15 11:26:35.796435] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:52.316 [2024-07-15 11:26:35.796453] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:52.316 [2024-07-15 11:26:35.796500] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:52.316 [2024-07-15 11:26:35.796551] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:52.316 [2024-07-15 11:26:35.796563] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x17e6ea0 name raid_bdev1, state offline 00:15:52.316 11:26:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:52.316 11:26:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # jq -r '.[]' 00:15:52.572 11:26:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # raid_bdev= 00:15:52.572 11:26:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # '[' -n '' ']' 00:15:52.572 11:26:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:15:52.572 11:26:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt1 00:15:52.829 11:26:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:15:52.829 11:26:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:15:53.087 11:26:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:15:53.087 11:26:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt3 00:15:53.344 11:26:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:15:53.344 11:26:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs 00:15:53.602 11:26:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # '[' false == true ']' 00:15:53.602 11:26:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@456 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'malloc1 malloc2 malloc3' -n raid_bdev1 00:15:53.602 11:26:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@648 -- # local es=0 00:15:53.602 11:26:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'malloc1 malloc2 malloc3' -n raid_bdev1 00:15:53.602 11:26:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:15:53.602 11:26:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:15:53.602 11:26:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:15:53.602 11:26:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:15:53.602 11:26:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:15:53.602 11:26:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:15:53.602 11:26:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:15:53.602 11:26:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py ]] 00:15:53.603 11:26:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'malloc1 malloc2 malloc3' -n raid_bdev1 00:15:53.603 [2024-07-15 11:26:37.188064] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:15:53.603 [2024-07-15 11:26:37.189463] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:15:53.603 [2024-07-15 11:26:37.189507] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:15:53.603 [2024-07-15 11:26:37.189555] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:15:53.603 [2024-07-15 11:26:37.189595] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:15:53.603 [2024-07-15 11:26:37.189617] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:15:53.603 [2024-07-15 11:26:37.189636] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:53.603 [2024-07-15 11:26:37.189646] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1991ff0 name raid_bdev1, state configuring 00:15:53.603 request: 00:15:53.603 { 00:15:53.603 "name": "raid_bdev1", 00:15:53.603 "raid_level": "concat", 00:15:53.603 "base_bdevs": [ 00:15:53.603 "malloc1", 00:15:53.603 "malloc2", 00:15:53.603 "malloc3" 00:15:53.603 ], 00:15:53.603 "strip_size_kb": 64, 00:15:53.603 "superblock": false, 00:15:53.603 "method": "bdev_raid_create", 00:15:53.603 "req_id": 1 00:15:53.603 } 00:15:53.603 Got JSON-RPC error response 00:15:53.603 response: 00:15:53.603 { 00:15:53.603 "code": -17, 00:15:53.603 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:15:53.603 } 00:15:53.860 11:26:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@651 -- # es=1 00:15:53.860 11:26:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:15:53.860 11:26:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:15:53.860 11:26:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:15:53.860 11:26:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:53.860 11:26:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # jq -r '.[]' 00:15:53.860 11:26:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # raid_bdev= 00:15:53.861 11:26:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # '[' -n '' ']' 00:15:53.861 11:26:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@464 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:54.118 [2024-07-15 11:26:37.677288] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:54.118 [2024-07-15 11:26:37.677331] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:54.118 [2024-07-15 11:26:37.677352] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x17ee7a0 00:15:54.118 [2024-07-15 11:26:37.677364] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:54.118 [2024-07-15 11:26:37.678973] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:54.118 [2024-07-15 11:26:37.679003] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:54.118 [2024-07-15 11:26:37.679069] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:54.118 [2024-07-15 11:26:37.679096] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:54.118 pt1 00:15:54.118 11:26:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@467 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:15:54.118 11:26:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:15:54.118 11:26:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:15:54.118 11:26:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:54.118 11:26:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:54.118 11:26:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:54.118 11:26:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:54.118 11:26:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:54.118 11:26:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:54.118 11:26:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:54.118 11:26:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:54.118 11:26:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:54.376 11:26:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:54.376 "name": "raid_bdev1", 00:15:54.376 "uuid": "9bd35c1a-b498-4611-8f66-2ad677fd6a6e", 00:15:54.376 "strip_size_kb": 64, 00:15:54.376 "state": "configuring", 00:15:54.376 "raid_level": "concat", 00:15:54.376 "superblock": true, 00:15:54.376 "num_base_bdevs": 3, 00:15:54.376 "num_base_bdevs_discovered": 1, 00:15:54.376 "num_base_bdevs_operational": 3, 00:15:54.376 "base_bdevs_list": [ 00:15:54.376 { 00:15:54.376 "name": "pt1", 00:15:54.376 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:54.376 "is_configured": true, 00:15:54.376 "data_offset": 2048, 00:15:54.376 "data_size": 63488 00:15:54.376 }, 00:15:54.376 { 00:15:54.376 "name": null, 00:15:54.376 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:54.376 "is_configured": false, 00:15:54.376 "data_offset": 2048, 00:15:54.376 "data_size": 63488 00:15:54.376 }, 00:15:54.376 { 00:15:54.376 "name": null, 00:15:54.376 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:54.376 "is_configured": false, 00:15:54.376 "data_offset": 2048, 00:15:54.376 "data_size": 63488 00:15:54.376 } 00:15:54.376 ] 00:15:54.376 }' 00:15:54.376 11:26:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:54.376 11:26:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:54.943 11:26:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@469 -- # '[' 3 -gt 2 ']' 00:15:54.943 11:26:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@471 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:55.201 [2024-07-15 11:26:38.683961] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:55.201 [2024-07-15 11:26:38.684012] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:55.201 [2024-07-15 11:26:38.684030] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x17e5c70 00:15:55.201 [2024-07-15 11:26:38.684042] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:55.201 [2024-07-15 11:26:38.684393] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:55.201 [2024-07-15 11:26:38.684412] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:55.201 [2024-07-15 11:26:38.684472] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:55.201 [2024-07-15 11:26:38.684491] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:55.201 pt2 00:15:55.201 11:26:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:15:55.460 [2024-07-15 11:26:38.928619] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:15:55.460 11:26:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:15:55.460 11:26:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:15:55.460 11:26:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:15:55.460 11:26:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:55.460 11:26:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:55.460 11:26:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:55.460 11:26:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:55.460 11:26:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:55.460 11:26:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:55.460 11:26:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:55.460 11:26:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:55.460 11:26:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:55.719 11:26:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:55.719 "name": "raid_bdev1", 00:15:55.719 "uuid": "9bd35c1a-b498-4611-8f66-2ad677fd6a6e", 00:15:55.719 "strip_size_kb": 64, 00:15:55.719 "state": "configuring", 00:15:55.719 "raid_level": "concat", 00:15:55.719 "superblock": true, 00:15:55.719 "num_base_bdevs": 3, 00:15:55.719 "num_base_bdevs_discovered": 1, 00:15:55.719 "num_base_bdevs_operational": 3, 00:15:55.719 "base_bdevs_list": [ 00:15:55.719 { 00:15:55.719 "name": "pt1", 00:15:55.719 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:55.719 "is_configured": true, 00:15:55.719 "data_offset": 2048, 00:15:55.719 "data_size": 63488 00:15:55.719 }, 00:15:55.719 { 00:15:55.719 "name": null, 00:15:55.719 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:55.719 "is_configured": false, 00:15:55.719 "data_offset": 2048, 00:15:55.719 "data_size": 63488 00:15:55.719 }, 00:15:55.719 { 00:15:55.719 "name": null, 00:15:55.719 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:55.719 "is_configured": false, 00:15:55.719 "data_offset": 2048, 00:15:55.719 "data_size": 63488 00:15:55.719 } 00:15:55.719 ] 00:15:55.719 }' 00:15:55.719 11:26:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:55.719 11:26:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:56.286 11:26:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i = 1 )) 00:15:56.286 11:26:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:15:56.286 11:26:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:56.545 [2024-07-15 11:26:39.955343] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:56.545 [2024-07-15 11:26:39.955396] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:56.545 [2024-07-15 11:26:39.955419] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x17eea10 00:15:56.545 [2024-07-15 11:26:39.955432] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:56.545 [2024-07-15 11:26:39.955769] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:56.545 [2024-07-15 11:26:39.955787] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:56.545 [2024-07-15 11:26:39.955850] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:56.545 [2024-07-15 11:26:39.955867] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:56.545 pt2 00:15:56.545 11:26:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:15:56.545 11:26:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:15:56.545 11:26:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:15:56.545 [2024-07-15 11:26:40.131821] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:15:56.545 [2024-07-15 11:26:40.131859] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:56.545 [2024-07-15 11:26:40.131875] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1988740 00:15:56.545 [2024-07-15 11:26:40.131887] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:56.545 [2024-07-15 11:26:40.132194] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:56.545 [2024-07-15 11:26:40.132212] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:15:56.545 [2024-07-15 11:26:40.132266] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:15:56.545 [2024-07-15 11:26:40.132283] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:15:56.545 [2024-07-15 11:26:40.132392] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1988c00 00:15:56.545 [2024-07-15 11:26:40.132403] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:56.545 [2024-07-15 11:26:40.132572] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x17eda40 00:15:56.545 [2024-07-15 11:26:40.132696] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1988c00 00:15:56.545 [2024-07-15 11:26:40.132705] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x1988c00 00:15:56.545 [2024-07-15 11:26:40.132799] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:56.545 pt3 00:15:56.804 11:26:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:15:56.804 11:26:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:15:56.804 11:26:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@482 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:15:56.804 11:26:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:15:56.804 11:26:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:15:56.804 11:26:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:15:56.804 11:26:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:15:56.804 11:26:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:15:56.804 11:26:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:15:56.804 11:26:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:15:56.804 11:26:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:15:56.804 11:26:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:15:56.804 11:26:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:15:56.804 11:26:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:57.063 11:26:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:15:57.063 "name": "raid_bdev1", 00:15:57.063 "uuid": "9bd35c1a-b498-4611-8f66-2ad677fd6a6e", 00:15:57.063 "strip_size_kb": 64, 00:15:57.063 "state": "online", 00:15:57.063 "raid_level": "concat", 00:15:57.063 "superblock": true, 00:15:57.063 "num_base_bdevs": 3, 00:15:57.063 "num_base_bdevs_discovered": 3, 00:15:57.063 "num_base_bdevs_operational": 3, 00:15:57.063 "base_bdevs_list": [ 00:15:57.063 { 00:15:57.063 "name": "pt1", 00:15:57.063 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:57.063 "is_configured": true, 00:15:57.063 "data_offset": 2048, 00:15:57.063 "data_size": 63488 00:15:57.063 }, 00:15:57.063 { 00:15:57.063 "name": "pt2", 00:15:57.063 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:57.063 "is_configured": true, 00:15:57.063 "data_offset": 2048, 00:15:57.063 "data_size": 63488 00:15:57.063 }, 00:15:57.063 { 00:15:57.063 "name": "pt3", 00:15:57.063 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:57.063 "is_configured": true, 00:15:57.063 "data_offset": 2048, 00:15:57.063 "data_size": 63488 00:15:57.063 } 00:15:57.063 ] 00:15:57.063 }' 00:15:57.063 11:26:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:15:57.063 11:26:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:57.631 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_properties raid_bdev1 00:15:57.631 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:15:57.631 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:15:57.631 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:15:57.631 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:15:57.631 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # local name 00:15:57.632 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:15:57.632 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:15:57.891 [2024-07-15 11:26:41.231019] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:57.891 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:15:57.891 "name": "raid_bdev1", 00:15:57.891 "aliases": [ 00:15:57.891 "9bd35c1a-b498-4611-8f66-2ad677fd6a6e" 00:15:57.891 ], 00:15:57.891 "product_name": "Raid Volume", 00:15:57.891 "block_size": 512, 00:15:57.891 "num_blocks": 190464, 00:15:57.891 "uuid": "9bd35c1a-b498-4611-8f66-2ad677fd6a6e", 00:15:57.891 "assigned_rate_limits": { 00:15:57.891 "rw_ios_per_sec": 0, 00:15:57.891 "rw_mbytes_per_sec": 0, 00:15:57.891 "r_mbytes_per_sec": 0, 00:15:57.891 "w_mbytes_per_sec": 0 00:15:57.891 }, 00:15:57.891 "claimed": false, 00:15:57.891 "zoned": false, 00:15:57.891 "supported_io_types": { 00:15:57.891 "read": true, 00:15:57.891 "write": true, 00:15:57.891 "unmap": true, 00:15:57.891 "flush": true, 00:15:57.891 "reset": true, 00:15:57.891 "nvme_admin": false, 00:15:57.891 "nvme_io": false, 00:15:57.891 "nvme_io_md": false, 00:15:57.891 "write_zeroes": true, 00:15:57.891 "zcopy": false, 00:15:57.891 "get_zone_info": false, 00:15:57.891 "zone_management": false, 00:15:57.891 "zone_append": false, 00:15:57.891 "compare": false, 00:15:57.891 "compare_and_write": false, 00:15:57.891 "abort": false, 00:15:57.891 "seek_hole": false, 00:15:57.891 "seek_data": false, 00:15:57.891 "copy": false, 00:15:57.891 "nvme_iov_md": false 00:15:57.891 }, 00:15:57.891 "memory_domains": [ 00:15:57.891 { 00:15:57.891 "dma_device_id": "system", 00:15:57.891 "dma_device_type": 1 00:15:57.891 }, 00:15:57.891 { 00:15:57.891 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:57.891 "dma_device_type": 2 00:15:57.891 }, 00:15:57.891 { 00:15:57.891 "dma_device_id": "system", 00:15:57.891 "dma_device_type": 1 00:15:57.891 }, 00:15:57.891 { 00:15:57.891 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:57.891 "dma_device_type": 2 00:15:57.891 }, 00:15:57.891 { 00:15:57.891 "dma_device_id": "system", 00:15:57.891 "dma_device_type": 1 00:15:57.891 }, 00:15:57.891 { 00:15:57.891 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:57.891 "dma_device_type": 2 00:15:57.891 } 00:15:57.891 ], 00:15:57.891 "driver_specific": { 00:15:57.891 "raid": { 00:15:57.891 "uuid": "9bd35c1a-b498-4611-8f66-2ad677fd6a6e", 00:15:57.891 "strip_size_kb": 64, 00:15:57.891 "state": "online", 00:15:57.891 "raid_level": "concat", 00:15:57.891 "superblock": true, 00:15:57.891 "num_base_bdevs": 3, 00:15:57.891 "num_base_bdevs_discovered": 3, 00:15:57.891 "num_base_bdevs_operational": 3, 00:15:57.891 "base_bdevs_list": [ 00:15:57.891 { 00:15:57.891 "name": "pt1", 00:15:57.891 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:57.891 "is_configured": true, 00:15:57.891 "data_offset": 2048, 00:15:57.891 "data_size": 63488 00:15:57.891 }, 00:15:57.891 { 00:15:57.891 "name": "pt2", 00:15:57.892 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:57.892 "is_configured": true, 00:15:57.892 "data_offset": 2048, 00:15:57.892 "data_size": 63488 00:15:57.892 }, 00:15:57.892 { 00:15:57.892 "name": "pt3", 00:15:57.892 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:57.892 "is_configured": true, 00:15:57.892 "data_offset": 2048, 00:15:57.892 "data_size": 63488 00:15:57.892 } 00:15:57.892 ] 00:15:57.892 } 00:15:57.892 } 00:15:57.892 }' 00:15:57.892 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:57.892 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:15:57.892 pt2 00:15:57.892 pt3' 00:15:57.892 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:15:57.892 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:15:57.892 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:15:57.892 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:15:57.892 "name": "pt1", 00:15:57.892 "aliases": [ 00:15:57.892 "00000000-0000-0000-0000-000000000001" 00:15:57.892 ], 00:15:57.892 "product_name": "passthru", 00:15:57.892 "block_size": 512, 00:15:57.892 "num_blocks": 65536, 00:15:57.892 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:57.892 "assigned_rate_limits": { 00:15:57.892 "rw_ios_per_sec": 0, 00:15:57.892 "rw_mbytes_per_sec": 0, 00:15:57.892 "r_mbytes_per_sec": 0, 00:15:57.892 "w_mbytes_per_sec": 0 00:15:57.892 }, 00:15:57.892 "claimed": true, 00:15:57.892 "claim_type": "exclusive_write", 00:15:57.892 "zoned": false, 00:15:57.892 "supported_io_types": { 00:15:57.892 "read": true, 00:15:57.892 "write": true, 00:15:57.892 "unmap": true, 00:15:57.892 "flush": true, 00:15:57.892 "reset": true, 00:15:57.892 "nvme_admin": false, 00:15:57.892 "nvme_io": false, 00:15:57.892 "nvme_io_md": false, 00:15:57.892 "write_zeroes": true, 00:15:57.892 "zcopy": true, 00:15:57.892 "get_zone_info": false, 00:15:57.892 "zone_management": false, 00:15:57.892 "zone_append": false, 00:15:57.892 "compare": false, 00:15:57.892 "compare_and_write": false, 00:15:57.892 "abort": true, 00:15:57.892 "seek_hole": false, 00:15:57.892 "seek_data": false, 00:15:57.892 "copy": true, 00:15:57.892 "nvme_iov_md": false 00:15:57.892 }, 00:15:57.892 "memory_domains": [ 00:15:57.892 { 00:15:57.892 "dma_device_id": "system", 00:15:57.892 "dma_device_type": 1 00:15:57.892 }, 00:15:57.892 { 00:15:57.892 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:57.892 "dma_device_type": 2 00:15:57.892 } 00:15:57.892 ], 00:15:57.892 "driver_specific": { 00:15:57.892 "passthru": { 00:15:57.892 "name": "pt1", 00:15:57.892 "base_bdev_name": "malloc1" 00:15:57.892 } 00:15:57.892 } 00:15:57.892 }' 00:15:57.892 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:58.151 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:58.151 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:15:58.151 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:58.151 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:58.151 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:15:58.151 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:58.151 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:58.151 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:15:58.151 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:58.410 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:58.410 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:15:58.410 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:15:58.410 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:15:58.410 11:26:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:15:58.670 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:15:58.670 "name": "pt2", 00:15:58.670 "aliases": [ 00:15:58.670 "00000000-0000-0000-0000-000000000002" 00:15:58.670 ], 00:15:58.670 "product_name": "passthru", 00:15:58.670 "block_size": 512, 00:15:58.670 "num_blocks": 65536, 00:15:58.670 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:58.670 "assigned_rate_limits": { 00:15:58.670 "rw_ios_per_sec": 0, 00:15:58.670 "rw_mbytes_per_sec": 0, 00:15:58.670 "r_mbytes_per_sec": 0, 00:15:58.670 "w_mbytes_per_sec": 0 00:15:58.670 }, 00:15:58.670 "claimed": true, 00:15:58.670 "claim_type": "exclusive_write", 00:15:58.670 "zoned": false, 00:15:58.670 "supported_io_types": { 00:15:58.670 "read": true, 00:15:58.670 "write": true, 00:15:58.670 "unmap": true, 00:15:58.670 "flush": true, 00:15:58.670 "reset": true, 00:15:58.670 "nvme_admin": false, 00:15:58.670 "nvme_io": false, 00:15:58.670 "nvme_io_md": false, 00:15:58.670 "write_zeroes": true, 00:15:58.670 "zcopy": true, 00:15:58.670 "get_zone_info": false, 00:15:58.670 "zone_management": false, 00:15:58.670 "zone_append": false, 00:15:58.670 "compare": false, 00:15:58.670 "compare_and_write": false, 00:15:58.670 "abort": true, 00:15:58.670 "seek_hole": false, 00:15:58.670 "seek_data": false, 00:15:58.670 "copy": true, 00:15:58.670 "nvme_iov_md": false 00:15:58.670 }, 00:15:58.670 "memory_domains": [ 00:15:58.670 { 00:15:58.670 "dma_device_id": "system", 00:15:58.670 "dma_device_type": 1 00:15:58.670 }, 00:15:58.670 { 00:15:58.670 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:58.670 "dma_device_type": 2 00:15:58.670 } 00:15:58.670 ], 00:15:58.670 "driver_specific": { 00:15:58.670 "passthru": { 00:15:58.670 "name": "pt2", 00:15:58.670 "base_bdev_name": "malloc2" 00:15:58.670 } 00:15:58.670 } 00:15:58.670 }' 00:15:58.670 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:58.670 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:58.670 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:15:58.670 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:58.670 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:58.670 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:15:58.670 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:58.930 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:58.930 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:15:58.930 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:58.930 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:58.930 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:15:58.930 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:15:58.930 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt3 00:15:58.930 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:15:59.189 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:15:59.189 "name": "pt3", 00:15:59.189 "aliases": [ 00:15:59.189 "00000000-0000-0000-0000-000000000003" 00:15:59.189 ], 00:15:59.189 "product_name": "passthru", 00:15:59.189 "block_size": 512, 00:15:59.189 "num_blocks": 65536, 00:15:59.189 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:59.189 "assigned_rate_limits": { 00:15:59.189 "rw_ios_per_sec": 0, 00:15:59.189 "rw_mbytes_per_sec": 0, 00:15:59.189 "r_mbytes_per_sec": 0, 00:15:59.189 "w_mbytes_per_sec": 0 00:15:59.189 }, 00:15:59.189 "claimed": true, 00:15:59.189 "claim_type": "exclusive_write", 00:15:59.189 "zoned": false, 00:15:59.189 "supported_io_types": { 00:15:59.189 "read": true, 00:15:59.189 "write": true, 00:15:59.189 "unmap": true, 00:15:59.189 "flush": true, 00:15:59.189 "reset": true, 00:15:59.189 "nvme_admin": false, 00:15:59.189 "nvme_io": false, 00:15:59.189 "nvme_io_md": false, 00:15:59.189 "write_zeroes": true, 00:15:59.189 "zcopy": true, 00:15:59.189 "get_zone_info": false, 00:15:59.189 "zone_management": false, 00:15:59.189 "zone_append": false, 00:15:59.189 "compare": false, 00:15:59.189 "compare_and_write": false, 00:15:59.189 "abort": true, 00:15:59.189 "seek_hole": false, 00:15:59.190 "seek_data": false, 00:15:59.190 "copy": true, 00:15:59.190 "nvme_iov_md": false 00:15:59.190 }, 00:15:59.190 "memory_domains": [ 00:15:59.190 { 00:15:59.190 "dma_device_id": "system", 00:15:59.190 "dma_device_type": 1 00:15:59.190 }, 00:15:59.190 { 00:15:59.190 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:59.190 "dma_device_type": 2 00:15:59.190 } 00:15:59.190 ], 00:15:59.190 "driver_specific": { 00:15:59.190 "passthru": { 00:15:59.190 "name": "pt3", 00:15:59.190 "base_bdev_name": "malloc3" 00:15:59.190 } 00:15:59.190 } 00:15:59.190 }' 00:15:59.190 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:59.190 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:15:59.190 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:15:59.190 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:59.190 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:15:59.448 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:15:59.448 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:59.448 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:15:59.448 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:15:59.448 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:59.448 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:15:59.448 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:15:59.448 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:15:59.448 11:26:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # jq -r '.[] | .uuid' 00:15:59.707 [2024-07-15 11:26:43.192232] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:59.707 11:26:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # '[' 9bd35c1a-b498-4611-8f66-2ad677fd6a6e '!=' 9bd35c1a-b498-4611-8f66-2ad677fd6a6e ']' 00:15:59.707 11:26:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@490 -- # has_redundancy concat 00:15:59.707 11:26:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:15:59.707 11:26:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@215 -- # return 1 00:15:59.707 11:26:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@562 -- # killprocess 904277 00:15:59.707 11:26:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@948 -- # '[' -z 904277 ']' 00:15:59.707 11:26:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@952 -- # kill -0 904277 00:15:59.707 11:26:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@953 -- # uname 00:15:59.707 11:26:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:15:59.707 11:26:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 904277 00:15:59.707 11:26:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:15:59.707 11:26:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:15:59.707 11:26:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 904277' 00:15:59.707 killing process with pid 904277 00:15:59.707 11:26:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@967 -- # kill 904277 00:15:59.707 [2024-07-15 11:26:43.267158] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:59.707 [2024-07-15 11:26:43.267214] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:59.707 [2024-07-15 11:26:43.267274] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:59.707 [2024-07-15 11:26:43.267287] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1988c00 name raid_bdev1, state offline 00:15:59.707 11:26:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # wait 904277 00:15:59.707 [2024-07-15 11:26:43.296490] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:59.965 11:26:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@564 -- # return 0 00:15:59.965 00:15:59.965 real 0m13.700s 00:15:59.965 user 0m24.584s 00:15:59.965 sys 0m2.512s 00:15:59.965 11:26:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:15:59.965 11:26:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.965 ************************************ 00:15:59.965 END TEST raid_superblock_test 00:15:59.966 ************************************ 00:16:00.224 11:26:43 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:16:00.224 11:26:43 bdev_raid -- bdev/bdev_raid.sh@870 -- # run_test raid_read_error_test raid_io_error_test concat 3 read 00:16:00.224 11:26:43 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:16:00.224 11:26:43 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:16:00.224 11:26:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:00.224 ************************************ 00:16:00.224 START TEST raid_read_error_test 00:16:00.224 ************************************ 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1123 -- # raid_io_error_test concat 3 read 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@788 -- # local raid_level=concat 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@789 -- # local num_base_bdevs=3 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local error_io_type=read 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i = 1 )) 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev1 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev2 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev3 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local base_bdevs 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local raid_bdev_name=raid_bdev1 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local strip_size 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local create_arg 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local bdevperf_log 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local fail_per_s 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # '[' concat '!=' raid1 ']' 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@799 -- # strip_size=64 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # create_arg+=' -z 64' 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@805 -- # mktemp -p /raidtest 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@805 -- # bdevperf_log=/raidtest/tmp.FfGsGtaPWA 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@808 -- # raid_pid=906368 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # waitforlisten 906368 /var/tmp/spdk-raid.sock 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@829 -- # '[' -z 906368 ']' 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:16:00.224 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:16:00.224 11:26:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.224 [2024-07-15 11:26:43.680230] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:16:00.224 [2024-07-15 11:26:43.680294] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid906368 ] 00:16:00.224 [2024-07-15 11:26:43.810986] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:00.484 [2024-07-15 11:26:43.921006] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:16:00.484 [2024-07-15 11:26:43.988550] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:00.484 [2024-07-15 11:26:43.988589] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:01.051 11:26:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:16:01.051 11:26:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@862 -- # return 0 00:16:01.051 11:26:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:16:01.051 11:26:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:16:01.311 BaseBdev1_malloc 00:16:01.311 11:26:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev1_malloc 00:16:01.570 true 00:16:01.570 11:26:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:16:01.830 [2024-07-15 11:26:45.319656] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:16:01.830 [2024-07-15 11:26:45.319702] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:01.830 [2024-07-15 11:26:45.319722] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xaaa0d0 00:16:01.830 [2024-07-15 11:26:45.319735] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:01.830 [2024-07-15 11:26:45.321584] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:01.830 [2024-07-15 11:26:45.321612] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:01.830 BaseBdev1 00:16:01.830 11:26:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:16:01.830 11:26:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:16:02.088 BaseBdev2_malloc 00:16:02.088 11:26:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev2_malloc 00:16:02.348 true 00:16:02.348 11:26:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:16:02.607 [2024-07-15 11:26:46.051477] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:16:02.607 [2024-07-15 11:26:46.051520] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:02.607 [2024-07-15 11:26:46.051541] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xaae910 00:16:02.607 [2024-07-15 11:26:46.051554] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:02.607 [2024-07-15 11:26:46.053124] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:02.607 [2024-07-15 11:26:46.053150] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:16:02.607 BaseBdev2 00:16:02.607 11:26:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:16:02.607 11:26:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:16:02.866 BaseBdev3_malloc 00:16:02.866 11:26:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev3_malloc 00:16:03.125 true 00:16:03.125 11:26:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:16:03.384 [2024-07-15 11:26:46.773913] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:16:03.384 [2024-07-15 11:26:46.773964] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:03.384 [2024-07-15 11:26:46.773985] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xab0bd0 00:16:03.384 [2024-07-15 11:26:46.773998] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:03.384 [2024-07-15 11:26:46.775605] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:03.384 [2024-07-15 11:26:46.775631] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:16:03.385 BaseBdev3 00:16:03.385 11:26:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@819 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n raid_bdev1 -s 00:16:03.643 [2024-07-15 11:26:47.006564] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:03.643 [2024-07-15 11:26:47.007935] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:03.643 [2024-07-15 11:26:47.008007] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:16:03.643 [2024-07-15 11:26:47.008222] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xab2280 00:16:03.643 [2024-07-15 11:26:47.008234] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:16:03.643 [2024-07-15 11:26:47.008430] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xab1e20 00:16:03.643 [2024-07-15 11:26:47.008578] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xab2280 00:16:03.643 [2024-07-15 11:26:47.008589] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0xab2280 00:16:03.643 [2024-07-15 11:26:47.008693] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:03.643 11:26:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@820 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:16:03.643 11:26:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:16:03.643 11:26:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:16:03.643 11:26:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:16:03.643 11:26:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:16:03.643 11:26:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:03.643 11:26:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:03.643 11:26:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:03.643 11:26:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:03.643 11:26:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:03.643 11:26:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:03.643 11:26:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:03.902 11:26:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:03.902 "name": "raid_bdev1", 00:16:03.902 "uuid": "c39d3717-c5d2-4987-a083-897a76fba90f", 00:16:03.902 "strip_size_kb": 64, 00:16:03.902 "state": "online", 00:16:03.902 "raid_level": "concat", 00:16:03.902 "superblock": true, 00:16:03.902 "num_base_bdevs": 3, 00:16:03.902 "num_base_bdevs_discovered": 3, 00:16:03.902 "num_base_bdevs_operational": 3, 00:16:03.902 "base_bdevs_list": [ 00:16:03.902 { 00:16:03.902 "name": "BaseBdev1", 00:16:03.902 "uuid": "0748c44c-1693-5cc2-a4d9-11ac9d936b99", 00:16:03.902 "is_configured": true, 00:16:03.902 "data_offset": 2048, 00:16:03.902 "data_size": 63488 00:16:03.902 }, 00:16:03.902 { 00:16:03.902 "name": "BaseBdev2", 00:16:03.902 "uuid": "4ee8774f-fc17-50f9-90d5-647cca1cf627", 00:16:03.902 "is_configured": true, 00:16:03.902 "data_offset": 2048, 00:16:03.902 "data_size": 63488 00:16:03.902 }, 00:16:03.902 { 00:16:03.902 "name": "BaseBdev3", 00:16:03.902 "uuid": "9d00e992-bd43-531c-87b7-3ef21d8a69ae", 00:16:03.902 "is_configured": true, 00:16:03.902 "data_offset": 2048, 00:16:03.902 "data_size": 63488 00:16:03.902 } 00:16:03.902 ] 00:16:03.902 }' 00:16:03.902 11:26:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:03.902 11:26:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:16:04.470 11:26:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@824 -- # sleep 1 00:16:04.470 11:26:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@823 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:16:04.470 [2024-07-15 11:26:47.929298] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x9004d0 00:16:05.423 11:26:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@827 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:16:05.705 11:26:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # local expected_num_base_bdevs 00:16:05.705 11:26:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@830 -- # [[ concat = \r\a\i\d\1 ]] 00:16:05.705 11:26:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=3 00:16:05.705 11:26:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:16:05.705 11:26:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:16:05.705 11:26:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:16:05.705 11:26:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:16:05.705 11:26:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:16:05.705 11:26:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:05.705 11:26:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:05.705 11:26:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:05.705 11:26:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:05.705 11:26:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:05.705 11:26:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:05.705 11:26:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:05.705 11:26:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:05.705 "name": "raid_bdev1", 00:16:05.705 "uuid": "c39d3717-c5d2-4987-a083-897a76fba90f", 00:16:05.705 "strip_size_kb": 64, 00:16:05.705 "state": "online", 00:16:05.705 "raid_level": "concat", 00:16:05.705 "superblock": true, 00:16:05.705 "num_base_bdevs": 3, 00:16:05.705 "num_base_bdevs_discovered": 3, 00:16:05.705 "num_base_bdevs_operational": 3, 00:16:05.705 "base_bdevs_list": [ 00:16:05.705 { 00:16:05.705 "name": "BaseBdev1", 00:16:05.705 "uuid": "0748c44c-1693-5cc2-a4d9-11ac9d936b99", 00:16:05.705 "is_configured": true, 00:16:05.705 "data_offset": 2048, 00:16:05.705 "data_size": 63488 00:16:05.705 }, 00:16:05.705 { 00:16:05.705 "name": "BaseBdev2", 00:16:05.705 "uuid": "4ee8774f-fc17-50f9-90d5-647cca1cf627", 00:16:05.705 "is_configured": true, 00:16:05.705 "data_offset": 2048, 00:16:05.705 "data_size": 63488 00:16:05.705 }, 00:16:05.705 { 00:16:05.705 "name": "BaseBdev3", 00:16:05.705 "uuid": "9d00e992-bd43-531c-87b7-3ef21d8a69ae", 00:16:05.705 "is_configured": true, 00:16:05.705 "data_offset": 2048, 00:16:05.705 "data_size": 63488 00:16:05.705 } 00:16:05.705 ] 00:16:05.705 }' 00:16:05.705 11:26:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:05.705 11:26:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:16:06.274 11:26:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:16:06.533 [2024-07-15 11:26:50.041741] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:06.533 [2024-07-15 11:26:50.041775] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:06.533 [2024-07-15 11:26:50.044943] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:06.533 [2024-07-15 11:26:50.044981] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:06.533 [2024-07-15 11:26:50.045022] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:06.533 [2024-07-15 11:26:50.045033] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xab2280 name raid_bdev1, state offline 00:16:06.533 0 00:16:06.533 11:26:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # killprocess 906368 00:16:06.533 11:26:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@948 -- # '[' -z 906368 ']' 00:16:06.533 11:26:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@952 -- # kill -0 906368 00:16:06.533 11:26:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@953 -- # uname 00:16:06.533 11:26:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:16:06.533 11:26:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 906368 00:16:06.533 11:26:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:16:06.533 11:26:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:16:06.533 11:26:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 906368' 00:16:06.533 killing process with pid 906368 00:16:06.533 11:26:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@967 -- # kill 906368 00:16:06.533 [2024-07-15 11:26:50.098217] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:06.533 11:26:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # wait 906368 00:16:06.533 [2024-07-15 11:26:50.118980] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:06.792 11:26:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # grep -v Job /raidtest/tmp.FfGsGtaPWA 00:16:06.792 11:26:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # grep raid_bdev1 00:16:06.792 11:26:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # awk '{print $6}' 00:16:06.792 11:26:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # fail_per_s=0.48 00:16:06.792 11:26:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@844 -- # has_redundancy concat 00:16:06.792 11:26:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:16:06.792 11:26:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@215 -- # return 1 00:16:06.792 11:26:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.48 != \0\.\0\0 ]] 00:16:06.792 00:16:06.792 real 0m6.748s 00:16:06.792 user 0m10.594s 00:16:06.792 sys 0m1.208s 00:16:06.792 11:26:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:16:06.792 11:26:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:16:06.792 ************************************ 00:16:06.792 END TEST raid_read_error_test 00:16:06.792 ************************************ 00:16:07.053 11:26:50 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:16:07.053 11:26:50 bdev_raid -- bdev/bdev_raid.sh@871 -- # run_test raid_write_error_test raid_io_error_test concat 3 write 00:16:07.053 11:26:50 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:16:07.053 11:26:50 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:16:07.053 11:26:50 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:07.053 ************************************ 00:16:07.053 START TEST raid_write_error_test 00:16:07.053 ************************************ 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1123 -- # raid_io_error_test concat 3 write 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@788 -- # local raid_level=concat 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@789 -- # local num_base_bdevs=3 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local error_io_type=write 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i = 1 )) 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev1 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev2 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev3 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local base_bdevs 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local raid_bdev_name=raid_bdev1 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local strip_size 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local create_arg 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local bdevperf_log 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local fail_per_s 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # '[' concat '!=' raid1 ']' 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@799 -- # strip_size=64 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # create_arg+=' -z 64' 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@805 -- # mktemp -p /raidtest 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@805 -- # bdevperf_log=/raidtest/tmp.4JN0ntUyWs 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@808 -- # raid_pid=907347 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # waitforlisten 907347 /var/tmp/spdk-raid.sock 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@829 -- # '[' -z 907347 ']' 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:16:07.053 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:16:07.053 11:26:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:16:07.053 [2024-07-15 11:26:50.506165] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:16:07.053 [2024-07-15 11:26:50.506230] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid907347 ] 00:16:07.053 [2024-07-15 11:26:50.636101] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:07.312 [2024-07-15 11:26:50.742986] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:16:07.312 [2024-07-15 11:26:50.806117] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:07.312 [2024-07-15 11:26:50.806148] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:07.880 11:26:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:16:07.880 11:26:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@862 -- # return 0 00:16:07.880 11:26:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:16:07.880 11:26:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:16:08.139 BaseBdev1_malloc 00:16:08.139 11:26:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev1_malloc 00:16:08.398 true 00:16:08.398 11:26:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:16:08.657 [2024-07-15 11:26:52.135361] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:16:08.657 [2024-07-15 11:26:52.135405] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:08.657 [2024-07-15 11:26:52.135426] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1e300d0 00:16:08.657 [2024-07-15 11:26:52.135439] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:08.657 [2024-07-15 11:26:52.137442] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:08.657 [2024-07-15 11:26:52.137472] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:08.657 BaseBdev1 00:16:08.657 11:26:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:16:08.657 11:26:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:16:08.916 BaseBdev2_malloc 00:16:08.916 11:26:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev2_malloc 00:16:09.174 true 00:16:09.174 11:26:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:16:09.433 [2024-07-15 11:26:52.871094] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:16:09.433 [2024-07-15 11:26:52.871139] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:09.433 [2024-07-15 11:26:52.871160] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1e34910 00:16:09.433 [2024-07-15 11:26:52.871172] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:09.433 [2024-07-15 11:26:52.872777] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:09.433 [2024-07-15 11:26:52.872807] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:16:09.433 BaseBdev2 00:16:09.433 11:26:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:16:09.433 11:26:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:16:09.693 BaseBdev3_malloc 00:16:09.693 11:26:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev3_malloc 00:16:09.952 true 00:16:09.952 11:26:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:16:10.210 [2024-07-15 11:26:53.605582] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:16:10.210 [2024-07-15 11:26:53.605625] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:10.210 [2024-07-15 11:26:53.605644] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1e36bd0 00:16:10.210 [2024-07-15 11:26:53.605657] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:10.210 [2024-07-15 11:26:53.607235] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:10.210 [2024-07-15 11:26:53.607261] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:16:10.210 BaseBdev3 00:16:10.210 11:26:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@819 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n raid_bdev1 -s 00:16:10.468 [2024-07-15 11:26:53.846246] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:10.468 [2024-07-15 11:26:53.847618] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:10.468 [2024-07-15 11:26:53.847688] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:16:10.468 [2024-07-15 11:26:53.847900] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1e38280 00:16:10.468 [2024-07-15 11:26:53.847912] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:16:10.468 [2024-07-15 11:26:53.848125] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1e37e20 00:16:10.468 [2024-07-15 11:26:53.848275] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1e38280 00:16:10.468 [2024-07-15 11:26:53.848285] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x1e38280 00:16:10.468 [2024-07-15 11:26:53.848390] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:10.468 11:26:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@820 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:16:10.468 11:26:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:16:10.468 11:26:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:16:10.468 11:26:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:16:10.468 11:26:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:16:10.468 11:26:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:10.468 11:26:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:10.468 11:26:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:10.468 11:26:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:10.468 11:26:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:10.468 11:26:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:10.468 11:26:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:10.726 11:26:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:10.726 "name": "raid_bdev1", 00:16:10.726 "uuid": "867f45ee-fd61-4105-9de2-511160387f06", 00:16:10.726 "strip_size_kb": 64, 00:16:10.726 "state": "online", 00:16:10.726 "raid_level": "concat", 00:16:10.726 "superblock": true, 00:16:10.726 "num_base_bdevs": 3, 00:16:10.726 "num_base_bdevs_discovered": 3, 00:16:10.726 "num_base_bdevs_operational": 3, 00:16:10.726 "base_bdevs_list": [ 00:16:10.726 { 00:16:10.726 "name": "BaseBdev1", 00:16:10.726 "uuid": "a8a62604-4d1c-544a-8ddc-614d35c09192", 00:16:10.726 "is_configured": true, 00:16:10.727 "data_offset": 2048, 00:16:10.727 "data_size": 63488 00:16:10.727 }, 00:16:10.727 { 00:16:10.727 "name": "BaseBdev2", 00:16:10.727 "uuid": "aece5d16-a9ab-5ade-a723-b2125e60617c", 00:16:10.727 "is_configured": true, 00:16:10.727 "data_offset": 2048, 00:16:10.727 "data_size": 63488 00:16:10.727 }, 00:16:10.727 { 00:16:10.727 "name": "BaseBdev3", 00:16:10.727 "uuid": "21911861-9f3a-572e-9e29-6c45af5946b2", 00:16:10.727 "is_configured": true, 00:16:10.727 "data_offset": 2048, 00:16:10.727 "data_size": 63488 00:16:10.727 } 00:16:10.727 ] 00:16:10.727 }' 00:16:10.727 11:26:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:10.727 11:26:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:16:11.294 11:26:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@824 -- # sleep 1 00:16:11.294 11:26:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@823 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:16:11.294 [2024-07-15 11:26:54.829125] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1c864d0 00:16:12.232 11:26:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@827 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:16:12.491 11:26:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # local expected_num_base_bdevs 00:16:12.491 11:26:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@830 -- # [[ concat = \r\a\i\d\1 ]] 00:16:12.491 11:26:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=3 00:16:12.491 11:26:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:16:12.491 11:26:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:16:12.491 11:26:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:16:12.491 11:26:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:16:12.491 11:26:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:16:12.491 11:26:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:12.491 11:26:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:12.491 11:26:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:12.491 11:26:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:12.491 11:26:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:12.491 11:26:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:12.491 11:26:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:12.749 11:26:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:12.749 "name": "raid_bdev1", 00:16:12.749 "uuid": "867f45ee-fd61-4105-9de2-511160387f06", 00:16:12.749 "strip_size_kb": 64, 00:16:12.749 "state": "online", 00:16:12.749 "raid_level": "concat", 00:16:12.749 "superblock": true, 00:16:12.749 "num_base_bdevs": 3, 00:16:12.750 "num_base_bdevs_discovered": 3, 00:16:12.750 "num_base_bdevs_operational": 3, 00:16:12.750 "base_bdevs_list": [ 00:16:12.750 { 00:16:12.750 "name": "BaseBdev1", 00:16:12.750 "uuid": "a8a62604-4d1c-544a-8ddc-614d35c09192", 00:16:12.750 "is_configured": true, 00:16:12.750 "data_offset": 2048, 00:16:12.750 "data_size": 63488 00:16:12.750 }, 00:16:12.750 { 00:16:12.750 "name": "BaseBdev2", 00:16:12.750 "uuid": "aece5d16-a9ab-5ade-a723-b2125e60617c", 00:16:12.750 "is_configured": true, 00:16:12.750 "data_offset": 2048, 00:16:12.750 "data_size": 63488 00:16:12.750 }, 00:16:12.750 { 00:16:12.750 "name": "BaseBdev3", 00:16:12.750 "uuid": "21911861-9f3a-572e-9e29-6c45af5946b2", 00:16:12.750 "is_configured": true, 00:16:12.750 "data_offset": 2048, 00:16:12.750 "data_size": 63488 00:16:12.750 } 00:16:12.750 ] 00:16:12.750 }' 00:16:12.750 11:26:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:12.750 11:26:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:16:13.316 11:26:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:16:13.575 [2024-07-15 11:26:57.070815] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:13.575 [2024-07-15 11:26:57.070849] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:13.575 [2024-07-15 11:26:57.074065] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:13.575 [2024-07-15 11:26:57.074104] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:13.575 [2024-07-15 11:26:57.074140] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:13.575 [2024-07-15 11:26:57.074151] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1e38280 name raid_bdev1, state offline 00:16:13.575 0 00:16:13.575 11:26:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # killprocess 907347 00:16:13.575 11:26:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@948 -- # '[' -z 907347 ']' 00:16:13.575 11:26:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@952 -- # kill -0 907347 00:16:13.575 11:26:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@953 -- # uname 00:16:13.575 11:26:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:16:13.575 11:26:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 907347 00:16:13.575 11:26:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:16:13.575 11:26:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:16:13.575 11:26:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 907347' 00:16:13.575 killing process with pid 907347 00:16:13.575 11:26:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@967 -- # kill 907347 00:16:13.575 [2024-07-15 11:26:57.140498] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:13.575 11:26:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # wait 907347 00:16:13.575 [2024-07-15 11:26:57.162689] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:13.833 11:26:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # grep -v Job /raidtest/tmp.4JN0ntUyWs 00:16:13.833 11:26:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # grep raid_bdev1 00:16:13.833 11:26:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # awk '{print $6}' 00:16:13.833 11:26:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # fail_per_s=0.45 00:16:13.833 11:26:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@844 -- # has_redundancy concat 00:16:13.833 11:26:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:16:13.833 11:26:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@215 -- # return 1 00:16:13.833 11:26:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.45 != \0\.\0\0 ]] 00:16:13.833 00:16:13.833 real 0m6.977s 00:16:13.834 user 0m11.058s 00:16:13.834 sys 0m1.203s 00:16:13.834 11:26:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:16:13.834 11:26:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:16:13.834 ************************************ 00:16:13.834 END TEST raid_write_error_test 00:16:13.834 ************************************ 00:16:14.092 11:26:57 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:16:14.092 11:26:57 bdev_raid -- bdev/bdev_raid.sh@866 -- # for level in raid0 concat raid1 00:16:14.092 11:26:57 bdev_raid -- bdev/bdev_raid.sh@867 -- # run_test raid_state_function_test raid_state_function_test raid1 3 false 00:16:14.092 11:26:57 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:16:14.092 11:26:57 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:16:14.092 11:26:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:14.092 ************************************ 00:16:14.092 START TEST raid_state_function_test 00:16:14.092 ************************************ 00:16:14.092 11:26:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1123 -- # raid_state_function_test raid1 3 false 00:16:14.092 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@220 -- # local raid_level=raid1 00:16:14.092 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=3 00:16:14.092 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # local superblock=false 00:16:14.092 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:16:14.092 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev3 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # local strip_size 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # '[' raid1 '!=' raid1 ']' 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@234 -- # strip_size=0 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # '[' false = true ']' 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@240 -- # superblock_create_arg= 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # raid_pid=908324 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 908324' 00:16:14.093 Process raid pid: 908324 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@246 -- # waitforlisten 908324 /var/tmp/spdk-raid.sock 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@829 -- # '[' -z 908324 ']' 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:16:14.093 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:16:14.093 11:26:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:16:14.093 [2024-07-15 11:26:57.564000] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:16:14.093 [2024-07-15 11:26:57.564054] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:14.093 [2024-07-15 11:26:57.677816] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:14.352 [2024-07-15 11:26:57.781287] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:16:14.352 [2024-07-15 11:26:57.842802] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:14.352 [2024-07-15 11:26:57.842839] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:14.918 11:26:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:16:14.918 11:26:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@862 -- # return 0 00:16:14.919 11:26:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:16:15.177 [2024-07-15 11:26:58.720657] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:15.177 [2024-07-15 11:26:58.720702] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:15.177 [2024-07-15 11:26:58.720713] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:15.177 [2024-07-15 11:26:58.720725] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:15.177 [2024-07-15 11:26:58.720734] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:16:15.177 [2024-07-15 11:26:58.720745] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:16:15.177 11:26:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:16:15.177 11:26:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:15.177 11:26:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:16:15.177 11:26:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:15.177 11:26:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:15.177 11:26:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:15.177 11:26:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:15.177 11:26:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:15.177 11:26:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:15.177 11:26:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:15.177 11:26:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:15.177 11:26:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:15.435 11:26:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:15.435 "name": "Existed_Raid", 00:16:15.435 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:15.435 "strip_size_kb": 0, 00:16:15.435 "state": "configuring", 00:16:15.435 "raid_level": "raid1", 00:16:15.435 "superblock": false, 00:16:15.435 "num_base_bdevs": 3, 00:16:15.435 "num_base_bdevs_discovered": 0, 00:16:15.435 "num_base_bdevs_operational": 3, 00:16:15.435 "base_bdevs_list": [ 00:16:15.435 { 00:16:15.435 "name": "BaseBdev1", 00:16:15.435 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:15.435 "is_configured": false, 00:16:15.435 "data_offset": 0, 00:16:15.435 "data_size": 0 00:16:15.435 }, 00:16:15.435 { 00:16:15.435 "name": "BaseBdev2", 00:16:15.435 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:15.435 "is_configured": false, 00:16:15.435 "data_offset": 0, 00:16:15.435 "data_size": 0 00:16:15.435 }, 00:16:15.435 { 00:16:15.435 "name": "BaseBdev3", 00:16:15.436 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:15.436 "is_configured": false, 00:16:15.436 "data_offset": 0, 00:16:15.436 "data_size": 0 00:16:15.436 } 00:16:15.436 ] 00:16:15.436 }' 00:16:15.436 11:26:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:15.436 11:26:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:16:16.002 11:26:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:16:16.261 [2024-07-15 11:26:59.819423] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:16.261 [2024-07-15 11:26:59.819452] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x195ea80 name Existed_Raid, state configuring 00:16:16.261 11:26:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:16:16.518 [2024-07-15 11:27:00.076122] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:16.518 [2024-07-15 11:27:00.076161] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:16.518 [2024-07-15 11:27:00.076171] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:16.518 [2024-07-15 11:27:00.076182] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:16.518 [2024-07-15 11:27:00.076191] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:16:16.518 [2024-07-15 11:27:00.076202] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:16:16.518 11:27:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:16:16.777 [2024-07-15 11:27:00.330673] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:16.777 BaseBdev1 00:16:16.777 11:27:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:16:16.777 11:27:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:16:16.777 11:27:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:16:16.777 11:27:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:16:16.777 11:27:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:16:16.777 11:27:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:16:16.777 11:27:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:16:17.036 11:27:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:16:17.294 [ 00:16:17.294 { 00:16:17.294 "name": "BaseBdev1", 00:16:17.294 "aliases": [ 00:16:17.294 "382f59dd-8bfb-42a7-9e3f-302dd6ed652c" 00:16:17.294 ], 00:16:17.294 "product_name": "Malloc disk", 00:16:17.294 "block_size": 512, 00:16:17.294 "num_blocks": 65536, 00:16:17.294 "uuid": "382f59dd-8bfb-42a7-9e3f-302dd6ed652c", 00:16:17.294 "assigned_rate_limits": { 00:16:17.294 "rw_ios_per_sec": 0, 00:16:17.294 "rw_mbytes_per_sec": 0, 00:16:17.294 "r_mbytes_per_sec": 0, 00:16:17.294 "w_mbytes_per_sec": 0 00:16:17.294 }, 00:16:17.294 "claimed": true, 00:16:17.294 "claim_type": "exclusive_write", 00:16:17.294 "zoned": false, 00:16:17.294 "supported_io_types": { 00:16:17.294 "read": true, 00:16:17.294 "write": true, 00:16:17.294 "unmap": true, 00:16:17.294 "flush": true, 00:16:17.294 "reset": true, 00:16:17.294 "nvme_admin": false, 00:16:17.294 "nvme_io": false, 00:16:17.294 "nvme_io_md": false, 00:16:17.294 "write_zeroes": true, 00:16:17.294 "zcopy": true, 00:16:17.294 "get_zone_info": false, 00:16:17.294 "zone_management": false, 00:16:17.294 "zone_append": false, 00:16:17.294 "compare": false, 00:16:17.294 "compare_and_write": false, 00:16:17.294 "abort": true, 00:16:17.294 "seek_hole": false, 00:16:17.294 "seek_data": false, 00:16:17.294 "copy": true, 00:16:17.294 "nvme_iov_md": false 00:16:17.294 }, 00:16:17.294 "memory_domains": [ 00:16:17.294 { 00:16:17.294 "dma_device_id": "system", 00:16:17.294 "dma_device_type": 1 00:16:17.294 }, 00:16:17.294 { 00:16:17.294 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:17.294 "dma_device_type": 2 00:16:17.294 } 00:16:17.294 ], 00:16:17.294 "driver_specific": {} 00:16:17.294 } 00:16:17.294 ] 00:16:17.294 11:27:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:16:17.294 11:27:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:16:17.294 11:27:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:17.294 11:27:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:16:17.294 11:27:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:17.294 11:27:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:17.294 11:27:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:17.294 11:27:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:17.294 11:27:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:17.294 11:27:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:17.294 11:27:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:17.294 11:27:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:17.294 11:27:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:17.553 11:27:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:17.553 "name": "Existed_Raid", 00:16:17.553 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:17.553 "strip_size_kb": 0, 00:16:17.553 "state": "configuring", 00:16:17.553 "raid_level": "raid1", 00:16:17.553 "superblock": false, 00:16:17.553 "num_base_bdevs": 3, 00:16:17.553 "num_base_bdevs_discovered": 1, 00:16:17.553 "num_base_bdevs_operational": 3, 00:16:17.553 "base_bdevs_list": [ 00:16:17.553 { 00:16:17.553 "name": "BaseBdev1", 00:16:17.553 "uuid": "382f59dd-8bfb-42a7-9e3f-302dd6ed652c", 00:16:17.553 "is_configured": true, 00:16:17.553 "data_offset": 0, 00:16:17.553 "data_size": 65536 00:16:17.553 }, 00:16:17.553 { 00:16:17.553 "name": "BaseBdev2", 00:16:17.553 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:17.553 "is_configured": false, 00:16:17.553 "data_offset": 0, 00:16:17.553 "data_size": 0 00:16:17.553 }, 00:16:17.553 { 00:16:17.553 "name": "BaseBdev3", 00:16:17.553 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:17.553 "is_configured": false, 00:16:17.553 "data_offset": 0, 00:16:17.553 "data_size": 0 00:16:17.553 } 00:16:17.553 ] 00:16:17.553 }' 00:16:17.553 11:27:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:17.553 11:27:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:16:18.120 11:27:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:16:18.379 [2024-07-15 11:27:01.774506] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:18.379 [2024-07-15 11:27:01.774546] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x195e310 name Existed_Raid, state configuring 00:16:18.379 11:27:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:16:18.639 [2024-07-15 11:27:02.023195] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:18.639 [2024-07-15 11:27:02.024660] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:18.639 [2024-07-15 11:27:02.024695] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:18.639 [2024-07-15 11:27:02.024705] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:16:18.639 [2024-07-15 11:27:02.024716] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:16:18.639 11:27:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:16:18.639 11:27:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:16:18.639 11:27:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:16:18.639 11:27:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:18.639 11:27:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:16:18.639 11:27:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:18.639 11:27:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:18.639 11:27:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:18.639 11:27:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:18.639 11:27:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:18.639 11:27:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:18.639 11:27:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:18.639 11:27:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:18.639 11:27:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:18.898 11:27:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:18.898 "name": "Existed_Raid", 00:16:18.898 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:18.898 "strip_size_kb": 0, 00:16:18.898 "state": "configuring", 00:16:18.898 "raid_level": "raid1", 00:16:18.898 "superblock": false, 00:16:18.898 "num_base_bdevs": 3, 00:16:18.898 "num_base_bdevs_discovered": 1, 00:16:18.898 "num_base_bdevs_operational": 3, 00:16:18.898 "base_bdevs_list": [ 00:16:18.898 { 00:16:18.898 "name": "BaseBdev1", 00:16:18.898 "uuid": "382f59dd-8bfb-42a7-9e3f-302dd6ed652c", 00:16:18.898 "is_configured": true, 00:16:18.898 "data_offset": 0, 00:16:18.898 "data_size": 65536 00:16:18.898 }, 00:16:18.898 { 00:16:18.898 "name": "BaseBdev2", 00:16:18.898 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:18.898 "is_configured": false, 00:16:18.898 "data_offset": 0, 00:16:18.898 "data_size": 0 00:16:18.898 }, 00:16:18.898 { 00:16:18.898 "name": "BaseBdev3", 00:16:18.898 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:18.898 "is_configured": false, 00:16:18.898 "data_offset": 0, 00:16:18.898 "data_size": 0 00:16:18.898 } 00:16:18.898 ] 00:16:18.898 }' 00:16:18.898 11:27:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:18.898 11:27:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:16:19.465 11:27:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:16:19.770 [2024-07-15 11:27:03.097513] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:19.770 BaseBdev2 00:16:19.770 11:27:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:16:19.770 11:27:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:16:19.770 11:27:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:16:19.770 11:27:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:16:19.770 11:27:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:16:19.770 11:27:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:16:19.770 11:27:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:16:19.770 11:27:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:16:20.045 [ 00:16:20.045 { 00:16:20.045 "name": "BaseBdev2", 00:16:20.045 "aliases": [ 00:16:20.045 "91c19df1-59e3-4895-ba71-79ced3520be9" 00:16:20.045 ], 00:16:20.045 "product_name": "Malloc disk", 00:16:20.045 "block_size": 512, 00:16:20.045 "num_blocks": 65536, 00:16:20.045 "uuid": "91c19df1-59e3-4895-ba71-79ced3520be9", 00:16:20.045 "assigned_rate_limits": { 00:16:20.045 "rw_ios_per_sec": 0, 00:16:20.045 "rw_mbytes_per_sec": 0, 00:16:20.045 "r_mbytes_per_sec": 0, 00:16:20.045 "w_mbytes_per_sec": 0 00:16:20.045 }, 00:16:20.045 "claimed": true, 00:16:20.045 "claim_type": "exclusive_write", 00:16:20.045 "zoned": false, 00:16:20.045 "supported_io_types": { 00:16:20.045 "read": true, 00:16:20.045 "write": true, 00:16:20.045 "unmap": true, 00:16:20.045 "flush": true, 00:16:20.045 "reset": true, 00:16:20.045 "nvme_admin": false, 00:16:20.045 "nvme_io": false, 00:16:20.045 "nvme_io_md": false, 00:16:20.045 "write_zeroes": true, 00:16:20.045 "zcopy": true, 00:16:20.045 "get_zone_info": false, 00:16:20.045 "zone_management": false, 00:16:20.045 "zone_append": false, 00:16:20.045 "compare": false, 00:16:20.045 "compare_and_write": false, 00:16:20.045 "abort": true, 00:16:20.045 "seek_hole": false, 00:16:20.045 "seek_data": false, 00:16:20.045 "copy": true, 00:16:20.045 "nvme_iov_md": false 00:16:20.045 }, 00:16:20.045 "memory_domains": [ 00:16:20.045 { 00:16:20.045 "dma_device_id": "system", 00:16:20.045 "dma_device_type": 1 00:16:20.045 }, 00:16:20.045 { 00:16:20.045 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:20.045 "dma_device_type": 2 00:16:20.045 } 00:16:20.045 ], 00:16:20.045 "driver_specific": {} 00:16:20.045 } 00:16:20.045 ] 00:16:20.045 11:27:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:16:20.045 11:27:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:16:20.045 11:27:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:16:20.045 11:27:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:16:20.045 11:27:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:20.045 11:27:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:16:20.045 11:27:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:20.045 11:27:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:20.045 11:27:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:20.045 11:27:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:20.045 11:27:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:20.045 11:27:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:20.045 11:27:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:20.045 11:27:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:20.045 11:27:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:20.303 11:27:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:20.303 "name": "Existed_Raid", 00:16:20.303 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:20.303 "strip_size_kb": 0, 00:16:20.303 "state": "configuring", 00:16:20.303 "raid_level": "raid1", 00:16:20.303 "superblock": false, 00:16:20.303 "num_base_bdevs": 3, 00:16:20.303 "num_base_bdevs_discovered": 2, 00:16:20.303 "num_base_bdevs_operational": 3, 00:16:20.303 "base_bdevs_list": [ 00:16:20.303 { 00:16:20.303 "name": "BaseBdev1", 00:16:20.303 "uuid": "382f59dd-8bfb-42a7-9e3f-302dd6ed652c", 00:16:20.303 "is_configured": true, 00:16:20.303 "data_offset": 0, 00:16:20.303 "data_size": 65536 00:16:20.303 }, 00:16:20.303 { 00:16:20.303 "name": "BaseBdev2", 00:16:20.303 "uuid": "91c19df1-59e3-4895-ba71-79ced3520be9", 00:16:20.303 "is_configured": true, 00:16:20.303 "data_offset": 0, 00:16:20.303 "data_size": 65536 00:16:20.303 }, 00:16:20.303 { 00:16:20.303 "name": "BaseBdev3", 00:16:20.303 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:20.303 "is_configured": false, 00:16:20.303 "data_offset": 0, 00:16:20.304 "data_size": 0 00:16:20.304 } 00:16:20.304 ] 00:16:20.304 }' 00:16:20.304 11:27:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:20.304 11:27:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:16:20.871 11:27:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:16:21.129 [2024-07-15 11:27:04.576820] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:16:21.129 [2024-07-15 11:27:04.576860] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x195f400 00:16:21.129 [2024-07-15 11:27:04.576869] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:16:21.129 [2024-07-15 11:27:04.577130] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x195eef0 00:16:21.129 [2024-07-15 11:27:04.577258] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x195f400 00:16:21.129 [2024-07-15 11:27:04.577268] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x195f400 00:16:21.129 [2024-07-15 11:27:04.577433] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:21.129 BaseBdev3 00:16:21.129 11:27:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev3 00:16:21.129 11:27:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:16:21.129 11:27:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:16:21.129 11:27:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:16:21.129 11:27:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:16:21.129 11:27:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:16:21.129 11:27:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:16:21.388 11:27:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:16:21.647 [ 00:16:21.647 { 00:16:21.647 "name": "BaseBdev3", 00:16:21.647 "aliases": [ 00:16:21.647 "657b8c71-a29c-4810-827f-bd3d6112282e" 00:16:21.647 ], 00:16:21.647 "product_name": "Malloc disk", 00:16:21.647 "block_size": 512, 00:16:21.647 "num_blocks": 65536, 00:16:21.647 "uuid": "657b8c71-a29c-4810-827f-bd3d6112282e", 00:16:21.647 "assigned_rate_limits": { 00:16:21.647 "rw_ios_per_sec": 0, 00:16:21.647 "rw_mbytes_per_sec": 0, 00:16:21.647 "r_mbytes_per_sec": 0, 00:16:21.647 "w_mbytes_per_sec": 0 00:16:21.647 }, 00:16:21.647 "claimed": true, 00:16:21.647 "claim_type": "exclusive_write", 00:16:21.647 "zoned": false, 00:16:21.647 "supported_io_types": { 00:16:21.647 "read": true, 00:16:21.647 "write": true, 00:16:21.647 "unmap": true, 00:16:21.647 "flush": true, 00:16:21.647 "reset": true, 00:16:21.647 "nvme_admin": false, 00:16:21.647 "nvme_io": false, 00:16:21.647 "nvme_io_md": false, 00:16:21.647 "write_zeroes": true, 00:16:21.647 "zcopy": true, 00:16:21.647 "get_zone_info": false, 00:16:21.647 "zone_management": false, 00:16:21.647 "zone_append": false, 00:16:21.647 "compare": false, 00:16:21.647 "compare_and_write": false, 00:16:21.647 "abort": true, 00:16:21.647 "seek_hole": false, 00:16:21.647 "seek_data": false, 00:16:21.647 "copy": true, 00:16:21.647 "nvme_iov_md": false 00:16:21.647 }, 00:16:21.647 "memory_domains": [ 00:16:21.647 { 00:16:21.647 "dma_device_id": "system", 00:16:21.647 "dma_device_type": 1 00:16:21.647 }, 00:16:21.647 { 00:16:21.647 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:21.647 "dma_device_type": 2 00:16:21.647 } 00:16:21.647 ], 00:16:21.647 "driver_specific": {} 00:16:21.647 } 00:16:21.647 ] 00:16:21.648 11:27:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:16:21.648 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:16:21.648 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:16:21.648 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:16:21.648 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:21.648 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:16:21.648 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:21.648 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:21.648 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:21.648 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:21.648 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:21.648 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:21.648 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:21.648 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:21.648 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:21.907 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:21.907 "name": "Existed_Raid", 00:16:21.907 "uuid": "78c823d2-837c-40e1-9407-d5d1a9819a18", 00:16:21.907 "strip_size_kb": 0, 00:16:21.907 "state": "online", 00:16:21.907 "raid_level": "raid1", 00:16:21.907 "superblock": false, 00:16:21.907 "num_base_bdevs": 3, 00:16:21.907 "num_base_bdevs_discovered": 3, 00:16:21.907 "num_base_bdevs_operational": 3, 00:16:21.907 "base_bdevs_list": [ 00:16:21.907 { 00:16:21.907 "name": "BaseBdev1", 00:16:21.907 "uuid": "382f59dd-8bfb-42a7-9e3f-302dd6ed652c", 00:16:21.907 "is_configured": true, 00:16:21.907 "data_offset": 0, 00:16:21.907 "data_size": 65536 00:16:21.907 }, 00:16:21.907 { 00:16:21.907 "name": "BaseBdev2", 00:16:21.907 "uuid": "91c19df1-59e3-4895-ba71-79ced3520be9", 00:16:21.907 "is_configured": true, 00:16:21.907 "data_offset": 0, 00:16:21.907 "data_size": 65536 00:16:21.907 }, 00:16:21.907 { 00:16:21.907 "name": "BaseBdev3", 00:16:21.907 "uuid": "657b8c71-a29c-4810-827f-bd3d6112282e", 00:16:21.907 "is_configured": true, 00:16:21.907 "data_offset": 0, 00:16:21.907 "data_size": 65536 00:16:21.907 } 00:16:21.907 ] 00:16:21.907 }' 00:16:21.907 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:21.907 11:27:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:16:22.475 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:16:22.475 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:16:22.475 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:16:22.475 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:16:22.475 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:16:22.475 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # local name 00:16:22.475 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:16:22.475 11:27:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:16:22.734 [2024-07-15 11:27:06.109212] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:22.734 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:16:22.734 "name": "Existed_Raid", 00:16:22.734 "aliases": [ 00:16:22.734 "78c823d2-837c-40e1-9407-d5d1a9819a18" 00:16:22.734 ], 00:16:22.734 "product_name": "Raid Volume", 00:16:22.734 "block_size": 512, 00:16:22.734 "num_blocks": 65536, 00:16:22.734 "uuid": "78c823d2-837c-40e1-9407-d5d1a9819a18", 00:16:22.734 "assigned_rate_limits": { 00:16:22.734 "rw_ios_per_sec": 0, 00:16:22.734 "rw_mbytes_per_sec": 0, 00:16:22.734 "r_mbytes_per_sec": 0, 00:16:22.734 "w_mbytes_per_sec": 0 00:16:22.734 }, 00:16:22.734 "claimed": false, 00:16:22.734 "zoned": false, 00:16:22.734 "supported_io_types": { 00:16:22.734 "read": true, 00:16:22.734 "write": true, 00:16:22.734 "unmap": false, 00:16:22.734 "flush": false, 00:16:22.734 "reset": true, 00:16:22.734 "nvme_admin": false, 00:16:22.734 "nvme_io": false, 00:16:22.734 "nvme_io_md": false, 00:16:22.734 "write_zeroes": true, 00:16:22.734 "zcopy": false, 00:16:22.734 "get_zone_info": false, 00:16:22.734 "zone_management": false, 00:16:22.734 "zone_append": false, 00:16:22.734 "compare": false, 00:16:22.734 "compare_and_write": false, 00:16:22.734 "abort": false, 00:16:22.734 "seek_hole": false, 00:16:22.734 "seek_data": false, 00:16:22.734 "copy": false, 00:16:22.734 "nvme_iov_md": false 00:16:22.734 }, 00:16:22.734 "memory_domains": [ 00:16:22.734 { 00:16:22.734 "dma_device_id": "system", 00:16:22.734 "dma_device_type": 1 00:16:22.734 }, 00:16:22.734 { 00:16:22.734 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:22.735 "dma_device_type": 2 00:16:22.735 }, 00:16:22.735 { 00:16:22.735 "dma_device_id": "system", 00:16:22.735 "dma_device_type": 1 00:16:22.735 }, 00:16:22.735 { 00:16:22.735 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:22.735 "dma_device_type": 2 00:16:22.735 }, 00:16:22.735 { 00:16:22.735 "dma_device_id": "system", 00:16:22.735 "dma_device_type": 1 00:16:22.735 }, 00:16:22.735 { 00:16:22.735 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:22.735 "dma_device_type": 2 00:16:22.735 } 00:16:22.735 ], 00:16:22.735 "driver_specific": { 00:16:22.735 "raid": { 00:16:22.735 "uuid": "78c823d2-837c-40e1-9407-d5d1a9819a18", 00:16:22.735 "strip_size_kb": 0, 00:16:22.735 "state": "online", 00:16:22.735 "raid_level": "raid1", 00:16:22.735 "superblock": false, 00:16:22.735 "num_base_bdevs": 3, 00:16:22.735 "num_base_bdevs_discovered": 3, 00:16:22.735 "num_base_bdevs_operational": 3, 00:16:22.735 "base_bdevs_list": [ 00:16:22.735 { 00:16:22.735 "name": "BaseBdev1", 00:16:22.735 "uuid": "382f59dd-8bfb-42a7-9e3f-302dd6ed652c", 00:16:22.735 "is_configured": true, 00:16:22.735 "data_offset": 0, 00:16:22.735 "data_size": 65536 00:16:22.735 }, 00:16:22.735 { 00:16:22.735 "name": "BaseBdev2", 00:16:22.735 "uuid": "91c19df1-59e3-4895-ba71-79ced3520be9", 00:16:22.735 "is_configured": true, 00:16:22.735 "data_offset": 0, 00:16:22.735 "data_size": 65536 00:16:22.735 }, 00:16:22.735 { 00:16:22.735 "name": "BaseBdev3", 00:16:22.735 "uuid": "657b8c71-a29c-4810-827f-bd3d6112282e", 00:16:22.735 "is_configured": true, 00:16:22.735 "data_offset": 0, 00:16:22.735 "data_size": 65536 00:16:22.735 } 00:16:22.735 ] 00:16:22.735 } 00:16:22.735 } 00:16:22.735 }' 00:16:22.735 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:22.735 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:16:22.735 BaseBdev2 00:16:22.735 BaseBdev3' 00:16:22.735 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:16:22.735 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:16:22.735 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:16:23.008 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:16:23.009 "name": "BaseBdev1", 00:16:23.009 "aliases": [ 00:16:23.009 "382f59dd-8bfb-42a7-9e3f-302dd6ed652c" 00:16:23.009 ], 00:16:23.009 "product_name": "Malloc disk", 00:16:23.009 "block_size": 512, 00:16:23.009 "num_blocks": 65536, 00:16:23.009 "uuid": "382f59dd-8bfb-42a7-9e3f-302dd6ed652c", 00:16:23.009 "assigned_rate_limits": { 00:16:23.009 "rw_ios_per_sec": 0, 00:16:23.009 "rw_mbytes_per_sec": 0, 00:16:23.009 "r_mbytes_per_sec": 0, 00:16:23.009 "w_mbytes_per_sec": 0 00:16:23.009 }, 00:16:23.009 "claimed": true, 00:16:23.009 "claim_type": "exclusive_write", 00:16:23.009 "zoned": false, 00:16:23.009 "supported_io_types": { 00:16:23.009 "read": true, 00:16:23.009 "write": true, 00:16:23.009 "unmap": true, 00:16:23.009 "flush": true, 00:16:23.009 "reset": true, 00:16:23.009 "nvme_admin": false, 00:16:23.009 "nvme_io": false, 00:16:23.009 "nvme_io_md": false, 00:16:23.009 "write_zeroes": true, 00:16:23.009 "zcopy": true, 00:16:23.009 "get_zone_info": false, 00:16:23.009 "zone_management": false, 00:16:23.009 "zone_append": false, 00:16:23.009 "compare": false, 00:16:23.009 "compare_and_write": false, 00:16:23.009 "abort": true, 00:16:23.009 "seek_hole": false, 00:16:23.009 "seek_data": false, 00:16:23.009 "copy": true, 00:16:23.009 "nvme_iov_md": false 00:16:23.009 }, 00:16:23.009 "memory_domains": [ 00:16:23.009 { 00:16:23.009 "dma_device_id": "system", 00:16:23.009 "dma_device_type": 1 00:16:23.009 }, 00:16:23.009 { 00:16:23.009 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:23.009 "dma_device_type": 2 00:16:23.009 } 00:16:23.009 ], 00:16:23.009 "driver_specific": {} 00:16:23.009 }' 00:16:23.009 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:16:23.009 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:16:23.009 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:16:23.009 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:16:23.009 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:16:23.009 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:16:23.009 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:16:23.009 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:16:23.009 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:16:23.009 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:16:23.268 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:16:23.268 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:16:23.268 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:16:23.268 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:16:23.268 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:16:23.527 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:16:23.527 "name": "BaseBdev2", 00:16:23.527 "aliases": [ 00:16:23.527 "91c19df1-59e3-4895-ba71-79ced3520be9" 00:16:23.527 ], 00:16:23.527 "product_name": "Malloc disk", 00:16:23.527 "block_size": 512, 00:16:23.527 "num_blocks": 65536, 00:16:23.527 "uuid": "91c19df1-59e3-4895-ba71-79ced3520be9", 00:16:23.527 "assigned_rate_limits": { 00:16:23.527 "rw_ios_per_sec": 0, 00:16:23.527 "rw_mbytes_per_sec": 0, 00:16:23.527 "r_mbytes_per_sec": 0, 00:16:23.527 "w_mbytes_per_sec": 0 00:16:23.527 }, 00:16:23.527 "claimed": true, 00:16:23.527 "claim_type": "exclusive_write", 00:16:23.527 "zoned": false, 00:16:23.527 "supported_io_types": { 00:16:23.527 "read": true, 00:16:23.527 "write": true, 00:16:23.527 "unmap": true, 00:16:23.527 "flush": true, 00:16:23.527 "reset": true, 00:16:23.527 "nvme_admin": false, 00:16:23.527 "nvme_io": false, 00:16:23.527 "nvme_io_md": false, 00:16:23.527 "write_zeroes": true, 00:16:23.527 "zcopy": true, 00:16:23.527 "get_zone_info": false, 00:16:23.527 "zone_management": false, 00:16:23.527 "zone_append": false, 00:16:23.527 "compare": false, 00:16:23.527 "compare_and_write": false, 00:16:23.527 "abort": true, 00:16:23.527 "seek_hole": false, 00:16:23.527 "seek_data": false, 00:16:23.527 "copy": true, 00:16:23.527 "nvme_iov_md": false 00:16:23.527 }, 00:16:23.527 "memory_domains": [ 00:16:23.527 { 00:16:23.527 "dma_device_id": "system", 00:16:23.527 "dma_device_type": 1 00:16:23.527 }, 00:16:23.527 { 00:16:23.527 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:23.527 "dma_device_type": 2 00:16:23.527 } 00:16:23.527 ], 00:16:23.527 "driver_specific": {} 00:16:23.527 }' 00:16:23.527 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:16:23.527 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:16:23.527 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:16:23.527 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:16:23.527 11:27:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:16:23.527 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:16:23.527 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:16:23.527 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:16:23.527 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:16:23.527 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:16:23.787 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:16:23.787 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:16:23.787 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:16:23.787 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:16:23.787 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:16:24.045 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:16:24.045 "name": "BaseBdev3", 00:16:24.045 "aliases": [ 00:16:24.045 "657b8c71-a29c-4810-827f-bd3d6112282e" 00:16:24.045 ], 00:16:24.045 "product_name": "Malloc disk", 00:16:24.045 "block_size": 512, 00:16:24.045 "num_blocks": 65536, 00:16:24.045 "uuid": "657b8c71-a29c-4810-827f-bd3d6112282e", 00:16:24.045 "assigned_rate_limits": { 00:16:24.045 "rw_ios_per_sec": 0, 00:16:24.045 "rw_mbytes_per_sec": 0, 00:16:24.045 "r_mbytes_per_sec": 0, 00:16:24.045 "w_mbytes_per_sec": 0 00:16:24.045 }, 00:16:24.045 "claimed": true, 00:16:24.045 "claim_type": "exclusive_write", 00:16:24.045 "zoned": false, 00:16:24.045 "supported_io_types": { 00:16:24.045 "read": true, 00:16:24.045 "write": true, 00:16:24.045 "unmap": true, 00:16:24.045 "flush": true, 00:16:24.045 "reset": true, 00:16:24.045 "nvme_admin": false, 00:16:24.045 "nvme_io": false, 00:16:24.045 "nvme_io_md": false, 00:16:24.045 "write_zeroes": true, 00:16:24.045 "zcopy": true, 00:16:24.045 "get_zone_info": false, 00:16:24.045 "zone_management": false, 00:16:24.045 "zone_append": false, 00:16:24.045 "compare": false, 00:16:24.045 "compare_and_write": false, 00:16:24.045 "abort": true, 00:16:24.045 "seek_hole": false, 00:16:24.045 "seek_data": false, 00:16:24.045 "copy": true, 00:16:24.045 "nvme_iov_md": false 00:16:24.045 }, 00:16:24.045 "memory_domains": [ 00:16:24.045 { 00:16:24.045 "dma_device_id": "system", 00:16:24.045 "dma_device_type": 1 00:16:24.045 }, 00:16:24.045 { 00:16:24.046 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:24.046 "dma_device_type": 2 00:16:24.046 } 00:16:24.046 ], 00:16:24.046 "driver_specific": {} 00:16:24.046 }' 00:16:24.046 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:16:24.046 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:16:24.046 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:16:24.046 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:16:24.046 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:16:24.046 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:16:24.046 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:16:24.046 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:16:24.304 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:16:24.304 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:16:24.304 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:16:24.305 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:16:24.305 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:16:24.564 [2024-07-15 11:27:07.933831] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:24.564 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@275 -- # local expected_state 00:16:24.564 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # has_redundancy raid1 00:16:24.564 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:16:24.564 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@214 -- # return 0 00:16:24.564 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # expected_state=online 00:16:24.564 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:16:24.564 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:24.564 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:16:24.564 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:24.564 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:24.564 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:16:24.564 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:24.564 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:24.564 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:24.564 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:24.564 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:24.564 11:27:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:24.823 11:27:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:24.823 "name": "Existed_Raid", 00:16:24.823 "uuid": "78c823d2-837c-40e1-9407-d5d1a9819a18", 00:16:24.823 "strip_size_kb": 0, 00:16:24.823 "state": "online", 00:16:24.823 "raid_level": "raid1", 00:16:24.823 "superblock": false, 00:16:24.823 "num_base_bdevs": 3, 00:16:24.823 "num_base_bdevs_discovered": 2, 00:16:24.823 "num_base_bdevs_operational": 2, 00:16:24.823 "base_bdevs_list": [ 00:16:24.823 { 00:16:24.823 "name": null, 00:16:24.823 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:24.823 "is_configured": false, 00:16:24.823 "data_offset": 0, 00:16:24.823 "data_size": 65536 00:16:24.823 }, 00:16:24.823 { 00:16:24.823 "name": "BaseBdev2", 00:16:24.823 "uuid": "91c19df1-59e3-4895-ba71-79ced3520be9", 00:16:24.823 "is_configured": true, 00:16:24.823 "data_offset": 0, 00:16:24.823 "data_size": 65536 00:16:24.823 }, 00:16:24.823 { 00:16:24.823 "name": "BaseBdev3", 00:16:24.823 "uuid": "657b8c71-a29c-4810-827f-bd3d6112282e", 00:16:24.823 "is_configured": true, 00:16:24.823 "data_offset": 0, 00:16:24.823 "data_size": 65536 00:16:24.823 } 00:16:24.823 ] 00:16:24.823 }' 00:16:24.823 11:27:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:24.823 11:27:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:16:25.392 11:27:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:16:25.392 11:27:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:16:25.392 11:27:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:25.392 11:27:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:16:25.392 11:27:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:16:25.392 11:27:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:16:25.392 11:27:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:16:25.651 [2024-07-15 11:27:09.195102] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:16:25.651 11:27:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:16:25.651 11:27:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:16:25.652 11:27:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:25.652 11:27:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:16:25.911 11:27:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:16:25.911 11:27:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:16:25.911 11:27:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev3 00:16:26.169 [2024-07-15 11:27:09.628584] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:16:26.169 [2024-07-15 11:27:09.628671] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:26.169 [2024-07-15 11:27:09.641301] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:26.169 [2024-07-15 11:27:09.641337] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:26.169 [2024-07-15 11:27:09.641350] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x195f400 name Existed_Raid, state offline 00:16:26.169 11:27:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:16:26.169 11:27:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:16:26.169 11:27:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:26.169 11:27:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:16:26.428 11:27:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:16:26.428 11:27:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:16:26.428 11:27:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # '[' 3 -gt 2 ']' 00:16:26.428 11:27:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i = 1 )) 00:16:26.428 11:27:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:16:26.428 11:27:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:16:26.686 BaseBdev2 00:16:26.686 11:27:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev2 00:16:26.686 11:27:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:16:26.686 11:27:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:16:26.686 11:27:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:16:26.686 11:27:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:16:26.686 11:27:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:16:26.686 11:27:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:16:26.945 11:27:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:16:27.203 [ 00:16:27.203 { 00:16:27.203 "name": "BaseBdev2", 00:16:27.203 "aliases": [ 00:16:27.203 "72471cee-93e5-4daa-9f23-7ac80116a58e" 00:16:27.203 ], 00:16:27.203 "product_name": "Malloc disk", 00:16:27.203 "block_size": 512, 00:16:27.203 "num_blocks": 65536, 00:16:27.203 "uuid": "72471cee-93e5-4daa-9f23-7ac80116a58e", 00:16:27.203 "assigned_rate_limits": { 00:16:27.203 "rw_ios_per_sec": 0, 00:16:27.203 "rw_mbytes_per_sec": 0, 00:16:27.203 "r_mbytes_per_sec": 0, 00:16:27.203 "w_mbytes_per_sec": 0 00:16:27.203 }, 00:16:27.203 "claimed": false, 00:16:27.203 "zoned": false, 00:16:27.203 "supported_io_types": { 00:16:27.203 "read": true, 00:16:27.203 "write": true, 00:16:27.203 "unmap": true, 00:16:27.203 "flush": true, 00:16:27.203 "reset": true, 00:16:27.203 "nvme_admin": false, 00:16:27.203 "nvme_io": false, 00:16:27.203 "nvme_io_md": false, 00:16:27.203 "write_zeroes": true, 00:16:27.203 "zcopy": true, 00:16:27.203 "get_zone_info": false, 00:16:27.203 "zone_management": false, 00:16:27.203 "zone_append": false, 00:16:27.203 "compare": false, 00:16:27.203 "compare_and_write": false, 00:16:27.203 "abort": true, 00:16:27.203 "seek_hole": false, 00:16:27.203 "seek_data": false, 00:16:27.203 "copy": true, 00:16:27.203 "nvme_iov_md": false 00:16:27.203 }, 00:16:27.203 "memory_domains": [ 00:16:27.203 { 00:16:27.203 "dma_device_id": "system", 00:16:27.203 "dma_device_type": 1 00:16:27.203 }, 00:16:27.203 { 00:16:27.203 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:27.203 "dma_device_type": 2 00:16:27.203 } 00:16:27.203 ], 00:16:27.203 "driver_specific": {} 00:16:27.203 } 00:16:27.203 ] 00:16:27.203 11:27:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:16:27.203 11:27:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:16:27.203 11:27:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:16:27.203 11:27:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:16:27.463 BaseBdev3 00:16:27.463 11:27:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev3 00:16:27.463 11:27:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:16:27.463 11:27:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:16:27.463 11:27:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:16:27.463 11:27:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:16:27.463 11:27:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:16:27.463 11:27:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:16:27.723 11:27:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:16:27.982 [ 00:16:27.982 { 00:16:27.982 "name": "BaseBdev3", 00:16:27.982 "aliases": [ 00:16:27.982 "6e57ce22-2bfd-4293-b1bc-6b5f58d2641a" 00:16:27.982 ], 00:16:27.982 "product_name": "Malloc disk", 00:16:27.982 "block_size": 512, 00:16:27.982 "num_blocks": 65536, 00:16:27.982 "uuid": "6e57ce22-2bfd-4293-b1bc-6b5f58d2641a", 00:16:27.982 "assigned_rate_limits": { 00:16:27.982 "rw_ios_per_sec": 0, 00:16:27.982 "rw_mbytes_per_sec": 0, 00:16:27.982 "r_mbytes_per_sec": 0, 00:16:27.982 "w_mbytes_per_sec": 0 00:16:27.982 }, 00:16:27.982 "claimed": false, 00:16:27.982 "zoned": false, 00:16:27.982 "supported_io_types": { 00:16:27.982 "read": true, 00:16:27.982 "write": true, 00:16:27.982 "unmap": true, 00:16:27.982 "flush": true, 00:16:27.982 "reset": true, 00:16:27.982 "nvme_admin": false, 00:16:27.982 "nvme_io": false, 00:16:27.982 "nvme_io_md": false, 00:16:27.982 "write_zeroes": true, 00:16:27.982 "zcopy": true, 00:16:27.982 "get_zone_info": false, 00:16:27.982 "zone_management": false, 00:16:27.982 "zone_append": false, 00:16:27.982 "compare": false, 00:16:27.982 "compare_and_write": false, 00:16:27.982 "abort": true, 00:16:27.982 "seek_hole": false, 00:16:27.982 "seek_data": false, 00:16:27.982 "copy": true, 00:16:27.982 "nvme_iov_md": false 00:16:27.982 }, 00:16:27.982 "memory_domains": [ 00:16:27.982 { 00:16:27.982 "dma_device_id": "system", 00:16:27.982 "dma_device_type": 1 00:16:27.982 }, 00:16:27.982 { 00:16:27.982 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:27.982 "dma_device_type": 2 00:16:27.982 } 00:16:27.982 ], 00:16:27.982 "driver_specific": {} 00:16:27.982 } 00:16:27.982 ] 00:16:27.982 11:27:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:16:27.982 11:27:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:16:27.982 11:27:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:16:27.982 11:27:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@305 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:16:28.241 [2024-07-15 11:27:11.586874] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:28.241 [2024-07-15 11:27:11.586922] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:28.241 [2024-07-15 11:27:11.586953] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:28.241 [2024-07-15 11:27:11.588360] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:16:28.241 11:27:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:16:28.241 11:27:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:28.241 11:27:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:16:28.241 11:27:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:28.241 11:27:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:28.241 11:27:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:28.241 11:27:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:28.241 11:27:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:28.241 11:27:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:28.241 11:27:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:28.241 11:27:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:28.241 11:27:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:28.499 11:27:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:28.499 "name": "Existed_Raid", 00:16:28.499 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:28.499 "strip_size_kb": 0, 00:16:28.499 "state": "configuring", 00:16:28.499 "raid_level": "raid1", 00:16:28.499 "superblock": false, 00:16:28.499 "num_base_bdevs": 3, 00:16:28.499 "num_base_bdevs_discovered": 2, 00:16:28.499 "num_base_bdevs_operational": 3, 00:16:28.499 "base_bdevs_list": [ 00:16:28.499 { 00:16:28.499 "name": "BaseBdev1", 00:16:28.499 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:28.499 "is_configured": false, 00:16:28.499 "data_offset": 0, 00:16:28.499 "data_size": 0 00:16:28.499 }, 00:16:28.499 { 00:16:28.499 "name": "BaseBdev2", 00:16:28.499 "uuid": "72471cee-93e5-4daa-9f23-7ac80116a58e", 00:16:28.499 "is_configured": true, 00:16:28.499 "data_offset": 0, 00:16:28.499 "data_size": 65536 00:16:28.499 }, 00:16:28.499 { 00:16:28.499 "name": "BaseBdev3", 00:16:28.499 "uuid": "6e57ce22-2bfd-4293-b1bc-6b5f58d2641a", 00:16:28.499 "is_configured": true, 00:16:28.499 "data_offset": 0, 00:16:28.499 "data_size": 65536 00:16:28.499 } 00:16:28.499 ] 00:16:28.499 }' 00:16:28.499 11:27:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:28.499 11:27:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:16:29.069 11:27:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev2 00:16:29.069 [2024-07-15 11:27:12.541394] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:16:29.069 11:27:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@309 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:16:29.069 11:27:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:29.069 11:27:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:16:29.069 11:27:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:29.069 11:27:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:29.069 11:27:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:29.069 11:27:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:29.069 11:27:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:29.069 11:27:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:29.069 11:27:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:29.069 11:27:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:29.069 11:27:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:29.328 11:27:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:29.328 "name": "Existed_Raid", 00:16:29.328 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:29.328 "strip_size_kb": 0, 00:16:29.328 "state": "configuring", 00:16:29.328 "raid_level": "raid1", 00:16:29.328 "superblock": false, 00:16:29.328 "num_base_bdevs": 3, 00:16:29.328 "num_base_bdevs_discovered": 1, 00:16:29.328 "num_base_bdevs_operational": 3, 00:16:29.328 "base_bdevs_list": [ 00:16:29.328 { 00:16:29.328 "name": "BaseBdev1", 00:16:29.328 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:29.328 "is_configured": false, 00:16:29.328 "data_offset": 0, 00:16:29.328 "data_size": 0 00:16:29.328 }, 00:16:29.328 { 00:16:29.328 "name": null, 00:16:29.328 "uuid": "72471cee-93e5-4daa-9f23-7ac80116a58e", 00:16:29.328 "is_configured": false, 00:16:29.328 "data_offset": 0, 00:16:29.328 "data_size": 65536 00:16:29.328 }, 00:16:29.328 { 00:16:29.328 "name": "BaseBdev3", 00:16:29.328 "uuid": "6e57ce22-2bfd-4293-b1bc-6b5f58d2641a", 00:16:29.328 "is_configured": true, 00:16:29.328 "data_offset": 0, 00:16:29.328 "data_size": 65536 00:16:29.328 } 00:16:29.328 ] 00:16:29.328 }' 00:16:29.328 11:27:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:29.328 11:27:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:16:29.895 11:27:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:29.895 11:27:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:16:30.154 11:27:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # [[ false == \f\a\l\s\e ]] 00:16:30.154 11:27:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:16:30.413 [2024-07-15 11:27:13.889292] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:30.413 BaseBdev1 00:16:30.413 11:27:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@313 -- # waitforbdev BaseBdev1 00:16:30.413 11:27:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:16:30.413 11:27:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:16:30.413 11:27:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:16:30.413 11:27:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:16:30.413 11:27:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:16:30.413 11:27:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:16:30.671 11:27:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:16:30.931 [ 00:16:30.931 { 00:16:30.931 "name": "BaseBdev1", 00:16:30.931 "aliases": [ 00:16:30.931 "78596337-2549-4db2-820a-efa6407c6ee6" 00:16:30.931 ], 00:16:30.931 "product_name": "Malloc disk", 00:16:30.931 "block_size": 512, 00:16:30.931 "num_blocks": 65536, 00:16:30.931 "uuid": "78596337-2549-4db2-820a-efa6407c6ee6", 00:16:30.931 "assigned_rate_limits": { 00:16:30.931 "rw_ios_per_sec": 0, 00:16:30.931 "rw_mbytes_per_sec": 0, 00:16:30.931 "r_mbytes_per_sec": 0, 00:16:30.931 "w_mbytes_per_sec": 0 00:16:30.931 }, 00:16:30.931 "claimed": true, 00:16:30.931 "claim_type": "exclusive_write", 00:16:30.931 "zoned": false, 00:16:30.931 "supported_io_types": { 00:16:30.931 "read": true, 00:16:30.931 "write": true, 00:16:30.931 "unmap": true, 00:16:30.931 "flush": true, 00:16:30.931 "reset": true, 00:16:30.931 "nvme_admin": false, 00:16:30.931 "nvme_io": false, 00:16:30.931 "nvme_io_md": false, 00:16:30.931 "write_zeroes": true, 00:16:30.931 "zcopy": true, 00:16:30.931 "get_zone_info": false, 00:16:30.931 "zone_management": false, 00:16:30.931 "zone_append": false, 00:16:30.931 "compare": false, 00:16:30.931 "compare_and_write": false, 00:16:30.931 "abort": true, 00:16:30.931 "seek_hole": false, 00:16:30.931 "seek_data": false, 00:16:30.931 "copy": true, 00:16:30.931 "nvme_iov_md": false 00:16:30.931 }, 00:16:30.931 "memory_domains": [ 00:16:30.931 { 00:16:30.931 "dma_device_id": "system", 00:16:30.931 "dma_device_type": 1 00:16:30.931 }, 00:16:30.931 { 00:16:30.931 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:30.931 "dma_device_type": 2 00:16:30.931 } 00:16:30.931 ], 00:16:30.931 "driver_specific": {} 00:16:30.931 } 00:16:30.931 ] 00:16:30.931 11:27:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:16:30.931 11:27:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:16:30.931 11:27:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:30.931 11:27:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:16:30.931 11:27:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:30.931 11:27:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:30.931 11:27:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:30.931 11:27:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:30.931 11:27:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:30.931 11:27:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:30.931 11:27:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:30.931 11:27:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:30.931 11:27:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:31.190 11:27:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:31.190 "name": "Existed_Raid", 00:16:31.190 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:31.190 "strip_size_kb": 0, 00:16:31.190 "state": "configuring", 00:16:31.190 "raid_level": "raid1", 00:16:31.190 "superblock": false, 00:16:31.190 "num_base_bdevs": 3, 00:16:31.190 "num_base_bdevs_discovered": 2, 00:16:31.190 "num_base_bdevs_operational": 3, 00:16:31.190 "base_bdevs_list": [ 00:16:31.190 { 00:16:31.190 "name": "BaseBdev1", 00:16:31.190 "uuid": "78596337-2549-4db2-820a-efa6407c6ee6", 00:16:31.190 "is_configured": true, 00:16:31.190 "data_offset": 0, 00:16:31.190 "data_size": 65536 00:16:31.190 }, 00:16:31.190 { 00:16:31.190 "name": null, 00:16:31.190 "uuid": "72471cee-93e5-4daa-9f23-7ac80116a58e", 00:16:31.190 "is_configured": false, 00:16:31.190 "data_offset": 0, 00:16:31.190 "data_size": 65536 00:16:31.190 }, 00:16:31.190 { 00:16:31.190 "name": "BaseBdev3", 00:16:31.190 "uuid": "6e57ce22-2bfd-4293-b1bc-6b5f58d2641a", 00:16:31.190 "is_configured": true, 00:16:31.190 "data_offset": 0, 00:16:31.190 "data_size": 65536 00:16:31.190 } 00:16:31.190 ] 00:16:31.190 }' 00:16:31.190 11:27:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:31.190 11:27:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:16:31.758 11:27:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:31.758 11:27:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:16:32.018 11:27:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # [[ true == \t\r\u\e ]] 00:16:32.018 11:27:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@317 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev3 00:16:32.587 [2024-07-15 11:27:15.934757] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:16:32.587 11:27:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:16:32.588 11:27:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:32.588 11:27:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:16:32.588 11:27:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:32.588 11:27:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:32.588 11:27:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:32.588 11:27:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:32.588 11:27:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:32.588 11:27:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:32.588 11:27:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:32.588 11:27:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:32.588 11:27:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:32.847 11:27:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:32.847 "name": "Existed_Raid", 00:16:32.847 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:32.847 "strip_size_kb": 0, 00:16:32.847 "state": "configuring", 00:16:32.847 "raid_level": "raid1", 00:16:32.847 "superblock": false, 00:16:32.847 "num_base_bdevs": 3, 00:16:32.847 "num_base_bdevs_discovered": 1, 00:16:32.847 "num_base_bdevs_operational": 3, 00:16:32.847 "base_bdevs_list": [ 00:16:32.847 { 00:16:32.847 "name": "BaseBdev1", 00:16:32.847 "uuid": "78596337-2549-4db2-820a-efa6407c6ee6", 00:16:32.847 "is_configured": true, 00:16:32.847 "data_offset": 0, 00:16:32.847 "data_size": 65536 00:16:32.847 }, 00:16:32.847 { 00:16:32.847 "name": null, 00:16:32.847 "uuid": "72471cee-93e5-4daa-9f23-7ac80116a58e", 00:16:32.847 "is_configured": false, 00:16:32.847 "data_offset": 0, 00:16:32.847 "data_size": 65536 00:16:32.847 }, 00:16:32.847 { 00:16:32.847 "name": null, 00:16:32.847 "uuid": "6e57ce22-2bfd-4293-b1bc-6b5f58d2641a", 00:16:32.847 "is_configured": false, 00:16:32.847 "data_offset": 0, 00:16:32.847 "data_size": 65536 00:16:32.847 } 00:16:32.847 ] 00:16:32.847 }' 00:16:32.847 11:27:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:32.847 11:27:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:16:33.454 11:27:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:33.454 11:27:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:16:33.454 11:27:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # [[ false == \f\a\l\s\e ]] 00:16:33.454 11:27:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:16:33.713 [2024-07-15 11:27:17.226213] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:16:33.713 11:27:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@322 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:16:33.713 11:27:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:33.713 11:27:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:16:33.713 11:27:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:33.713 11:27:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:33.713 11:27:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:33.713 11:27:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:33.713 11:27:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:33.713 11:27:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:33.713 11:27:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:33.713 11:27:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:33.713 11:27:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:33.973 11:27:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:33.973 "name": "Existed_Raid", 00:16:33.973 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:33.973 "strip_size_kb": 0, 00:16:33.973 "state": "configuring", 00:16:33.973 "raid_level": "raid1", 00:16:33.973 "superblock": false, 00:16:33.973 "num_base_bdevs": 3, 00:16:33.973 "num_base_bdevs_discovered": 2, 00:16:33.973 "num_base_bdevs_operational": 3, 00:16:33.973 "base_bdevs_list": [ 00:16:33.973 { 00:16:33.973 "name": "BaseBdev1", 00:16:33.973 "uuid": "78596337-2549-4db2-820a-efa6407c6ee6", 00:16:33.973 "is_configured": true, 00:16:33.973 "data_offset": 0, 00:16:33.973 "data_size": 65536 00:16:33.973 }, 00:16:33.973 { 00:16:33.973 "name": null, 00:16:33.973 "uuid": "72471cee-93e5-4daa-9f23-7ac80116a58e", 00:16:33.973 "is_configured": false, 00:16:33.973 "data_offset": 0, 00:16:33.973 "data_size": 65536 00:16:33.973 }, 00:16:33.973 { 00:16:33.973 "name": "BaseBdev3", 00:16:33.973 "uuid": "6e57ce22-2bfd-4293-b1bc-6b5f58d2641a", 00:16:33.973 "is_configured": true, 00:16:33.973 "data_offset": 0, 00:16:33.973 "data_size": 65536 00:16:33.973 } 00:16:33.973 ] 00:16:33.973 }' 00:16:33.973 11:27:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:33.973 11:27:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:16:34.541 11:27:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:16:34.541 11:27:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:35.109 11:27:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # [[ true == \t\r\u\e ]] 00:16:35.109 11:27:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@325 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:16:35.368 [2024-07-15 11:27:18.730374] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:35.368 11:27:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:16:35.368 11:27:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:35.368 11:27:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:16:35.368 11:27:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:35.368 11:27:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:35.368 11:27:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:35.368 11:27:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:35.368 11:27:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:35.368 11:27:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:35.368 11:27:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:35.368 11:27:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:35.368 11:27:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:35.625 11:27:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:35.625 "name": "Existed_Raid", 00:16:35.625 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:35.625 "strip_size_kb": 0, 00:16:35.625 "state": "configuring", 00:16:35.625 "raid_level": "raid1", 00:16:35.625 "superblock": false, 00:16:35.625 "num_base_bdevs": 3, 00:16:35.625 "num_base_bdevs_discovered": 1, 00:16:35.625 "num_base_bdevs_operational": 3, 00:16:35.625 "base_bdevs_list": [ 00:16:35.625 { 00:16:35.625 "name": null, 00:16:35.625 "uuid": "78596337-2549-4db2-820a-efa6407c6ee6", 00:16:35.625 "is_configured": false, 00:16:35.625 "data_offset": 0, 00:16:35.625 "data_size": 65536 00:16:35.625 }, 00:16:35.625 { 00:16:35.625 "name": null, 00:16:35.625 "uuid": "72471cee-93e5-4daa-9f23-7ac80116a58e", 00:16:35.625 "is_configured": false, 00:16:35.625 "data_offset": 0, 00:16:35.625 "data_size": 65536 00:16:35.625 }, 00:16:35.625 { 00:16:35.626 "name": "BaseBdev3", 00:16:35.626 "uuid": "6e57ce22-2bfd-4293-b1bc-6b5f58d2641a", 00:16:35.626 "is_configured": true, 00:16:35.626 "data_offset": 0, 00:16:35.626 "data_size": 65536 00:16:35.626 } 00:16:35.626 ] 00:16:35.626 }' 00:16:35.626 11:27:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:35.626 11:27:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:16:36.192 11:27:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@327 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:36.192 11:27:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@327 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:16:36.192 11:27:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@327 -- # [[ false == \f\a\l\s\e ]] 00:16:36.192 11:27:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@329 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:16:36.451 [2024-07-15 11:27:19.921906] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:36.451 11:27:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@330 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:16:36.451 11:27:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:36.451 11:27:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:16:36.451 11:27:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:36.451 11:27:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:36.451 11:27:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:36.451 11:27:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:36.451 11:27:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:36.451 11:27:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:36.451 11:27:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:36.451 11:27:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:36.451 11:27:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:37.017 11:27:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:37.017 "name": "Existed_Raid", 00:16:37.017 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:37.017 "strip_size_kb": 0, 00:16:37.017 "state": "configuring", 00:16:37.017 "raid_level": "raid1", 00:16:37.017 "superblock": false, 00:16:37.017 "num_base_bdevs": 3, 00:16:37.017 "num_base_bdevs_discovered": 2, 00:16:37.017 "num_base_bdevs_operational": 3, 00:16:37.017 "base_bdevs_list": [ 00:16:37.017 { 00:16:37.017 "name": null, 00:16:37.017 "uuid": "78596337-2549-4db2-820a-efa6407c6ee6", 00:16:37.017 "is_configured": false, 00:16:37.017 "data_offset": 0, 00:16:37.017 "data_size": 65536 00:16:37.017 }, 00:16:37.017 { 00:16:37.017 "name": "BaseBdev2", 00:16:37.017 "uuid": "72471cee-93e5-4daa-9f23-7ac80116a58e", 00:16:37.017 "is_configured": true, 00:16:37.017 "data_offset": 0, 00:16:37.017 "data_size": 65536 00:16:37.017 }, 00:16:37.017 { 00:16:37.017 "name": "BaseBdev3", 00:16:37.017 "uuid": "6e57ce22-2bfd-4293-b1bc-6b5f58d2641a", 00:16:37.017 "is_configured": true, 00:16:37.017 "data_offset": 0, 00:16:37.017 "data_size": 65536 00:16:37.017 } 00:16:37.017 ] 00:16:37.017 }' 00:16:37.017 11:27:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:37.017 11:27:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:16:37.582 11:27:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@331 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:37.582 11:27:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@331 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:16:37.840 11:27:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@331 -- # [[ true == \t\r\u\e ]] 00:16:37.840 11:27:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:37.840 11:27:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@333 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:16:38.099 11:27:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b NewBaseBdev -u 78596337-2549-4db2-820a-efa6407c6ee6 00:16:38.099 [2024-07-15 11:27:21.686110] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:16:38.099 [2024-07-15 11:27:21.686152] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1962e40 00:16:38.099 [2024-07-15 11:27:21.686160] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:16:38.099 [2024-07-15 11:27:21.686352] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x195fe60 00:16:38.099 [2024-07-15 11:27:21.686476] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1962e40 00:16:38.099 [2024-07-15 11:27:21.686486] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x1962e40 00:16:38.099 [2024-07-15 11:27:21.686660] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:38.099 NewBaseBdev 00:16:38.358 11:27:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@334 -- # waitforbdev NewBaseBdev 00:16:38.358 11:27:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=NewBaseBdev 00:16:38.358 11:27:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:16:38.358 11:27:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:16:38.358 11:27:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:16:38.358 11:27:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:16:38.358 11:27:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:16:38.358 11:27:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev -t 2000 00:16:38.925 [ 00:16:38.925 { 00:16:38.925 "name": "NewBaseBdev", 00:16:38.925 "aliases": [ 00:16:38.925 "78596337-2549-4db2-820a-efa6407c6ee6" 00:16:38.925 ], 00:16:38.925 "product_name": "Malloc disk", 00:16:38.925 "block_size": 512, 00:16:38.925 "num_blocks": 65536, 00:16:38.925 "uuid": "78596337-2549-4db2-820a-efa6407c6ee6", 00:16:38.925 "assigned_rate_limits": { 00:16:38.925 "rw_ios_per_sec": 0, 00:16:38.925 "rw_mbytes_per_sec": 0, 00:16:38.925 "r_mbytes_per_sec": 0, 00:16:38.925 "w_mbytes_per_sec": 0 00:16:38.925 }, 00:16:38.925 "claimed": true, 00:16:38.925 "claim_type": "exclusive_write", 00:16:38.925 "zoned": false, 00:16:38.925 "supported_io_types": { 00:16:38.925 "read": true, 00:16:38.925 "write": true, 00:16:38.925 "unmap": true, 00:16:38.925 "flush": true, 00:16:38.925 "reset": true, 00:16:38.925 "nvme_admin": false, 00:16:38.925 "nvme_io": false, 00:16:38.925 "nvme_io_md": false, 00:16:38.925 "write_zeroes": true, 00:16:38.925 "zcopy": true, 00:16:38.925 "get_zone_info": false, 00:16:38.925 "zone_management": false, 00:16:38.925 "zone_append": false, 00:16:38.925 "compare": false, 00:16:38.925 "compare_and_write": false, 00:16:38.925 "abort": true, 00:16:38.925 "seek_hole": false, 00:16:38.925 "seek_data": false, 00:16:38.925 "copy": true, 00:16:38.925 "nvme_iov_md": false 00:16:38.925 }, 00:16:38.925 "memory_domains": [ 00:16:38.925 { 00:16:38.925 "dma_device_id": "system", 00:16:38.925 "dma_device_type": 1 00:16:38.925 }, 00:16:38.925 { 00:16:38.925 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:38.925 "dma_device_type": 2 00:16:38.925 } 00:16:38.925 ], 00:16:38.925 "driver_specific": {} 00:16:38.925 } 00:16:38.925 ] 00:16:38.925 11:27:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:16:38.925 11:27:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@335 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:16:38.925 11:27:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:38.926 11:27:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:16:38.926 11:27:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:38.926 11:27:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:38.926 11:27:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:38.926 11:27:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:38.926 11:27:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:38.926 11:27:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:38.926 11:27:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:38.926 11:27:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:38.926 11:27:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:39.185 11:27:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:39.185 "name": "Existed_Raid", 00:16:39.185 "uuid": "3030d910-4c79-4a98-b5cb-818e76ffc8d8", 00:16:39.185 "strip_size_kb": 0, 00:16:39.185 "state": "online", 00:16:39.185 "raid_level": "raid1", 00:16:39.185 "superblock": false, 00:16:39.185 "num_base_bdevs": 3, 00:16:39.185 "num_base_bdevs_discovered": 3, 00:16:39.185 "num_base_bdevs_operational": 3, 00:16:39.185 "base_bdevs_list": [ 00:16:39.185 { 00:16:39.185 "name": "NewBaseBdev", 00:16:39.185 "uuid": "78596337-2549-4db2-820a-efa6407c6ee6", 00:16:39.185 "is_configured": true, 00:16:39.185 "data_offset": 0, 00:16:39.185 "data_size": 65536 00:16:39.185 }, 00:16:39.185 { 00:16:39.185 "name": "BaseBdev2", 00:16:39.185 "uuid": "72471cee-93e5-4daa-9f23-7ac80116a58e", 00:16:39.185 "is_configured": true, 00:16:39.185 "data_offset": 0, 00:16:39.185 "data_size": 65536 00:16:39.185 }, 00:16:39.185 { 00:16:39.185 "name": "BaseBdev3", 00:16:39.185 "uuid": "6e57ce22-2bfd-4293-b1bc-6b5f58d2641a", 00:16:39.185 "is_configured": true, 00:16:39.185 "data_offset": 0, 00:16:39.185 "data_size": 65536 00:16:39.185 } 00:16:39.185 ] 00:16:39.185 }' 00:16:39.185 11:27:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:39.185 11:27:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:16:40.136 11:27:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@336 -- # verify_raid_bdev_properties Existed_Raid 00:16:40.136 11:27:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:16:40.136 11:27:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:16:40.136 11:27:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:16:40.136 11:27:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:16:40.136 11:27:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # local name 00:16:40.136 11:27:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:16:40.136 11:27:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:16:40.395 [2024-07-15 11:27:23.812053] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:40.395 11:27:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:16:40.395 "name": "Existed_Raid", 00:16:40.395 "aliases": [ 00:16:40.395 "3030d910-4c79-4a98-b5cb-818e76ffc8d8" 00:16:40.395 ], 00:16:40.395 "product_name": "Raid Volume", 00:16:40.395 "block_size": 512, 00:16:40.395 "num_blocks": 65536, 00:16:40.395 "uuid": "3030d910-4c79-4a98-b5cb-818e76ffc8d8", 00:16:40.395 "assigned_rate_limits": { 00:16:40.395 "rw_ios_per_sec": 0, 00:16:40.395 "rw_mbytes_per_sec": 0, 00:16:40.395 "r_mbytes_per_sec": 0, 00:16:40.395 "w_mbytes_per_sec": 0 00:16:40.395 }, 00:16:40.395 "claimed": false, 00:16:40.395 "zoned": false, 00:16:40.395 "supported_io_types": { 00:16:40.395 "read": true, 00:16:40.395 "write": true, 00:16:40.395 "unmap": false, 00:16:40.395 "flush": false, 00:16:40.395 "reset": true, 00:16:40.395 "nvme_admin": false, 00:16:40.395 "nvme_io": false, 00:16:40.395 "nvme_io_md": false, 00:16:40.395 "write_zeroes": true, 00:16:40.395 "zcopy": false, 00:16:40.395 "get_zone_info": false, 00:16:40.395 "zone_management": false, 00:16:40.395 "zone_append": false, 00:16:40.395 "compare": false, 00:16:40.395 "compare_and_write": false, 00:16:40.395 "abort": false, 00:16:40.395 "seek_hole": false, 00:16:40.395 "seek_data": false, 00:16:40.395 "copy": false, 00:16:40.395 "nvme_iov_md": false 00:16:40.395 }, 00:16:40.395 "memory_domains": [ 00:16:40.395 { 00:16:40.395 "dma_device_id": "system", 00:16:40.395 "dma_device_type": 1 00:16:40.395 }, 00:16:40.395 { 00:16:40.395 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:40.395 "dma_device_type": 2 00:16:40.395 }, 00:16:40.395 { 00:16:40.395 "dma_device_id": "system", 00:16:40.395 "dma_device_type": 1 00:16:40.395 }, 00:16:40.395 { 00:16:40.395 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:40.395 "dma_device_type": 2 00:16:40.395 }, 00:16:40.395 { 00:16:40.395 "dma_device_id": "system", 00:16:40.395 "dma_device_type": 1 00:16:40.395 }, 00:16:40.395 { 00:16:40.395 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:40.395 "dma_device_type": 2 00:16:40.395 } 00:16:40.395 ], 00:16:40.395 "driver_specific": { 00:16:40.395 "raid": { 00:16:40.395 "uuid": "3030d910-4c79-4a98-b5cb-818e76ffc8d8", 00:16:40.395 "strip_size_kb": 0, 00:16:40.395 "state": "online", 00:16:40.395 "raid_level": "raid1", 00:16:40.395 "superblock": false, 00:16:40.395 "num_base_bdevs": 3, 00:16:40.395 "num_base_bdevs_discovered": 3, 00:16:40.395 "num_base_bdevs_operational": 3, 00:16:40.395 "base_bdevs_list": [ 00:16:40.395 { 00:16:40.395 "name": "NewBaseBdev", 00:16:40.395 "uuid": "78596337-2549-4db2-820a-efa6407c6ee6", 00:16:40.395 "is_configured": true, 00:16:40.395 "data_offset": 0, 00:16:40.395 "data_size": 65536 00:16:40.395 }, 00:16:40.395 { 00:16:40.395 "name": "BaseBdev2", 00:16:40.395 "uuid": "72471cee-93e5-4daa-9f23-7ac80116a58e", 00:16:40.395 "is_configured": true, 00:16:40.395 "data_offset": 0, 00:16:40.395 "data_size": 65536 00:16:40.395 }, 00:16:40.395 { 00:16:40.395 "name": "BaseBdev3", 00:16:40.395 "uuid": "6e57ce22-2bfd-4293-b1bc-6b5f58d2641a", 00:16:40.395 "is_configured": true, 00:16:40.395 "data_offset": 0, 00:16:40.395 "data_size": 65536 00:16:40.395 } 00:16:40.395 ] 00:16:40.395 } 00:16:40.395 } 00:16:40.395 }' 00:16:40.395 11:27:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:40.395 11:27:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='NewBaseBdev 00:16:40.395 BaseBdev2 00:16:40.395 BaseBdev3' 00:16:40.395 11:27:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:16:40.395 11:27:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev 00:16:40.395 11:27:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:16:40.654 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:16:40.654 "name": "NewBaseBdev", 00:16:40.654 "aliases": [ 00:16:40.654 "78596337-2549-4db2-820a-efa6407c6ee6" 00:16:40.654 ], 00:16:40.654 "product_name": "Malloc disk", 00:16:40.654 "block_size": 512, 00:16:40.654 "num_blocks": 65536, 00:16:40.654 "uuid": "78596337-2549-4db2-820a-efa6407c6ee6", 00:16:40.654 "assigned_rate_limits": { 00:16:40.654 "rw_ios_per_sec": 0, 00:16:40.654 "rw_mbytes_per_sec": 0, 00:16:40.654 "r_mbytes_per_sec": 0, 00:16:40.654 "w_mbytes_per_sec": 0 00:16:40.654 }, 00:16:40.654 "claimed": true, 00:16:40.654 "claim_type": "exclusive_write", 00:16:40.654 "zoned": false, 00:16:40.654 "supported_io_types": { 00:16:40.654 "read": true, 00:16:40.654 "write": true, 00:16:40.654 "unmap": true, 00:16:40.654 "flush": true, 00:16:40.654 "reset": true, 00:16:40.654 "nvme_admin": false, 00:16:40.654 "nvme_io": false, 00:16:40.654 "nvme_io_md": false, 00:16:40.654 "write_zeroes": true, 00:16:40.654 "zcopy": true, 00:16:40.654 "get_zone_info": false, 00:16:40.654 "zone_management": false, 00:16:40.654 "zone_append": false, 00:16:40.654 "compare": false, 00:16:40.654 "compare_and_write": false, 00:16:40.654 "abort": true, 00:16:40.654 "seek_hole": false, 00:16:40.654 "seek_data": false, 00:16:40.654 "copy": true, 00:16:40.654 "nvme_iov_md": false 00:16:40.654 }, 00:16:40.654 "memory_domains": [ 00:16:40.654 { 00:16:40.654 "dma_device_id": "system", 00:16:40.654 "dma_device_type": 1 00:16:40.654 }, 00:16:40.654 { 00:16:40.654 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:40.654 "dma_device_type": 2 00:16:40.654 } 00:16:40.654 ], 00:16:40.654 "driver_specific": {} 00:16:40.654 }' 00:16:40.654 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:16:40.654 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:16:40.654 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:16:40.654 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:16:40.914 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:16:40.914 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:16:40.914 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:16:40.914 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:16:40.914 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:16:40.914 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:16:40.914 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:16:40.914 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:16:40.914 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:16:40.914 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:16:40.914 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:16:41.173 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:16:41.173 "name": "BaseBdev2", 00:16:41.173 "aliases": [ 00:16:41.173 "72471cee-93e5-4daa-9f23-7ac80116a58e" 00:16:41.173 ], 00:16:41.173 "product_name": "Malloc disk", 00:16:41.173 "block_size": 512, 00:16:41.173 "num_blocks": 65536, 00:16:41.174 "uuid": "72471cee-93e5-4daa-9f23-7ac80116a58e", 00:16:41.174 "assigned_rate_limits": { 00:16:41.174 "rw_ios_per_sec": 0, 00:16:41.174 "rw_mbytes_per_sec": 0, 00:16:41.174 "r_mbytes_per_sec": 0, 00:16:41.174 "w_mbytes_per_sec": 0 00:16:41.174 }, 00:16:41.174 "claimed": true, 00:16:41.174 "claim_type": "exclusive_write", 00:16:41.174 "zoned": false, 00:16:41.174 "supported_io_types": { 00:16:41.174 "read": true, 00:16:41.174 "write": true, 00:16:41.174 "unmap": true, 00:16:41.174 "flush": true, 00:16:41.174 "reset": true, 00:16:41.174 "nvme_admin": false, 00:16:41.174 "nvme_io": false, 00:16:41.174 "nvme_io_md": false, 00:16:41.174 "write_zeroes": true, 00:16:41.174 "zcopy": true, 00:16:41.174 "get_zone_info": false, 00:16:41.174 "zone_management": false, 00:16:41.174 "zone_append": false, 00:16:41.174 "compare": false, 00:16:41.174 "compare_and_write": false, 00:16:41.174 "abort": true, 00:16:41.174 "seek_hole": false, 00:16:41.174 "seek_data": false, 00:16:41.174 "copy": true, 00:16:41.174 "nvme_iov_md": false 00:16:41.174 }, 00:16:41.174 "memory_domains": [ 00:16:41.174 { 00:16:41.174 "dma_device_id": "system", 00:16:41.174 "dma_device_type": 1 00:16:41.174 }, 00:16:41.174 { 00:16:41.174 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:41.174 "dma_device_type": 2 00:16:41.174 } 00:16:41.174 ], 00:16:41.174 "driver_specific": {} 00:16:41.174 }' 00:16:41.174 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:16:41.434 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:16:41.434 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:16:41.434 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:16:41.434 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:16:41.434 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:16:41.434 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:16:41.434 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:16:41.434 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:16:41.434 11:27:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:16:41.434 11:27:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:16:41.693 11:27:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:16:41.693 11:27:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:16:41.693 11:27:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:16:41.693 11:27:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:16:41.952 11:27:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:16:41.952 "name": "BaseBdev3", 00:16:41.952 "aliases": [ 00:16:41.952 "6e57ce22-2bfd-4293-b1bc-6b5f58d2641a" 00:16:41.952 ], 00:16:41.952 "product_name": "Malloc disk", 00:16:41.952 "block_size": 512, 00:16:41.952 "num_blocks": 65536, 00:16:41.952 "uuid": "6e57ce22-2bfd-4293-b1bc-6b5f58d2641a", 00:16:41.952 "assigned_rate_limits": { 00:16:41.952 "rw_ios_per_sec": 0, 00:16:41.952 "rw_mbytes_per_sec": 0, 00:16:41.952 "r_mbytes_per_sec": 0, 00:16:41.952 "w_mbytes_per_sec": 0 00:16:41.952 }, 00:16:41.952 "claimed": true, 00:16:41.952 "claim_type": "exclusive_write", 00:16:41.952 "zoned": false, 00:16:41.952 "supported_io_types": { 00:16:41.952 "read": true, 00:16:41.952 "write": true, 00:16:41.952 "unmap": true, 00:16:41.952 "flush": true, 00:16:41.952 "reset": true, 00:16:41.952 "nvme_admin": false, 00:16:41.952 "nvme_io": false, 00:16:41.952 "nvme_io_md": false, 00:16:41.952 "write_zeroes": true, 00:16:41.952 "zcopy": true, 00:16:41.952 "get_zone_info": false, 00:16:41.952 "zone_management": false, 00:16:41.952 "zone_append": false, 00:16:41.952 "compare": false, 00:16:41.952 "compare_and_write": false, 00:16:41.952 "abort": true, 00:16:41.952 "seek_hole": false, 00:16:41.952 "seek_data": false, 00:16:41.952 "copy": true, 00:16:41.952 "nvme_iov_md": false 00:16:41.952 }, 00:16:41.952 "memory_domains": [ 00:16:41.952 { 00:16:41.952 "dma_device_id": "system", 00:16:41.952 "dma_device_type": 1 00:16:41.952 }, 00:16:41.952 { 00:16:41.952 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:41.952 "dma_device_type": 2 00:16:41.952 } 00:16:41.952 ], 00:16:41.952 "driver_specific": {} 00:16:41.952 }' 00:16:41.952 11:27:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:16:41.952 11:27:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:16:41.952 11:27:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:16:41.952 11:27:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:16:41.952 11:27:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:16:41.952 11:27:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:16:41.952 11:27:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:16:41.952 11:27:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:16:42.211 11:27:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:16:42.211 11:27:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:16:42.211 11:27:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:16:42.211 11:27:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:16:42.211 11:27:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@338 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:16:42.470 [2024-07-15 11:27:25.893297] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:42.470 [2024-07-15 11:27:25.893328] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:42.470 [2024-07-15 11:27:25.893387] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:42.470 [2024-07-15 11:27:25.893654] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:42.470 [2024-07-15 11:27:25.893668] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1962e40 name Existed_Raid, state offline 00:16:42.470 11:27:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@341 -- # killprocess 908324 00:16:42.470 11:27:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@948 -- # '[' -z 908324 ']' 00:16:42.470 11:27:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@952 -- # kill -0 908324 00:16:42.470 11:27:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@953 -- # uname 00:16:42.470 11:27:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:16:42.470 11:27:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 908324 00:16:42.470 11:27:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:16:42.470 11:27:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:16:42.470 11:27:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 908324' 00:16:42.470 killing process with pid 908324 00:16:42.470 11:27:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@967 -- # kill 908324 00:16:42.470 [2024-07-15 11:27:25.961118] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:42.470 11:27:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # wait 908324 00:16:42.470 [2024-07-15 11:27:25.991459] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@343 -- # return 0 00:16:42.729 00:16:42.729 real 0m28.715s 00:16:42.729 user 0m52.671s 00:16:42.729 sys 0m5.022s 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:16:42.729 ************************************ 00:16:42.729 END TEST raid_state_function_test 00:16:42.729 ************************************ 00:16:42.729 11:27:26 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:16:42.729 11:27:26 bdev_raid -- bdev/bdev_raid.sh@868 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 3 true 00:16:42.729 11:27:26 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:16:42.729 11:27:26 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:16:42.729 11:27:26 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:42.729 ************************************ 00:16:42.729 START TEST raid_state_function_test_sb 00:16:42.729 ************************************ 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1123 -- # raid_state_function_test raid1 3 true 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@220 -- # local raid_level=raid1 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=3 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # local superblock=true 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev3 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # local strip_size 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # '[' raid1 '!=' raid1 ']' 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@234 -- # strip_size=0 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # '[' true = true ']' 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@238 -- # superblock_create_arg=-s 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # raid_pid=913209 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 913209' 00:16:42.729 Process raid pid: 913209 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@246 -- # waitforlisten 913209 /var/tmp/spdk-raid.sock 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@829 -- # '[' -z 913209 ']' 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@834 -- # local max_retries=100 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:16:42.729 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # xtrace_disable 00:16:42.729 11:27:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:42.988 [2024-07-15 11:27:26.363432] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:16:42.988 [2024-07-15 11:27:26.363500] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:42.988 [2024-07-15 11:27:26.485993] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:43.247 [2024-07-15 11:27:26.583102] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:16:43.247 [2024-07-15 11:27:26.636805] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:43.247 [2024-07-15 11:27:26.636834] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:43.814 11:27:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:16:43.814 11:27:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@862 -- # return 0 00:16:43.814 11:27:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:16:44.073 [2024-07-15 11:27:27.532160] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:44.073 [2024-07-15 11:27:27.532209] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:44.073 [2024-07-15 11:27:27.532221] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:44.073 [2024-07-15 11:27:27.532232] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:44.073 [2024-07-15 11:27:27.532241] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:16:44.074 [2024-07-15 11:27:27.532253] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:16:44.074 11:27:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:16:44.074 11:27:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:44.074 11:27:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:16:44.074 11:27:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:44.074 11:27:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:44.074 11:27:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:44.074 11:27:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:44.074 11:27:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:44.074 11:27:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:44.074 11:27:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:44.074 11:27:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:44.074 11:27:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:44.332 11:27:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:44.332 "name": "Existed_Raid", 00:16:44.332 "uuid": "750723a1-9d02-46c2-b105-6126eb0a3c09", 00:16:44.332 "strip_size_kb": 0, 00:16:44.332 "state": "configuring", 00:16:44.332 "raid_level": "raid1", 00:16:44.332 "superblock": true, 00:16:44.332 "num_base_bdevs": 3, 00:16:44.332 "num_base_bdevs_discovered": 0, 00:16:44.333 "num_base_bdevs_operational": 3, 00:16:44.333 "base_bdevs_list": [ 00:16:44.333 { 00:16:44.333 "name": "BaseBdev1", 00:16:44.333 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:44.333 "is_configured": false, 00:16:44.333 "data_offset": 0, 00:16:44.333 "data_size": 0 00:16:44.333 }, 00:16:44.333 { 00:16:44.333 "name": "BaseBdev2", 00:16:44.333 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:44.333 "is_configured": false, 00:16:44.333 "data_offset": 0, 00:16:44.333 "data_size": 0 00:16:44.333 }, 00:16:44.333 { 00:16:44.333 "name": "BaseBdev3", 00:16:44.333 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:44.333 "is_configured": false, 00:16:44.333 "data_offset": 0, 00:16:44.333 "data_size": 0 00:16:44.333 } 00:16:44.333 ] 00:16:44.333 }' 00:16:44.333 11:27:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:44.333 11:27:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:44.900 11:27:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:16:45.467 [2024-07-15 11:27:28.867588] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:45.467 [2024-07-15 11:27:28.867622] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1b5ba80 name Existed_Raid, state configuring 00:16:45.467 11:27:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:16:45.726 [2024-07-15 11:27:29.124281] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:45.726 [2024-07-15 11:27:29.124312] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:45.726 [2024-07-15 11:27:29.124322] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:45.726 [2024-07-15 11:27:29.124342] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:45.726 [2024-07-15 11:27:29.124351] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:16:45.726 [2024-07-15 11:27:29.124369] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:16:45.726 11:27:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:16:46.294 [2024-07-15 11:27:29.632636] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:46.294 BaseBdev1 00:16:46.294 11:27:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:16:46.294 11:27:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:16:46.294 11:27:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:16:46.294 11:27:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:16:46.294 11:27:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:16:46.294 11:27:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:16:46.294 11:27:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:16:46.553 11:27:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:16:46.812 [ 00:16:46.812 { 00:16:46.812 "name": "BaseBdev1", 00:16:46.812 "aliases": [ 00:16:46.812 "2a422cb3-4bbd-471f-9ec8-56cf165742c0" 00:16:46.812 ], 00:16:46.812 "product_name": "Malloc disk", 00:16:46.812 "block_size": 512, 00:16:46.812 "num_blocks": 65536, 00:16:46.812 "uuid": "2a422cb3-4bbd-471f-9ec8-56cf165742c0", 00:16:46.812 "assigned_rate_limits": { 00:16:46.812 "rw_ios_per_sec": 0, 00:16:46.812 "rw_mbytes_per_sec": 0, 00:16:46.812 "r_mbytes_per_sec": 0, 00:16:46.812 "w_mbytes_per_sec": 0 00:16:46.812 }, 00:16:46.812 "claimed": true, 00:16:46.812 "claim_type": "exclusive_write", 00:16:46.812 "zoned": false, 00:16:46.812 "supported_io_types": { 00:16:46.812 "read": true, 00:16:46.812 "write": true, 00:16:46.812 "unmap": true, 00:16:46.812 "flush": true, 00:16:46.812 "reset": true, 00:16:46.812 "nvme_admin": false, 00:16:46.812 "nvme_io": false, 00:16:46.812 "nvme_io_md": false, 00:16:46.812 "write_zeroes": true, 00:16:46.812 "zcopy": true, 00:16:46.812 "get_zone_info": false, 00:16:46.812 "zone_management": false, 00:16:46.812 "zone_append": false, 00:16:46.812 "compare": false, 00:16:46.812 "compare_and_write": false, 00:16:46.812 "abort": true, 00:16:46.812 "seek_hole": false, 00:16:46.812 "seek_data": false, 00:16:46.812 "copy": true, 00:16:46.812 "nvme_iov_md": false 00:16:46.812 }, 00:16:46.812 "memory_domains": [ 00:16:46.812 { 00:16:46.812 "dma_device_id": "system", 00:16:46.812 "dma_device_type": 1 00:16:46.812 }, 00:16:46.812 { 00:16:46.812 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:46.812 "dma_device_type": 2 00:16:46.812 } 00:16:46.812 ], 00:16:46.812 "driver_specific": {} 00:16:46.812 } 00:16:46.812 ] 00:16:47.071 11:27:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:16:47.071 11:27:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:16:47.071 11:27:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:47.071 11:27:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:16:47.071 11:27:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:47.071 11:27:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:47.071 11:27:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:47.071 11:27:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:47.071 11:27:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:47.071 11:27:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:47.071 11:27:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:47.071 11:27:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:47.071 11:27:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:47.330 11:27:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:47.330 "name": "Existed_Raid", 00:16:47.330 "uuid": "f62a3e93-7e5e-45b0-9959-f03ba6497d96", 00:16:47.330 "strip_size_kb": 0, 00:16:47.330 "state": "configuring", 00:16:47.330 "raid_level": "raid1", 00:16:47.330 "superblock": true, 00:16:47.330 "num_base_bdevs": 3, 00:16:47.330 "num_base_bdevs_discovered": 1, 00:16:47.330 "num_base_bdevs_operational": 3, 00:16:47.330 "base_bdevs_list": [ 00:16:47.330 { 00:16:47.330 "name": "BaseBdev1", 00:16:47.330 "uuid": "2a422cb3-4bbd-471f-9ec8-56cf165742c0", 00:16:47.330 "is_configured": true, 00:16:47.330 "data_offset": 2048, 00:16:47.330 "data_size": 63488 00:16:47.330 }, 00:16:47.330 { 00:16:47.330 "name": "BaseBdev2", 00:16:47.330 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:47.330 "is_configured": false, 00:16:47.330 "data_offset": 0, 00:16:47.330 "data_size": 0 00:16:47.330 }, 00:16:47.330 { 00:16:47.330 "name": "BaseBdev3", 00:16:47.330 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:47.330 "is_configured": false, 00:16:47.330 "data_offset": 0, 00:16:47.330 "data_size": 0 00:16:47.330 } 00:16:47.330 ] 00:16:47.330 }' 00:16:47.330 11:27:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:47.330 11:27:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:47.991 11:27:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:16:47.991 [2024-07-15 11:27:31.425374] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:47.991 [2024-07-15 11:27:31.425412] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1b5b310 name Existed_Raid, state configuring 00:16:47.991 11:27:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:16:48.250 [2024-07-15 11:27:31.670060] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:48.250 [2024-07-15 11:27:31.671495] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:48.250 [2024-07-15 11:27:31.671528] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:48.250 [2024-07-15 11:27:31.671538] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:16:48.250 [2024-07-15 11:27:31.671549] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:16:48.250 11:27:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:16:48.250 11:27:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:16:48.250 11:27:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:16:48.250 11:27:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:48.250 11:27:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:16:48.250 11:27:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:48.250 11:27:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:48.250 11:27:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:48.250 11:27:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:48.250 11:27:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:48.250 11:27:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:48.250 11:27:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:48.250 11:27:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:48.250 11:27:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:48.510 11:27:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:48.510 "name": "Existed_Raid", 00:16:48.510 "uuid": "e4f7ba56-0138-4064-8488-860886fd716f", 00:16:48.510 "strip_size_kb": 0, 00:16:48.510 "state": "configuring", 00:16:48.510 "raid_level": "raid1", 00:16:48.510 "superblock": true, 00:16:48.510 "num_base_bdevs": 3, 00:16:48.510 "num_base_bdevs_discovered": 1, 00:16:48.510 "num_base_bdevs_operational": 3, 00:16:48.510 "base_bdevs_list": [ 00:16:48.510 { 00:16:48.510 "name": "BaseBdev1", 00:16:48.510 "uuid": "2a422cb3-4bbd-471f-9ec8-56cf165742c0", 00:16:48.510 "is_configured": true, 00:16:48.510 "data_offset": 2048, 00:16:48.510 "data_size": 63488 00:16:48.510 }, 00:16:48.510 { 00:16:48.510 "name": "BaseBdev2", 00:16:48.510 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:48.510 "is_configured": false, 00:16:48.510 "data_offset": 0, 00:16:48.510 "data_size": 0 00:16:48.510 }, 00:16:48.510 { 00:16:48.510 "name": "BaseBdev3", 00:16:48.510 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:48.510 "is_configured": false, 00:16:48.510 "data_offset": 0, 00:16:48.510 "data_size": 0 00:16:48.510 } 00:16:48.510 ] 00:16:48.510 }' 00:16:48.510 11:27:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:48.510 11:27:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:49.079 11:27:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:16:49.339 [2024-07-15 11:27:32.688144] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:49.339 BaseBdev2 00:16:49.339 11:27:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:16:49.339 11:27:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:16:49.339 11:27:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:16:49.339 11:27:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:16:49.339 11:27:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:16:49.339 11:27:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:16:49.339 11:27:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:16:49.598 11:27:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:16:49.598 [ 00:16:49.598 { 00:16:49.598 "name": "BaseBdev2", 00:16:49.598 "aliases": [ 00:16:49.598 "3d9b03b4-e613-4597-8049-d9ca4fc39ca5" 00:16:49.598 ], 00:16:49.598 "product_name": "Malloc disk", 00:16:49.598 "block_size": 512, 00:16:49.598 "num_blocks": 65536, 00:16:49.598 "uuid": "3d9b03b4-e613-4597-8049-d9ca4fc39ca5", 00:16:49.598 "assigned_rate_limits": { 00:16:49.598 "rw_ios_per_sec": 0, 00:16:49.598 "rw_mbytes_per_sec": 0, 00:16:49.598 "r_mbytes_per_sec": 0, 00:16:49.598 "w_mbytes_per_sec": 0 00:16:49.598 }, 00:16:49.598 "claimed": true, 00:16:49.598 "claim_type": "exclusive_write", 00:16:49.598 "zoned": false, 00:16:49.598 "supported_io_types": { 00:16:49.598 "read": true, 00:16:49.598 "write": true, 00:16:49.598 "unmap": true, 00:16:49.598 "flush": true, 00:16:49.598 "reset": true, 00:16:49.598 "nvme_admin": false, 00:16:49.598 "nvme_io": false, 00:16:49.598 "nvme_io_md": false, 00:16:49.598 "write_zeroes": true, 00:16:49.598 "zcopy": true, 00:16:49.598 "get_zone_info": false, 00:16:49.598 "zone_management": false, 00:16:49.598 "zone_append": false, 00:16:49.598 "compare": false, 00:16:49.598 "compare_and_write": false, 00:16:49.598 "abort": true, 00:16:49.598 "seek_hole": false, 00:16:49.598 "seek_data": false, 00:16:49.598 "copy": true, 00:16:49.598 "nvme_iov_md": false 00:16:49.598 }, 00:16:49.598 "memory_domains": [ 00:16:49.598 { 00:16:49.598 "dma_device_id": "system", 00:16:49.598 "dma_device_type": 1 00:16:49.598 }, 00:16:49.598 { 00:16:49.598 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:49.598 "dma_device_type": 2 00:16:49.598 } 00:16:49.598 ], 00:16:49.598 "driver_specific": {} 00:16:49.598 } 00:16:49.598 ] 00:16:49.879 11:27:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:16:49.879 11:27:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:16:49.879 11:27:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:16:49.879 11:27:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:16:49.879 11:27:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:49.879 11:27:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:16:49.879 11:27:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:49.879 11:27:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:49.879 11:27:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:49.879 11:27:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:49.879 11:27:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:49.879 11:27:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:49.879 11:27:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:49.879 11:27:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:49.879 11:27:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:49.879 11:27:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:49.879 "name": "Existed_Raid", 00:16:49.879 "uuid": "e4f7ba56-0138-4064-8488-860886fd716f", 00:16:49.879 "strip_size_kb": 0, 00:16:49.879 "state": "configuring", 00:16:49.879 "raid_level": "raid1", 00:16:49.879 "superblock": true, 00:16:49.879 "num_base_bdevs": 3, 00:16:49.879 "num_base_bdevs_discovered": 2, 00:16:49.879 "num_base_bdevs_operational": 3, 00:16:49.879 "base_bdevs_list": [ 00:16:49.879 { 00:16:49.879 "name": "BaseBdev1", 00:16:49.879 "uuid": "2a422cb3-4bbd-471f-9ec8-56cf165742c0", 00:16:49.879 "is_configured": true, 00:16:49.879 "data_offset": 2048, 00:16:49.879 "data_size": 63488 00:16:49.879 }, 00:16:49.879 { 00:16:49.879 "name": "BaseBdev2", 00:16:49.879 "uuid": "3d9b03b4-e613-4597-8049-d9ca4fc39ca5", 00:16:49.879 "is_configured": true, 00:16:49.879 "data_offset": 2048, 00:16:49.879 "data_size": 63488 00:16:49.879 }, 00:16:49.879 { 00:16:49.879 "name": "BaseBdev3", 00:16:49.879 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:49.879 "is_configured": false, 00:16:49.879 "data_offset": 0, 00:16:49.879 "data_size": 0 00:16:49.879 } 00:16:49.879 ] 00:16:49.879 }' 00:16:49.879 11:27:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:49.879 11:27:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:50.816 11:27:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:16:50.816 [2024-07-15 11:27:34.275723] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:16:50.816 [2024-07-15 11:27:34.275879] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1b5c400 00:16:50.816 [2024-07-15 11:27:34.275893] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:16:50.816 [2024-07-15 11:27:34.276071] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1b5bef0 00:16:50.816 [2024-07-15 11:27:34.276188] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1b5c400 00:16:50.816 [2024-07-15 11:27:34.276198] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x1b5c400 00:16:50.816 [2024-07-15 11:27:34.276287] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:50.816 BaseBdev3 00:16:50.816 11:27:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev3 00:16:50.816 11:27:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:16:50.816 11:27:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:16:50.816 11:27:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:16:50.816 11:27:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:16:50.816 11:27:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:16:50.816 11:27:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:16:51.073 11:27:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:16:51.330 [ 00:16:51.330 { 00:16:51.330 "name": "BaseBdev3", 00:16:51.330 "aliases": [ 00:16:51.330 "12b08651-c04b-4c82-ba89-399286cade07" 00:16:51.330 ], 00:16:51.330 "product_name": "Malloc disk", 00:16:51.330 "block_size": 512, 00:16:51.330 "num_blocks": 65536, 00:16:51.330 "uuid": "12b08651-c04b-4c82-ba89-399286cade07", 00:16:51.330 "assigned_rate_limits": { 00:16:51.330 "rw_ios_per_sec": 0, 00:16:51.330 "rw_mbytes_per_sec": 0, 00:16:51.330 "r_mbytes_per_sec": 0, 00:16:51.330 "w_mbytes_per_sec": 0 00:16:51.330 }, 00:16:51.331 "claimed": true, 00:16:51.331 "claim_type": "exclusive_write", 00:16:51.331 "zoned": false, 00:16:51.331 "supported_io_types": { 00:16:51.331 "read": true, 00:16:51.331 "write": true, 00:16:51.331 "unmap": true, 00:16:51.331 "flush": true, 00:16:51.331 "reset": true, 00:16:51.331 "nvme_admin": false, 00:16:51.331 "nvme_io": false, 00:16:51.331 "nvme_io_md": false, 00:16:51.331 "write_zeroes": true, 00:16:51.331 "zcopy": true, 00:16:51.331 "get_zone_info": false, 00:16:51.331 "zone_management": false, 00:16:51.331 "zone_append": false, 00:16:51.331 "compare": false, 00:16:51.331 "compare_and_write": false, 00:16:51.331 "abort": true, 00:16:51.331 "seek_hole": false, 00:16:51.331 "seek_data": false, 00:16:51.331 "copy": true, 00:16:51.331 "nvme_iov_md": false 00:16:51.331 }, 00:16:51.331 "memory_domains": [ 00:16:51.331 { 00:16:51.331 "dma_device_id": "system", 00:16:51.331 "dma_device_type": 1 00:16:51.331 }, 00:16:51.331 { 00:16:51.331 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:51.331 "dma_device_type": 2 00:16:51.331 } 00:16:51.331 ], 00:16:51.331 "driver_specific": {} 00:16:51.331 } 00:16:51.331 ] 00:16:51.331 11:27:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:16:51.331 11:27:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:16:51.331 11:27:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:16:51.331 11:27:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:16:51.331 11:27:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:51.331 11:27:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:16:51.331 11:27:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:51.331 11:27:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:51.331 11:27:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:51.331 11:27:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:51.331 11:27:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:51.331 11:27:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:51.331 11:27:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:51.331 11:27:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:51.331 11:27:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:51.589 11:27:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:51.589 "name": "Existed_Raid", 00:16:51.589 "uuid": "e4f7ba56-0138-4064-8488-860886fd716f", 00:16:51.589 "strip_size_kb": 0, 00:16:51.589 "state": "online", 00:16:51.589 "raid_level": "raid1", 00:16:51.589 "superblock": true, 00:16:51.589 "num_base_bdevs": 3, 00:16:51.589 "num_base_bdevs_discovered": 3, 00:16:51.589 "num_base_bdevs_operational": 3, 00:16:51.589 "base_bdevs_list": [ 00:16:51.589 { 00:16:51.589 "name": "BaseBdev1", 00:16:51.589 "uuid": "2a422cb3-4bbd-471f-9ec8-56cf165742c0", 00:16:51.589 "is_configured": true, 00:16:51.589 "data_offset": 2048, 00:16:51.589 "data_size": 63488 00:16:51.589 }, 00:16:51.589 { 00:16:51.589 "name": "BaseBdev2", 00:16:51.589 "uuid": "3d9b03b4-e613-4597-8049-d9ca4fc39ca5", 00:16:51.589 "is_configured": true, 00:16:51.589 "data_offset": 2048, 00:16:51.589 "data_size": 63488 00:16:51.589 }, 00:16:51.589 { 00:16:51.589 "name": "BaseBdev3", 00:16:51.589 "uuid": "12b08651-c04b-4c82-ba89-399286cade07", 00:16:51.589 "is_configured": true, 00:16:51.589 "data_offset": 2048, 00:16:51.589 "data_size": 63488 00:16:51.589 } 00:16:51.589 ] 00:16:51.589 }' 00:16:51.589 11:27:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:51.589 11:27:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:52.155 11:27:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:16:52.155 11:27:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:16:52.155 11:27:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:16:52.155 11:27:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:16:52.155 11:27:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:16:52.155 11:27:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # local name 00:16:52.155 11:27:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:16:52.155 11:27:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:16:52.414 [2024-07-15 11:27:35.792256] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:52.414 11:27:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:16:52.414 "name": "Existed_Raid", 00:16:52.414 "aliases": [ 00:16:52.414 "e4f7ba56-0138-4064-8488-860886fd716f" 00:16:52.414 ], 00:16:52.414 "product_name": "Raid Volume", 00:16:52.414 "block_size": 512, 00:16:52.414 "num_blocks": 63488, 00:16:52.414 "uuid": "e4f7ba56-0138-4064-8488-860886fd716f", 00:16:52.414 "assigned_rate_limits": { 00:16:52.414 "rw_ios_per_sec": 0, 00:16:52.414 "rw_mbytes_per_sec": 0, 00:16:52.414 "r_mbytes_per_sec": 0, 00:16:52.414 "w_mbytes_per_sec": 0 00:16:52.414 }, 00:16:52.414 "claimed": false, 00:16:52.414 "zoned": false, 00:16:52.414 "supported_io_types": { 00:16:52.414 "read": true, 00:16:52.414 "write": true, 00:16:52.414 "unmap": false, 00:16:52.414 "flush": false, 00:16:52.414 "reset": true, 00:16:52.414 "nvme_admin": false, 00:16:52.414 "nvme_io": false, 00:16:52.414 "nvme_io_md": false, 00:16:52.414 "write_zeroes": true, 00:16:52.414 "zcopy": false, 00:16:52.414 "get_zone_info": false, 00:16:52.414 "zone_management": false, 00:16:52.414 "zone_append": false, 00:16:52.414 "compare": false, 00:16:52.414 "compare_and_write": false, 00:16:52.414 "abort": false, 00:16:52.414 "seek_hole": false, 00:16:52.414 "seek_data": false, 00:16:52.414 "copy": false, 00:16:52.414 "nvme_iov_md": false 00:16:52.414 }, 00:16:52.414 "memory_domains": [ 00:16:52.414 { 00:16:52.414 "dma_device_id": "system", 00:16:52.414 "dma_device_type": 1 00:16:52.414 }, 00:16:52.414 { 00:16:52.414 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:52.414 "dma_device_type": 2 00:16:52.414 }, 00:16:52.414 { 00:16:52.414 "dma_device_id": "system", 00:16:52.414 "dma_device_type": 1 00:16:52.414 }, 00:16:52.414 { 00:16:52.414 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:52.414 "dma_device_type": 2 00:16:52.414 }, 00:16:52.414 { 00:16:52.414 "dma_device_id": "system", 00:16:52.414 "dma_device_type": 1 00:16:52.414 }, 00:16:52.414 { 00:16:52.414 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:52.414 "dma_device_type": 2 00:16:52.414 } 00:16:52.414 ], 00:16:52.414 "driver_specific": { 00:16:52.414 "raid": { 00:16:52.414 "uuid": "e4f7ba56-0138-4064-8488-860886fd716f", 00:16:52.414 "strip_size_kb": 0, 00:16:52.414 "state": "online", 00:16:52.414 "raid_level": "raid1", 00:16:52.414 "superblock": true, 00:16:52.414 "num_base_bdevs": 3, 00:16:52.414 "num_base_bdevs_discovered": 3, 00:16:52.414 "num_base_bdevs_operational": 3, 00:16:52.414 "base_bdevs_list": [ 00:16:52.414 { 00:16:52.414 "name": "BaseBdev1", 00:16:52.414 "uuid": "2a422cb3-4bbd-471f-9ec8-56cf165742c0", 00:16:52.414 "is_configured": true, 00:16:52.414 "data_offset": 2048, 00:16:52.414 "data_size": 63488 00:16:52.414 }, 00:16:52.414 { 00:16:52.414 "name": "BaseBdev2", 00:16:52.414 "uuid": "3d9b03b4-e613-4597-8049-d9ca4fc39ca5", 00:16:52.414 "is_configured": true, 00:16:52.414 "data_offset": 2048, 00:16:52.414 "data_size": 63488 00:16:52.414 }, 00:16:52.414 { 00:16:52.414 "name": "BaseBdev3", 00:16:52.414 "uuid": "12b08651-c04b-4c82-ba89-399286cade07", 00:16:52.414 "is_configured": true, 00:16:52.414 "data_offset": 2048, 00:16:52.414 "data_size": 63488 00:16:52.414 } 00:16:52.414 ] 00:16:52.414 } 00:16:52.414 } 00:16:52.414 }' 00:16:52.414 11:27:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:52.415 11:27:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:16:52.415 BaseBdev2 00:16:52.415 BaseBdev3' 00:16:52.415 11:27:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:16:52.415 11:27:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:16:52.415 11:27:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:16:52.673 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:16:52.673 "name": "BaseBdev1", 00:16:52.673 "aliases": [ 00:16:52.673 "2a422cb3-4bbd-471f-9ec8-56cf165742c0" 00:16:52.673 ], 00:16:52.673 "product_name": "Malloc disk", 00:16:52.673 "block_size": 512, 00:16:52.673 "num_blocks": 65536, 00:16:52.673 "uuid": "2a422cb3-4bbd-471f-9ec8-56cf165742c0", 00:16:52.673 "assigned_rate_limits": { 00:16:52.673 "rw_ios_per_sec": 0, 00:16:52.673 "rw_mbytes_per_sec": 0, 00:16:52.673 "r_mbytes_per_sec": 0, 00:16:52.673 "w_mbytes_per_sec": 0 00:16:52.673 }, 00:16:52.673 "claimed": true, 00:16:52.673 "claim_type": "exclusive_write", 00:16:52.673 "zoned": false, 00:16:52.673 "supported_io_types": { 00:16:52.673 "read": true, 00:16:52.673 "write": true, 00:16:52.673 "unmap": true, 00:16:52.673 "flush": true, 00:16:52.673 "reset": true, 00:16:52.673 "nvme_admin": false, 00:16:52.673 "nvme_io": false, 00:16:52.673 "nvme_io_md": false, 00:16:52.673 "write_zeroes": true, 00:16:52.673 "zcopy": true, 00:16:52.673 "get_zone_info": false, 00:16:52.673 "zone_management": false, 00:16:52.673 "zone_append": false, 00:16:52.673 "compare": false, 00:16:52.673 "compare_and_write": false, 00:16:52.673 "abort": true, 00:16:52.673 "seek_hole": false, 00:16:52.673 "seek_data": false, 00:16:52.673 "copy": true, 00:16:52.673 "nvme_iov_md": false 00:16:52.673 }, 00:16:52.673 "memory_domains": [ 00:16:52.673 { 00:16:52.673 "dma_device_id": "system", 00:16:52.673 "dma_device_type": 1 00:16:52.673 }, 00:16:52.673 { 00:16:52.673 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:52.673 "dma_device_type": 2 00:16:52.673 } 00:16:52.673 ], 00:16:52.673 "driver_specific": {} 00:16:52.673 }' 00:16:52.673 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:16:52.673 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:16:52.673 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:16:52.673 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:16:52.673 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:16:52.673 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:16:52.673 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:16:52.932 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:16:52.932 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:16:52.932 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:16:52.932 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:16:52.932 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:16:52.932 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:16:52.932 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:16:52.932 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:16:53.190 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:16:53.190 "name": "BaseBdev2", 00:16:53.190 "aliases": [ 00:16:53.190 "3d9b03b4-e613-4597-8049-d9ca4fc39ca5" 00:16:53.190 ], 00:16:53.190 "product_name": "Malloc disk", 00:16:53.190 "block_size": 512, 00:16:53.190 "num_blocks": 65536, 00:16:53.190 "uuid": "3d9b03b4-e613-4597-8049-d9ca4fc39ca5", 00:16:53.190 "assigned_rate_limits": { 00:16:53.190 "rw_ios_per_sec": 0, 00:16:53.190 "rw_mbytes_per_sec": 0, 00:16:53.190 "r_mbytes_per_sec": 0, 00:16:53.190 "w_mbytes_per_sec": 0 00:16:53.190 }, 00:16:53.190 "claimed": true, 00:16:53.190 "claim_type": "exclusive_write", 00:16:53.190 "zoned": false, 00:16:53.190 "supported_io_types": { 00:16:53.190 "read": true, 00:16:53.190 "write": true, 00:16:53.190 "unmap": true, 00:16:53.190 "flush": true, 00:16:53.190 "reset": true, 00:16:53.190 "nvme_admin": false, 00:16:53.190 "nvme_io": false, 00:16:53.190 "nvme_io_md": false, 00:16:53.190 "write_zeroes": true, 00:16:53.190 "zcopy": true, 00:16:53.190 "get_zone_info": false, 00:16:53.190 "zone_management": false, 00:16:53.190 "zone_append": false, 00:16:53.190 "compare": false, 00:16:53.190 "compare_and_write": false, 00:16:53.190 "abort": true, 00:16:53.190 "seek_hole": false, 00:16:53.190 "seek_data": false, 00:16:53.190 "copy": true, 00:16:53.190 "nvme_iov_md": false 00:16:53.190 }, 00:16:53.190 "memory_domains": [ 00:16:53.190 { 00:16:53.190 "dma_device_id": "system", 00:16:53.190 "dma_device_type": 1 00:16:53.190 }, 00:16:53.190 { 00:16:53.190 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:53.190 "dma_device_type": 2 00:16:53.190 } 00:16:53.190 ], 00:16:53.190 "driver_specific": {} 00:16:53.190 }' 00:16:53.190 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:16:53.190 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:16:53.190 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:16:53.190 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:16:53.448 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:16:53.448 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:16:53.448 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:16:53.448 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:16:53.448 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:16:53.448 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:16:53.448 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:16:53.448 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:16:53.448 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:16:53.448 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:16:53.448 11:27:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:16:53.705 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:16:53.705 "name": "BaseBdev3", 00:16:53.705 "aliases": [ 00:16:53.705 "12b08651-c04b-4c82-ba89-399286cade07" 00:16:53.705 ], 00:16:53.705 "product_name": "Malloc disk", 00:16:53.705 "block_size": 512, 00:16:53.705 "num_blocks": 65536, 00:16:53.705 "uuid": "12b08651-c04b-4c82-ba89-399286cade07", 00:16:53.705 "assigned_rate_limits": { 00:16:53.705 "rw_ios_per_sec": 0, 00:16:53.705 "rw_mbytes_per_sec": 0, 00:16:53.705 "r_mbytes_per_sec": 0, 00:16:53.705 "w_mbytes_per_sec": 0 00:16:53.705 }, 00:16:53.705 "claimed": true, 00:16:53.705 "claim_type": "exclusive_write", 00:16:53.705 "zoned": false, 00:16:53.705 "supported_io_types": { 00:16:53.705 "read": true, 00:16:53.705 "write": true, 00:16:53.705 "unmap": true, 00:16:53.705 "flush": true, 00:16:53.705 "reset": true, 00:16:53.705 "nvme_admin": false, 00:16:53.705 "nvme_io": false, 00:16:53.705 "nvme_io_md": false, 00:16:53.705 "write_zeroes": true, 00:16:53.705 "zcopy": true, 00:16:53.705 "get_zone_info": false, 00:16:53.705 "zone_management": false, 00:16:53.705 "zone_append": false, 00:16:53.705 "compare": false, 00:16:53.705 "compare_and_write": false, 00:16:53.705 "abort": true, 00:16:53.705 "seek_hole": false, 00:16:53.705 "seek_data": false, 00:16:53.705 "copy": true, 00:16:53.705 "nvme_iov_md": false 00:16:53.705 }, 00:16:53.705 "memory_domains": [ 00:16:53.705 { 00:16:53.705 "dma_device_id": "system", 00:16:53.705 "dma_device_type": 1 00:16:53.705 }, 00:16:53.705 { 00:16:53.705 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:53.705 "dma_device_type": 2 00:16:53.705 } 00:16:53.705 ], 00:16:53.705 "driver_specific": {} 00:16:53.705 }' 00:16:53.705 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:16:53.705 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:16:53.705 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:16:53.963 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:16:53.963 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:16:53.963 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:16:53.963 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:16:53.963 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:16:53.963 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:16:53.963 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:16:53.963 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:16:54.221 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:16:54.221 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:16:54.221 [2024-07-15 11:27:37.785351] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:54.221 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@275 -- # local expected_state 00:16:54.221 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # has_redundancy raid1 00:16:54.221 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # case $1 in 00:16:54.221 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@214 -- # return 0 00:16:54.221 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # expected_state=online 00:16:54.221 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:16:54.221 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:54.221 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:16:54.221 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:54.221 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:54.221 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:16:54.221 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:54.221 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:54.221 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:54.221 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:54.221 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:54.221 11:27:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:54.479 11:27:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:54.479 "name": "Existed_Raid", 00:16:54.479 "uuid": "e4f7ba56-0138-4064-8488-860886fd716f", 00:16:54.479 "strip_size_kb": 0, 00:16:54.479 "state": "online", 00:16:54.479 "raid_level": "raid1", 00:16:54.479 "superblock": true, 00:16:54.479 "num_base_bdevs": 3, 00:16:54.479 "num_base_bdevs_discovered": 2, 00:16:54.479 "num_base_bdevs_operational": 2, 00:16:54.479 "base_bdevs_list": [ 00:16:54.479 { 00:16:54.479 "name": null, 00:16:54.479 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:54.479 "is_configured": false, 00:16:54.479 "data_offset": 2048, 00:16:54.479 "data_size": 63488 00:16:54.479 }, 00:16:54.479 { 00:16:54.479 "name": "BaseBdev2", 00:16:54.479 "uuid": "3d9b03b4-e613-4597-8049-d9ca4fc39ca5", 00:16:54.479 "is_configured": true, 00:16:54.479 "data_offset": 2048, 00:16:54.479 "data_size": 63488 00:16:54.479 }, 00:16:54.479 { 00:16:54.479 "name": "BaseBdev3", 00:16:54.479 "uuid": "12b08651-c04b-4c82-ba89-399286cade07", 00:16:54.479 "is_configured": true, 00:16:54.479 "data_offset": 2048, 00:16:54.479 "data_size": 63488 00:16:54.479 } 00:16:54.479 ] 00:16:54.479 }' 00:16:54.479 11:27:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:54.479 11:27:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:55.045 11:27:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:16:55.045 11:27:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:16:55.045 11:27:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:55.301 11:27:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:16:55.301 11:27:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:16:55.301 11:27:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:16:55.301 11:27:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:16:55.559 [2024-07-15 11:27:39.113933] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:16:55.559 11:27:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:16:55.559 11:27:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:16:55.559 11:27:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:55.816 11:27:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:16:55.816 11:27:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:16:55.816 11:27:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:16:55.816 11:27:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev3 00:16:56.074 [2024-07-15 11:27:39.549798] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:16:56.074 [2024-07-15 11:27:39.549886] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:56.074 [2024-07-15 11:27:39.562360] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:56.074 [2024-07-15 11:27:39.562397] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:56.074 [2024-07-15 11:27:39.562409] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1b5c400 name Existed_Raid, state offline 00:16:56.074 11:27:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:16:56.074 11:27:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:16:56.074 11:27:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:56.074 11:27:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:16:56.333 11:27:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:16:56.333 11:27:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:16:56.333 11:27:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # '[' 3 -gt 2 ']' 00:16:56.333 11:27:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i = 1 )) 00:16:56.333 11:27:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:16:56.333 11:27:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:16:56.591 BaseBdev2 00:16:56.591 11:27:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev2 00:16:56.591 11:27:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:16:56.591 11:27:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:16:56.591 11:27:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:16:56.591 11:27:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:16:56.591 11:27:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:16:56.591 11:27:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:16:56.849 11:27:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:16:57.107 [ 00:16:57.107 { 00:16:57.107 "name": "BaseBdev2", 00:16:57.107 "aliases": [ 00:16:57.107 "970c6643-2866-479e-91a5-0a4d62d2e695" 00:16:57.107 ], 00:16:57.107 "product_name": "Malloc disk", 00:16:57.107 "block_size": 512, 00:16:57.107 "num_blocks": 65536, 00:16:57.107 "uuid": "970c6643-2866-479e-91a5-0a4d62d2e695", 00:16:57.107 "assigned_rate_limits": { 00:16:57.107 "rw_ios_per_sec": 0, 00:16:57.107 "rw_mbytes_per_sec": 0, 00:16:57.107 "r_mbytes_per_sec": 0, 00:16:57.107 "w_mbytes_per_sec": 0 00:16:57.107 }, 00:16:57.107 "claimed": false, 00:16:57.107 "zoned": false, 00:16:57.107 "supported_io_types": { 00:16:57.107 "read": true, 00:16:57.107 "write": true, 00:16:57.107 "unmap": true, 00:16:57.107 "flush": true, 00:16:57.107 "reset": true, 00:16:57.107 "nvme_admin": false, 00:16:57.107 "nvme_io": false, 00:16:57.107 "nvme_io_md": false, 00:16:57.107 "write_zeroes": true, 00:16:57.107 "zcopy": true, 00:16:57.107 "get_zone_info": false, 00:16:57.107 "zone_management": false, 00:16:57.107 "zone_append": false, 00:16:57.107 "compare": false, 00:16:57.107 "compare_and_write": false, 00:16:57.107 "abort": true, 00:16:57.107 "seek_hole": false, 00:16:57.107 "seek_data": false, 00:16:57.107 "copy": true, 00:16:57.107 "nvme_iov_md": false 00:16:57.107 }, 00:16:57.107 "memory_domains": [ 00:16:57.107 { 00:16:57.107 "dma_device_id": "system", 00:16:57.107 "dma_device_type": 1 00:16:57.107 }, 00:16:57.107 { 00:16:57.107 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:57.107 "dma_device_type": 2 00:16:57.107 } 00:16:57.107 ], 00:16:57.107 "driver_specific": {} 00:16:57.107 } 00:16:57.107 ] 00:16:57.107 11:27:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:16:57.107 11:27:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:16:57.107 11:27:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:16:57.107 11:27:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:16:57.107 BaseBdev3 00:16:57.366 11:27:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev3 00:16:57.366 11:27:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:16:57.366 11:27:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:16:57.366 11:27:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:16:57.366 11:27:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:16:57.366 11:27:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:16:57.366 11:27:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:16:57.366 11:27:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:16:57.625 [ 00:16:57.625 { 00:16:57.625 "name": "BaseBdev3", 00:16:57.625 "aliases": [ 00:16:57.625 "d4aff6c8-27c6-4eb8-8134-29a9caa3c9ed" 00:16:57.625 ], 00:16:57.625 "product_name": "Malloc disk", 00:16:57.625 "block_size": 512, 00:16:57.625 "num_blocks": 65536, 00:16:57.625 "uuid": "d4aff6c8-27c6-4eb8-8134-29a9caa3c9ed", 00:16:57.625 "assigned_rate_limits": { 00:16:57.625 "rw_ios_per_sec": 0, 00:16:57.625 "rw_mbytes_per_sec": 0, 00:16:57.625 "r_mbytes_per_sec": 0, 00:16:57.625 "w_mbytes_per_sec": 0 00:16:57.625 }, 00:16:57.625 "claimed": false, 00:16:57.625 "zoned": false, 00:16:57.625 "supported_io_types": { 00:16:57.625 "read": true, 00:16:57.625 "write": true, 00:16:57.625 "unmap": true, 00:16:57.625 "flush": true, 00:16:57.625 "reset": true, 00:16:57.625 "nvme_admin": false, 00:16:57.625 "nvme_io": false, 00:16:57.625 "nvme_io_md": false, 00:16:57.625 "write_zeroes": true, 00:16:57.625 "zcopy": true, 00:16:57.625 "get_zone_info": false, 00:16:57.625 "zone_management": false, 00:16:57.625 "zone_append": false, 00:16:57.625 "compare": false, 00:16:57.625 "compare_and_write": false, 00:16:57.625 "abort": true, 00:16:57.625 "seek_hole": false, 00:16:57.625 "seek_data": false, 00:16:57.625 "copy": true, 00:16:57.625 "nvme_iov_md": false 00:16:57.625 }, 00:16:57.625 "memory_domains": [ 00:16:57.625 { 00:16:57.625 "dma_device_id": "system", 00:16:57.625 "dma_device_type": 1 00:16:57.625 }, 00:16:57.625 { 00:16:57.625 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:57.625 "dma_device_type": 2 00:16:57.625 } 00:16:57.625 ], 00:16:57.625 "driver_specific": {} 00:16:57.625 } 00:16:57.625 ] 00:16:57.625 11:27:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:16:57.625 11:27:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:16:57.625 11:27:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:16:57.625 11:27:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@305 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n Existed_Raid 00:16:57.884 [2024-07-15 11:27:41.407786] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:57.884 [2024-07-15 11:27:41.407829] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:57.884 [2024-07-15 11:27:41.407847] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:57.884 [2024-07-15 11:27:41.409200] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:16:57.884 11:27:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:16:57.884 11:27:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:57.884 11:27:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:16:57.884 11:27:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:57.884 11:27:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:57.885 11:27:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:57.885 11:27:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:57.885 11:27:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:57.885 11:27:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:57.885 11:27:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:57.885 11:27:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:57.885 11:27:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:58.144 11:27:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:58.144 "name": "Existed_Raid", 00:16:58.144 "uuid": "4408285a-a8a0-4733-9339-2c93677d2fac", 00:16:58.144 "strip_size_kb": 0, 00:16:58.144 "state": "configuring", 00:16:58.144 "raid_level": "raid1", 00:16:58.144 "superblock": true, 00:16:58.144 "num_base_bdevs": 3, 00:16:58.144 "num_base_bdevs_discovered": 2, 00:16:58.144 "num_base_bdevs_operational": 3, 00:16:58.144 "base_bdevs_list": [ 00:16:58.144 { 00:16:58.144 "name": "BaseBdev1", 00:16:58.144 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:58.144 "is_configured": false, 00:16:58.144 "data_offset": 0, 00:16:58.144 "data_size": 0 00:16:58.144 }, 00:16:58.144 { 00:16:58.144 "name": "BaseBdev2", 00:16:58.144 "uuid": "970c6643-2866-479e-91a5-0a4d62d2e695", 00:16:58.144 "is_configured": true, 00:16:58.144 "data_offset": 2048, 00:16:58.144 "data_size": 63488 00:16:58.144 }, 00:16:58.144 { 00:16:58.144 "name": "BaseBdev3", 00:16:58.144 "uuid": "d4aff6c8-27c6-4eb8-8134-29a9caa3c9ed", 00:16:58.144 "is_configured": true, 00:16:58.144 "data_offset": 2048, 00:16:58.144 "data_size": 63488 00:16:58.144 } 00:16:58.144 ] 00:16:58.144 }' 00:16:58.144 11:27:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:58.144 11:27:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:58.713 11:27:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev2 00:16:58.972 [2024-07-15 11:27:42.490634] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:16:58.972 11:27:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@309 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:16:58.972 11:27:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:16:58.972 11:27:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:16:58.972 11:27:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:16:58.972 11:27:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:16:58.972 11:27:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:16:58.972 11:27:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:16:58.972 11:27:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:16:58.972 11:27:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:16:58.972 11:27:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:16:58.972 11:27:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:58.972 11:27:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:59.231 11:27:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:16:59.231 "name": "Existed_Raid", 00:16:59.231 "uuid": "4408285a-a8a0-4733-9339-2c93677d2fac", 00:16:59.231 "strip_size_kb": 0, 00:16:59.231 "state": "configuring", 00:16:59.231 "raid_level": "raid1", 00:16:59.231 "superblock": true, 00:16:59.231 "num_base_bdevs": 3, 00:16:59.231 "num_base_bdevs_discovered": 1, 00:16:59.231 "num_base_bdevs_operational": 3, 00:16:59.231 "base_bdevs_list": [ 00:16:59.231 { 00:16:59.231 "name": "BaseBdev1", 00:16:59.231 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:59.231 "is_configured": false, 00:16:59.231 "data_offset": 0, 00:16:59.231 "data_size": 0 00:16:59.231 }, 00:16:59.231 { 00:16:59.231 "name": null, 00:16:59.231 "uuid": "970c6643-2866-479e-91a5-0a4d62d2e695", 00:16:59.231 "is_configured": false, 00:16:59.231 "data_offset": 2048, 00:16:59.231 "data_size": 63488 00:16:59.231 }, 00:16:59.231 { 00:16:59.231 "name": "BaseBdev3", 00:16:59.231 "uuid": "d4aff6c8-27c6-4eb8-8134-29a9caa3c9ed", 00:16:59.231 "is_configured": true, 00:16:59.231 "data_offset": 2048, 00:16:59.231 "data_size": 63488 00:16:59.231 } 00:16:59.231 ] 00:16:59.231 }' 00:16:59.231 11:27:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:16:59.231 11:27:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:59.801 11:27:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:16:59.801 11:27:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:17:00.061 11:27:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # [[ false == \f\a\l\s\e ]] 00:17:00.061 11:27:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:17:00.319 [2024-07-15 11:27:43.670364] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:00.319 BaseBdev1 00:17:00.319 11:27:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@313 -- # waitforbdev BaseBdev1 00:17:00.319 11:27:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:17:00.319 11:27:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:17:00.319 11:27:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:17:00.319 11:27:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:17:00.319 11:27:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:17:00.319 11:27:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:17:00.577 11:27:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:17:00.577 [ 00:17:00.577 { 00:17:00.577 "name": "BaseBdev1", 00:17:00.577 "aliases": [ 00:17:00.577 "b880da56-6dc1-4635-8567-75bc8625ca2b" 00:17:00.577 ], 00:17:00.577 "product_name": "Malloc disk", 00:17:00.577 "block_size": 512, 00:17:00.577 "num_blocks": 65536, 00:17:00.577 "uuid": "b880da56-6dc1-4635-8567-75bc8625ca2b", 00:17:00.577 "assigned_rate_limits": { 00:17:00.577 "rw_ios_per_sec": 0, 00:17:00.577 "rw_mbytes_per_sec": 0, 00:17:00.577 "r_mbytes_per_sec": 0, 00:17:00.577 "w_mbytes_per_sec": 0 00:17:00.577 }, 00:17:00.577 "claimed": true, 00:17:00.577 "claim_type": "exclusive_write", 00:17:00.577 "zoned": false, 00:17:00.577 "supported_io_types": { 00:17:00.577 "read": true, 00:17:00.577 "write": true, 00:17:00.577 "unmap": true, 00:17:00.577 "flush": true, 00:17:00.577 "reset": true, 00:17:00.577 "nvme_admin": false, 00:17:00.577 "nvme_io": false, 00:17:00.577 "nvme_io_md": false, 00:17:00.577 "write_zeroes": true, 00:17:00.577 "zcopy": true, 00:17:00.577 "get_zone_info": false, 00:17:00.577 "zone_management": false, 00:17:00.577 "zone_append": false, 00:17:00.577 "compare": false, 00:17:00.577 "compare_and_write": false, 00:17:00.577 "abort": true, 00:17:00.577 "seek_hole": false, 00:17:00.577 "seek_data": false, 00:17:00.577 "copy": true, 00:17:00.577 "nvme_iov_md": false 00:17:00.577 }, 00:17:00.577 "memory_domains": [ 00:17:00.577 { 00:17:00.577 "dma_device_id": "system", 00:17:00.577 "dma_device_type": 1 00:17:00.577 }, 00:17:00.577 { 00:17:00.577 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:00.577 "dma_device_type": 2 00:17:00.577 } 00:17:00.577 ], 00:17:00.577 "driver_specific": {} 00:17:00.577 } 00:17:00.577 ] 00:17:00.836 11:27:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:17:00.836 11:27:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:17:00.836 11:27:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:17:00.836 11:27:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:17:00.836 11:27:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:17:00.836 11:27:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:17:00.836 11:27:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:17:00.836 11:27:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:00.836 11:27:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:00.836 11:27:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:00.836 11:27:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:00.836 11:27:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:00.836 11:27:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:00.836 11:27:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:00.836 "name": "Existed_Raid", 00:17:00.836 "uuid": "4408285a-a8a0-4733-9339-2c93677d2fac", 00:17:00.836 "strip_size_kb": 0, 00:17:00.836 "state": "configuring", 00:17:00.836 "raid_level": "raid1", 00:17:00.836 "superblock": true, 00:17:00.836 "num_base_bdevs": 3, 00:17:00.836 "num_base_bdevs_discovered": 2, 00:17:00.836 "num_base_bdevs_operational": 3, 00:17:00.836 "base_bdevs_list": [ 00:17:00.836 { 00:17:00.836 "name": "BaseBdev1", 00:17:00.836 "uuid": "b880da56-6dc1-4635-8567-75bc8625ca2b", 00:17:00.836 "is_configured": true, 00:17:00.836 "data_offset": 2048, 00:17:00.836 "data_size": 63488 00:17:00.836 }, 00:17:00.836 { 00:17:00.836 "name": null, 00:17:00.836 "uuid": "970c6643-2866-479e-91a5-0a4d62d2e695", 00:17:00.836 "is_configured": false, 00:17:00.836 "data_offset": 2048, 00:17:00.836 "data_size": 63488 00:17:00.836 }, 00:17:00.836 { 00:17:00.836 "name": "BaseBdev3", 00:17:00.836 "uuid": "d4aff6c8-27c6-4eb8-8134-29a9caa3c9ed", 00:17:00.836 "is_configured": true, 00:17:00.836 "data_offset": 2048, 00:17:00.836 "data_size": 63488 00:17:00.836 } 00:17:00.836 ] 00:17:00.836 }' 00:17:00.836 11:27:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:00.836 11:27:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:17:01.405 11:27:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:01.405 11:27:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:17:01.665 11:27:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # [[ true == \t\r\u\e ]] 00:17:01.665 11:27:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@317 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev3 00:17:02.270 [2024-07-15 11:27:45.603526] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:17:02.270 11:27:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:17:02.270 11:27:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:17:02.270 11:27:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:17:02.270 11:27:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:17:02.270 11:27:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:17:02.271 11:27:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:17:02.271 11:27:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:02.271 11:27:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:02.271 11:27:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:02.271 11:27:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:02.271 11:27:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:02.271 11:27:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:02.530 11:27:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:02.530 "name": "Existed_Raid", 00:17:02.530 "uuid": "4408285a-a8a0-4733-9339-2c93677d2fac", 00:17:02.530 "strip_size_kb": 0, 00:17:02.530 "state": "configuring", 00:17:02.530 "raid_level": "raid1", 00:17:02.530 "superblock": true, 00:17:02.530 "num_base_bdevs": 3, 00:17:02.530 "num_base_bdevs_discovered": 1, 00:17:02.530 "num_base_bdevs_operational": 3, 00:17:02.530 "base_bdevs_list": [ 00:17:02.530 { 00:17:02.530 "name": "BaseBdev1", 00:17:02.530 "uuid": "b880da56-6dc1-4635-8567-75bc8625ca2b", 00:17:02.530 "is_configured": true, 00:17:02.530 "data_offset": 2048, 00:17:02.530 "data_size": 63488 00:17:02.530 }, 00:17:02.530 { 00:17:02.530 "name": null, 00:17:02.530 "uuid": "970c6643-2866-479e-91a5-0a4d62d2e695", 00:17:02.530 "is_configured": false, 00:17:02.530 "data_offset": 2048, 00:17:02.530 "data_size": 63488 00:17:02.530 }, 00:17:02.530 { 00:17:02.530 "name": null, 00:17:02.530 "uuid": "d4aff6c8-27c6-4eb8-8134-29a9caa3c9ed", 00:17:02.530 "is_configured": false, 00:17:02.530 "data_offset": 2048, 00:17:02.530 "data_size": 63488 00:17:02.530 } 00:17:02.530 ] 00:17:02.530 }' 00:17:02.530 11:27:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:02.530 11:27:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:17:03.098 11:27:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:03.098 11:27:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:17:03.098 11:27:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # [[ false == \f\a\l\s\e ]] 00:17:03.098 11:27:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:17:03.357 [2024-07-15 11:27:46.870900] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:17:03.358 11:27:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@322 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:17:03.358 11:27:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:17:03.358 11:27:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:17:03.358 11:27:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:17:03.358 11:27:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:17:03.358 11:27:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:17:03.358 11:27:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:03.358 11:27:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:03.358 11:27:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:03.358 11:27:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:03.358 11:27:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:03.358 11:27:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:03.617 11:27:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:03.617 "name": "Existed_Raid", 00:17:03.617 "uuid": "4408285a-a8a0-4733-9339-2c93677d2fac", 00:17:03.617 "strip_size_kb": 0, 00:17:03.617 "state": "configuring", 00:17:03.617 "raid_level": "raid1", 00:17:03.617 "superblock": true, 00:17:03.617 "num_base_bdevs": 3, 00:17:03.617 "num_base_bdevs_discovered": 2, 00:17:03.617 "num_base_bdevs_operational": 3, 00:17:03.617 "base_bdevs_list": [ 00:17:03.617 { 00:17:03.617 "name": "BaseBdev1", 00:17:03.617 "uuid": "b880da56-6dc1-4635-8567-75bc8625ca2b", 00:17:03.617 "is_configured": true, 00:17:03.617 "data_offset": 2048, 00:17:03.617 "data_size": 63488 00:17:03.617 }, 00:17:03.617 { 00:17:03.617 "name": null, 00:17:03.617 "uuid": "970c6643-2866-479e-91a5-0a4d62d2e695", 00:17:03.617 "is_configured": false, 00:17:03.617 "data_offset": 2048, 00:17:03.617 "data_size": 63488 00:17:03.617 }, 00:17:03.617 { 00:17:03.617 "name": "BaseBdev3", 00:17:03.617 "uuid": "d4aff6c8-27c6-4eb8-8134-29a9caa3c9ed", 00:17:03.617 "is_configured": true, 00:17:03.617 "data_offset": 2048, 00:17:03.617 "data_size": 63488 00:17:03.617 } 00:17:03.617 ] 00:17:03.617 }' 00:17:03.617 11:27:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:03.617 11:27:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:17:04.185 11:27:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:04.185 11:27:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:17:04.445 11:27:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # [[ true == \t\r\u\e ]] 00:17:04.445 11:27:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@325 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:17:04.704 [2024-07-15 11:27:48.202459] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:17:04.704 11:27:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:17:04.704 11:27:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:17:04.704 11:27:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:17:04.704 11:27:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:17:04.704 11:27:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:17:04.704 11:27:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:17:04.704 11:27:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:04.704 11:27:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:04.704 11:27:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:04.704 11:27:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:04.704 11:27:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:04.704 11:27:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:04.963 11:27:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:04.963 "name": "Existed_Raid", 00:17:04.963 "uuid": "4408285a-a8a0-4733-9339-2c93677d2fac", 00:17:04.963 "strip_size_kb": 0, 00:17:04.963 "state": "configuring", 00:17:04.963 "raid_level": "raid1", 00:17:04.963 "superblock": true, 00:17:04.963 "num_base_bdevs": 3, 00:17:04.963 "num_base_bdevs_discovered": 1, 00:17:04.963 "num_base_bdevs_operational": 3, 00:17:04.963 "base_bdevs_list": [ 00:17:04.963 { 00:17:04.963 "name": null, 00:17:04.963 "uuid": "b880da56-6dc1-4635-8567-75bc8625ca2b", 00:17:04.963 "is_configured": false, 00:17:04.963 "data_offset": 2048, 00:17:04.963 "data_size": 63488 00:17:04.963 }, 00:17:04.963 { 00:17:04.963 "name": null, 00:17:04.963 "uuid": "970c6643-2866-479e-91a5-0a4d62d2e695", 00:17:04.963 "is_configured": false, 00:17:04.963 "data_offset": 2048, 00:17:04.964 "data_size": 63488 00:17:04.964 }, 00:17:04.964 { 00:17:04.964 "name": "BaseBdev3", 00:17:04.964 "uuid": "d4aff6c8-27c6-4eb8-8134-29a9caa3c9ed", 00:17:04.964 "is_configured": true, 00:17:04.964 "data_offset": 2048, 00:17:04.964 "data_size": 63488 00:17:04.964 } 00:17:04.964 ] 00:17:04.964 }' 00:17:04.964 11:27:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:04.964 11:27:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:17:05.532 11:27:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@327 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:05.532 11:27:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@327 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:17:05.791 11:27:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@327 -- # [[ false == \f\a\l\s\e ]] 00:17:05.791 11:27:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@329 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:17:06.050 [2024-07-15 11:27:49.560756] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:06.050 11:27:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@330 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:17:06.050 11:27:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:17:06.050 11:27:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:17:06.050 11:27:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:17:06.050 11:27:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:17:06.050 11:27:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:17:06.050 11:27:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:06.050 11:27:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:06.050 11:27:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:06.050 11:27:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:06.050 11:27:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:06.050 11:27:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:06.310 11:27:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:06.310 "name": "Existed_Raid", 00:17:06.310 "uuid": "4408285a-a8a0-4733-9339-2c93677d2fac", 00:17:06.310 "strip_size_kb": 0, 00:17:06.310 "state": "configuring", 00:17:06.310 "raid_level": "raid1", 00:17:06.310 "superblock": true, 00:17:06.310 "num_base_bdevs": 3, 00:17:06.310 "num_base_bdevs_discovered": 2, 00:17:06.310 "num_base_bdevs_operational": 3, 00:17:06.310 "base_bdevs_list": [ 00:17:06.310 { 00:17:06.310 "name": null, 00:17:06.310 "uuid": "b880da56-6dc1-4635-8567-75bc8625ca2b", 00:17:06.310 "is_configured": false, 00:17:06.310 "data_offset": 2048, 00:17:06.310 "data_size": 63488 00:17:06.310 }, 00:17:06.310 { 00:17:06.310 "name": "BaseBdev2", 00:17:06.310 "uuid": "970c6643-2866-479e-91a5-0a4d62d2e695", 00:17:06.310 "is_configured": true, 00:17:06.310 "data_offset": 2048, 00:17:06.310 "data_size": 63488 00:17:06.310 }, 00:17:06.310 { 00:17:06.310 "name": "BaseBdev3", 00:17:06.310 "uuid": "d4aff6c8-27c6-4eb8-8134-29a9caa3c9ed", 00:17:06.310 "is_configured": true, 00:17:06.310 "data_offset": 2048, 00:17:06.310 "data_size": 63488 00:17:06.310 } 00:17:06.310 ] 00:17:06.310 }' 00:17:06.310 11:27:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:06.310 11:27:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:17:06.879 11:27:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@331 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:06.879 11:27:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@331 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:17:07.138 11:27:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@331 -- # [[ true == \t\r\u\e ]] 00:17:07.138 11:27:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:07.138 11:27:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@333 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:17:07.398 11:27:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b NewBaseBdev -u b880da56-6dc1-4635-8567-75bc8625ca2b 00:17:07.658 [2024-07-15 11:27:51.077049] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:17:07.658 [2024-07-15 11:27:51.077201] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1b521b0 00:17:07.658 [2024-07-15 11:27:51.077215] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:17:07.658 [2024-07-15 11:27:51.077387] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1d0e4f0 00:17:07.658 [2024-07-15 11:27:51.077508] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1b521b0 00:17:07.658 [2024-07-15 11:27:51.077518] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x1b521b0 00:17:07.658 [2024-07-15 11:27:51.077612] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:07.658 NewBaseBdev 00:17:07.658 11:27:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@334 -- # waitforbdev NewBaseBdev 00:17:07.658 11:27:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=NewBaseBdev 00:17:07.658 11:27:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:17:07.658 11:27:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:17:07.658 11:27:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:17:07.658 11:27:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:17:07.658 11:27:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:17:07.917 11:27:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev -t 2000 00:17:08.176 [ 00:17:08.176 { 00:17:08.176 "name": "NewBaseBdev", 00:17:08.176 "aliases": [ 00:17:08.176 "b880da56-6dc1-4635-8567-75bc8625ca2b" 00:17:08.176 ], 00:17:08.176 "product_name": "Malloc disk", 00:17:08.176 "block_size": 512, 00:17:08.176 "num_blocks": 65536, 00:17:08.176 "uuid": "b880da56-6dc1-4635-8567-75bc8625ca2b", 00:17:08.176 "assigned_rate_limits": { 00:17:08.176 "rw_ios_per_sec": 0, 00:17:08.176 "rw_mbytes_per_sec": 0, 00:17:08.176 "r_mbytes_per_sec": 0, 00:17:08.176 "w_mbytes_per_sec": 0 00:17:08.176 }, 00:17:08.176 "claimed": true, 00:17:08.176 "claim_type": "exclusive_write", 00:17:08.176 "zoned": false, 00:17:08.176 "supported_io_types": { 00:17:08.176 "read": true, 00:17:08.176 "write": true, 00:17:08.176 "unmap": true, 00:17:08.176 "flush": true, 00:17:08.176 "reset": true, 00:17:08.176 "nvme_admin": false, 00:17:08.176 "nvme_io": false, 00:17:08.176 "nvme_io_md": false, 00:17:08.176 "write_zeroes": true, 00:17:08.176 "zcopy": true, 00:17:08.176 "get_zone_info": false, 00:17:08.176 "zone_management": false, 00:17:08.176 "zone_append": false, 00:17:08.176 "compare": false, 00:17:08.176 "compare_and_write": false, 00:17:08.176 "abort": true, 00:17:08.176 "seek_hole": false, 00:17:08.176 "seek_data": false, 00:17:08.176 "copy": true, 00:17:08.176 "nvme_iov_md": false 00:17:08.176 }, 00:17:08.176 "memory_domains": [ 00:17:08.176 { 00:17:08.176 "dma_device_id": "system", 00:17:08.176 "dma_device_type": 1 00:17:08.176 }, 00:17:08.176 { 00:17:08.176 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:08.176 "dma_device_type": 2 00:17:08.176 } 00:17:08.176 ], 00:17:08.176 "driver_specific": {} 00:17:08.176 } 00:17:08.176 ] 00:17:08.176 11:27:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:17:08.176 11:27:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@335 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:17:08.176 11:27:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:17:08.176 11:27:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:17:08.176 11:27:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:17:08.176 11:27:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:17:08.176 11:27:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:17:08.176 11:27:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:08.176 11:27:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:08.176 11:27:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:08.176 11:27:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:08.176 11:27:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:08.176 11:27:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:08.435 11:27:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:08.435 "name": "Existed_Raid", 00:17:08.435 "uuid": "4408285a-a8a0-4733-9339-2c93677d2fac", 00:17:08.435 "strip_size_kb": 0, 00:17:08.435 "state": "online", 00:17:08.435 "raid_level": "raid1", 00:17:08.435 "superblock": true, 00:17:08.435 "num_base_bdevs": 3, 00:17:08.435 "num_base_bdevs_discovered": 3, 00:17:08.435 "num_base_bdevs_operational": 3, 00:17:08.435 "base_bdevs_list": [ 00:17:08.435 { 00:17:08.435 "name": "NewBaseBdev", 00:17:08.435 "uuid": "b880da56-6dc1-4635-8567-75bc8625ca2b", 00:17:08.435 "is_configured": true, 00:17:08.435 "data_offset": 2048, 00:17:08.435 "data_size": 63488 00:17:08.435 }, 00:17:08.435 { 00:17:08.435 "name": "BaseBdev2", 00:17:08.436 "uuid": "970c6643-2866-479e-91a5-0a4d62d2e695", 00:17:08.436 "is_configured": true, 00:17:08.436 "data_offset": 2048, 00:17:08.436 "data_size": 63488 00:17:08.436 }, 00:17:08.436 { 00:17:08.436 "name": "BaseBdev3", 00:17:08.436 "uuid": "d4aff6c8-27c6-4eb8-8134-29a9caa3c9ed", 00:17:08.436 "is_configured": true, 00:17:08.436 "data_offset": 2048, 00:17:08.436 "data_size": 63488 00:17:08.436 } 00:17:08.436 ] 00:17:08.436 }' 00:17:08.436 11:27:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:08.436 11:27:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:17:09.003 11:27:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@336 -- # verify_raid_bdev_properties Existed_Raid 00:17:09.003 11:27:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:17:09.003 11:27:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:17:09.003 11:27:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:17:09.003 11:27:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:17:09.003 11:27:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # local name 00:17:09.003 11:27:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:17:09.003 11:27:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:17:09.263 [2024-07-15 11:27:52.657547] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:09.263 11:27:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:17:09.263 "name": "Existed_Raid", 00:17:09.263 "aliases": [ 00:17:09.263 "4408285a-a8a0-4733-9339-2c93677d2fac" 00:17:09.263 ], 00:17:09.263 "product_name": "Raid Volume", 00:17:09.263 "block_size": 512, 00:17:09.263 "num_blocks": 63488, 00:17:09.263 "uuid": "4408285a-a8a0-4733-9339-2c93677d2fac", 00:17:09.263 "assigned_rate_limits": { 00:17:09.263 "rw_ios_per_sec": 0, 00:17:09.263 "rw_mbytes_per_sec": 0, 00:17:09.263 "r_mbytes_per_sec": 0, 00:17:09.263 "w_mbytes_per_sec": 0 00:17:09.263 }, 00:17:09.263 "claimed": false, 00:17:09.263 "zoned": false, 00:17:09.263 "supported_io_types": { 00:17:09.263 "read": true, 00:17:09.263 "write": true, 00:17:09.263 "unmap": false, 00:17:09.263 "flush": false, 00:17:09.263 "reset": true, 00:17:09.263 "nvme_admin": false, 00:17:09.263 "nvme_io": false, 00:17:09.263 "nvme_io_md": false, 00:17:09.263 "write_zeroes": true, 00:17:09.263 "zcopy": false, 00:17:09.263 "get_zone_info": false, 00:17:09.263 "zone_management": false, 00:17:09.263 "zone_append": false, 00:17:09.263 "compare": false, 00:17:09.263 "compare_and_write": false, 00:17:09.263 "abort": false, 00:17:09.263 "seek_hole": false, 00:17:09.263 "seek_data": false, 00:17:09.263 "copy": false, 00:17:09.263 "nvme_iov_md": false 00:17:09.263 }, 00:17:09.263 "memory_domains": [ 00:17:09.263 { 00:17:09.263 "dma_device_id": "system", 00:17:09.263 "dma_device_type": 1 00:17:09.263 }, 00:17:09.263 { 00:17:09.263 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:09.263 "dma_device_type": 2 00:17:09.263 }, 00:17:09.263 { 00:17:09.263 "dma_device_id": "system", 00:17:09.263 "dma_device_type": 1 00:17:09.263 }, 00:17:09.263 { 00:17:09.263 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:09.263 "dma_device_type": 2 00:17:09.263 }, 00:17:09.263 { 00:17:09.263 "dma_device_id": "system", 00:17:09.263 "dma_device_type": 1 00:17:09.263 }, 00:17:09.263 { 00:17:09.263 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:09.263 "dma_device_type": 2 00:17:09.263 } 00:17:09.263 ], 00:17:09.263 "driver_specific": { 00:17:09.263 "raid": { 00:17:09.263 "uuid": "4408285a-a8a0-4733-9339-2c93677d2fac", 00:17:09.263 "strip_size_kb": 0, 00:17:09.263 "state": "online", 00:17:09.263 "raid_level": "raid1", 00:17:09.263 "superblock": true, 00:17:09.263 "num_base_bdevs": 3, 00:17:09.263 "num_base_bdevs_discovered": 3, 00:17:09.263 "num_base_bdevs_operational": 3, 00:17:09.263 "base_bdevs_list": [ 00:17:09.263 { 00:17:09.263 "name": "NewBaseBdev", 00:17:09.263 "uuid": "b880da56-6dc1-4635-8567-75bc8625ca2b", 00:17:09.263 "is_configured": true, 00:17:09.263 "data_offset": 2048, 00:17:09.263 "data_size": 63488 00:17:09.263 }, 00:17:09.263 { 00:17:09.263 "name": "BaseBdev2", 00:17:09.263 "uuid": "970c6643-2866-479e-91a5-0a4d62d2e695", 00:17:09.263 "is_configured": true, 00:17:09.263 "data_offset": 2048, 00:17:09.263 "data_size": 63488 00:17:09.263 }, 00:17:09.263 { 00:17:09.263 "name": "BaseBdev3", 00:17:09.263 "uuid": "d4aff6c8-27c6-4eb8-8134-29a9caa3c9ed", 00:17:09.263 "is_configured": true, 00:17:09.263 "data_offset": 2048, 00:17:09.263 "data_size": 63488 00:17:09.263 } 00:17:09.263 ] 00:17:09.263 } 00:17:09.263 } 00:17:09.263 }' 00:17:09.263 11:27:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:17:09.263 11:27:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # base_bdev_names='NewBaseBdev 00:17:09.263 BaseBdev2 00:17:09.263 BaseBdev3' 00:17:09.263 11:27:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:17:09.263 11:27:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev 00:17:09.263 11:27:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:17:09.522 11:27:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:17:09.522 "name": "NewBaseBdev", 00:17:09.522 "aliases": [ 00:17:09.522 "b880da56-6dc1-4635-8567-75bc8625ca2b" 00:17:09.522 ], 00:17:09.522 "product_name": "Malloc disk", 00:17:09.522 "block_size": 512, 00:17:09.522 "num_blocks": 65536, 00:17:09.522 "uuid": "b880da56-6dc1-4635-8567-75bc8625ca2b", 00:17:09.522 "assigned_rate_limits": { 00:17:09.522 "rw_ios_per_sec": 0, 00:17:09.522 "rw_mbytes_per_sec": 0, 00:17:09.522 "r_mbytes_per_sec": 0, 00:17:09.522 "w_mbytes_per_sec": 0 00:17:09.522 }, 00:17:09.522 "claimed": true, 00:17:09.522 "claim_type": "exclusive_write", 00:17:09.522 "zoned": false, 00:17:09.522 "supported_io_types": { 00:17:09.522 "read": true, 00:17:09.522 "write": true, 00:17:09.522 "unmap": true, 00:17:09.522 "flush": true, 00:17:09.522 "reset": true, 00:17:09.522 "nvme_admin": false, 00:17:09.522 "nvme_io": false, 00:17:09.522 "nvme_io_md": false, 00:17:09.522 "write_zeroes": true, 00:17:09.522 "zcopy": true, 00:17:09.522 "get_zone_info": false, 00:17:09.522 "zone_management": false, 00:17:09.522 "zone_append": false, 00:17:09.522 "compare": false, 00:17:09.522 "compare_and_write": false, 00:17:09.522 "abort": true, 00:17:09.522 "seek_hole": false, 00:17:09.522 "seek_data": false, 00:17:09.522 "copy": true, 00:17:09.522 "nvme_iov_md": false 00:17:09.522 }, 00:17:09.522 "memory_domains": [ 00:17:09.522 { 00:17:09.522 "dma_device_id": "system", 00:17:09.522 "dma_device_type": 1 00:17:09.522 }, 00:17:09.522 { 00:17:09.522 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:09.522 "dma_device_type": 2 00:17:09.522 } 00:17:09.522 ], 00:17:09.522 "driver_specific": {} 00:17:09.523 }' 00:17:09.523 11:27:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:09.523 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:09.523 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:17:09.523 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:09.523 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:09.781 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:17:09.781 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:09.781 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:09.781 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:17:09.781 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:09.781 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:09.781 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:17:09.781 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:17:09.781 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:17:09.781 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:17:10.040 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:17:10.040 "name": "BaseBdev2", 00:17:10.040 "aliases": [ 00:17:10.040 "970c6643-2866-479e-91a5-0a4d62d2e695" 00:17:10.040 ], 00:17:10.040 "product_name": "Malloc disk", 00:17:10.040 "block_size": 512, 00:17:10.040 "num_blocks": 65536, 00:17:10.040 "uuid": "970c6643-2866-479e-91a5-0a4d62d2e695", 00:17:10.040 "assigned_rate_limits": { 00:17:10.040 "rw_ios_per_sec": 0, 00:17:10.040 "rw_mbytes_per_sec": 0, 00:17:10.040 "r_mbytes_per_sec": 0, 00:17:10.040 "w_mbytes_per_sec": 0 00:17:10.040 }, 00:17:10.040 "claimed": true, 00:17:10.040 "claim_type": "exclusive_write", 00:17:10.040 "zoned": false, 00:17:10.040 "supported_io_types": { 00:17:10.040 "read": true, 00:17:10.040 "write": true, 00:17:10.040 "unmap": true, 00:17:10.040 "flush": true, 00:17:10.040 "reset": true, 00:17:10.040 "nvme_admin": false, 00:17:10.040 "nvme_io": false, 00:17:10.040 "nvme_io_md": false, 00:17:10.040 "write_zeroes": true, 00:17:10.040 "zcopy": true, 00:17:10.040 "get_zone_info": false, 00:17:10.040 "zone_management": false, 00:17:10.040 "zone_append": false, 00:17:10.040 "compare": false, 00:17:10.040 "compare_and_write": false, 00:17:10.040 "abort": true, 00:17:10.040 "seek_hole": false, 00:17:10.040 "seek_data": false, 00:17:10.040 "copy": true, 00:17:10.040 "nvme_iov_md": false 00:17:10.041 }, 00:17:10.041 "memory_domains": [ 00:17:10.041 { 00:17:10.041 "dma_device_id": "system", 00:17:10.041 "dma_device_type": 1 00:17:10.041 }, 00:17:10.041 { 00:17:10.041 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:10.041 "dma_device_type": 2 00:17:10.041 } 00:17:10.041 ], 00:17:10.041 "driver_specific": {} 00:17:10.041 }' 00:17:10.041 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:10.041 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:10.299 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:17:10.299 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:10.299 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:10.299 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:17:10.299 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:10.299 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:10.299 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:17:10.299 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:10.299 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:10.557 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:17:10.557 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:17:10.557 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:17:10.557 11:27:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:17:10.557 11:27:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:17:10.557 "name": "BaseBdev3", 00:17:10.557 "aliases": [ 00:17:10.557 "d4aff6c8-27c6-4eb8-8134-29a9caa3c9ed" 00:17:10.557 ], 00:17:10.557 "product_name": "Malloc disk", 00:17:10.557 "block_size": 512, 00:17:10.557 "num_blocks": 65536, 00:17:10.557 "uuid": "d4aff6c8-27c6-4eb8-8134-29a9caa3c9ed", 00:17:10.557 "assigned_rate_limits": { 00:17:10.557 "rw_ios_per_sec": 0, 00:17:10.557 "rw_mbytes_per_sec": 0, 00:17:10.557 "r_mbytes_per_sec": 0, 00:17:10.557 "w_mbytes_per_sec": 0 00:17:10.557 }, 00:17:10.557 "claimed": true, 00:17:10.557 "claim_type": "exclusive_write", 00:17:10.557 "zoned": false, 00:17:10.557 "supported_io_types": { 00:17:10.557 "read": true, 00:17:10.557 "write": true, 00:17:10.557 "unmap": true, 00:17:10.557 "flush": true, 00:17:10.557 "reset": true, 00:17:10.557 "nvme_admin": false, 00:17:10.557 "nvme_io": false, 00:17:10.557 "nvme_io_md": false, 00:17:10.557 "write_zeroes": true, 00:17:10.557 "zcopy": true, 00:17:10.557 "get_zone_info": false, 00:17:10.557 "zone_management": false, 00:17:10.557 "zone_append": false, 00:17:10.557 "compare": false, 00:17:10.557 "compare_and_write": false, 00:17:10.557 "abort": true, 00:17:10.557 "seek_hole": false, 00:17:10.557 "seek_data": false, 00:17:10.557 "copy": true, 00:17:10.557 "nvme_iov_md": false 00:17:10.557 }, 00:17:10.557 "memory_domains": [ 00:17:10.557 { 00:17:10.557 "dma_device_id": "system", 00:17:10.557 "dma_device_type": 1 00:17:10.557 }, 00:17:10.557 { 00:17:10.557 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:10.557 "dma_device_type": 2 00:17:10.557 } 00:17:10.557 ], 00:17:10.557 "driver_specific": {} 00:17:10.557 }' 00:17:10.557 11:27:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:10.816 11:27:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:10.816 11:27:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:17:10.816 11:27:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:10.816 11:27:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:10.816 11:27:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:17:10.816 11:27:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:10.816 11:27:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:10.816 11:27:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:17:10.816 11:27:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:11.075 11:27:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:11.075 11:27:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:17:11.075 11:27:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@338 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:17:11.334 [2024-07-15 11:27:54.698703] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:17:11.334 [2024-07-15 11:27:54.698728] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:11.334 [2024-07-15 11:27:54.698779] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:11.334 [2024-07-15 11:27:54.699060] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:11.334 [2024-07-15 11:27:54.699073] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1b521b0 name Existed_Raid, state offline 00:17:11.334 11:27:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@341 -- # killprocess 913209 00:17:11.334 11:27:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@948 -- # '[' -z 913209 ']' 00:17:11.334 11:27:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@952 -- # kill -0 913209 00:17:11.334 11:27:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@953 -- # uname 00:17:11.334 11:27:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:17:11.334 11:27:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 913209 00:17:11.334 11:27:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:17:11.334 11:27:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:17:11.334 11:27:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@966 -- # echo 'killing process with pid 913209' 00:17:11.334 killing process with pid 913209 00:17:11.334 11:27:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@967 -- # kill 913209 00:17:11.334 [2024-07-15 11:27:54.765181] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:11.334 11:27:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # wait 913209 00:17:11.334 [2024-07-15 11:27:54.795122] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:11.593 11:27:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@343 -- # return 0 00:17:11.593 00:17:11.593 real 0m28.724s 00:17:11.593 user 0m52.682s 00:17:11.593 sys 0m5.117s 00:17:11.593 11:27:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1124 -- # xtrace_disable 00:17:11.593 11:27:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:17:11.593 ************************************ 00:17:11.593 END TEST raid_state_function_test_sb 00:17:11.593 ************************************ 00:17:11.593 11:27:55 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:17:11.593 11:27:55 bdev_raid -- bdev/bdev_raid.sh@869 -- # run_test raid_superblock_test raid_superblock_test raid1 3 00:17:11.593 11:27:55 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:17:11.593 11:27:55 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:17:11.593 11:27:55 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:11.593 ************************************ 00:17:11.593 START TEST raid_superblock_test 00:17:11.593 ************************************ 00:17:11.593 11:27:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1123 -- # raid_superblock_test raid1 3 00:17:11.593 11:27:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@392 -- # local raid_level=raid1 00:17:11.593 11:27:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local num_base_bdevs=3 00:17:11.593 11:27:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # base_bdevs_malloc=() 00:17:11.593 11:27:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local base_bdevs_malloc 00:17:11.593 11:27:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_pt=() 00:17:11.593 11:27:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_pt 00:17:11.593 11:27:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt_uuid=() 00:17:11.593 11:27:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt_uuid 00:17:11.593 11:27:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local raid_bdev_name=raid_bdev1 00:17:11.593 11:27:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local strip_size 00:17:11.593 11:27:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size_create_arg 00:17:11.593 11:27:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local raid_bdev_uuid 00:17:11.593 11:27:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev 00:17:11.593 11:27:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@403 -- # '[' raid1 '!=' raid1 ']' 00:17:11.593 11:27:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@407 -- # strip_size=0 00:17:11.593 11:27:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # raid_pid=917573 00:17:11.593 11:27:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # waitforlisten 917573 /var/tmp/spdk-raid.sock 00:17:11.594 11:27:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@410 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -L bdev_raid 00:17:11.594 11:27:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@829 -- # '[' -z 917573 ']' 00:17:11.594 11:27:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:17:11.594 11:27:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:17:11.594 11:27:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:17:11.594 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:17:11.594 11:27:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:17:11.594 11:27:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:17:11.594 [2024-07-15 11:27:55.165818] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:17:11.594 [2024-07-15 11:27:55.165873] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid917573 ] 00:17:11.853 [2024-07-15 11:27:55.278255] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:11.853 [2024-07-15 11:27:55.382490] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:17:11.853 [2024-07-15 11:27:55.443295] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:11.853 [2024-07-15 11:27:55.443333] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:12.788 11:27:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:17:12.788 11:27:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@862 -- # return 0 00:17:12.788 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i = 1 )) 00:17:12.788 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:17:12.788 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc1 00:17:12.788 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt1 00:17:12.788 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:17:12.788 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:17:12.788 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:17:12.789 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:17:12.789 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc1 00:17:12.789 malloc1 00:17:12.789 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:17:13.048 [2024-07-15 11:27:56.395193] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:17:13.048 [2024-07-15 11:27:56.395242] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:13.048 [2024-07-15 11:27:56.395263] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2114570 00:17:13.048 [2024-07-15 11:27:56.395275] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:13.048 [2024-07-15 11:27:56.396846] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:13.048 [2024-07-15 11:27:56.396876] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:17:13.048 pt1 00:17:13.048 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:17:13.048 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:17:13.048 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc2 00:17:13.048 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt2 00:17:13.048 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:17:13.048 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:17:13.048 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:17:13.048 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:17:13.048 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc2 00:17:13.048 malloc2 00:17:13.048 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:17:13.307 [2024-07-15 11:27:56.753043] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:17:13.307 [2024-07-15 11:27:56.753094] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:13.307 [2024-07-15 11:27:56.753112] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2115970 00:17:13.307 [2024-07-15 11:27:56.753125] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:13.307 [2024-07-15 11:27:56.754637] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:13.307 [2024-07-15 11:27:56.754670] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:17:13.307 pt2 00:17:13.307 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:17:13.307 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:17:13.307 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc3 00:17:13.307 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt3 00:17:13.307 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:17:13.307 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:17:13.307 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:17:13.307 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:17:13.307 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc3 00:17:13.566 malloc3 00:17:13.566 11:27:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:17:13.566 [2024-07-15 11:27:57.094738] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:17:13.566 [2024-07-15 11:27:57.094789] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:13.566 [2024-07-15 11:27:57.094808] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x22ac340 00:17:13.566 [2024-07-15 11:27:57.094821] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:13.566 [2024-07-15 11:27:57.096269] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:13.566 [2024-07-15 11:27:57.096298] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:17:13.566 pt3 00:17:13.566 11:27:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:17:13.566 11:27:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:17:13.566 11:27:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@429 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'pt1 pt2 pt3' -n raid_bdev1 -s 00:17:13.825 [2024-07-15 11:27:57.339408] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:17:13.825 [2024-07-15 11:27:57.340651] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:13.825 [2024-07-15 11:27:57.340706] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:17:13.825 [2024-07-15 11:27:57.340858] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x210cea0 00:17:13.825 [2024-07-15 11:27:57.340869] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:17:13.825 [2024-07-15 11:27:57.341080] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x2114240 00:17:13.825 [2024-07-15 11:27:57.341230] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x210cea0 00:17:13.825 [2024-07-15 11:27:57.341240] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x210cea0 00:17:13.825 [2024-07-15 11:27:57.341337] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:13.825 11:27:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:17:13.825 11:27:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:17:13.825 11:27:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:17:13.825 11:27:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:17:13.825 11:27:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:17:13.825 11:27:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:17:13.825 11:27:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:13.825 11:27:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:13.825 11:27:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:13.825 11:27:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:13.825 11:27:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:13.825 11:27:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:14.084 11:27:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:14.084 "name": "raid_bdev1", 00:17:14.084 "uuid": "356087b2-a385-4376-9bf5-a0fcd40d5fb1", 00:17:14.084 "strip_size_kb": 0, 00:17:14.084 "state": "online", 00:17:14.084 "raid_level": "raid1", 00:17:14.084 "superblock": true, 00:17:14.084 "num_base_bdevs": 3, 00:17:14.084 "num_base_bdevs_discovered": 3, 00:17:14.084 "num_base_bdevs_operational": 3, 00:17:14.084 "base_bdevs_list": [ 00:17:14.084 { 00:17:14.084 "name": "pt1", 00:17:14.084 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:14.084 "is_configured": true, 00:17:14.084 "data_offset": 2048, 00:17:14.084 "data_size": 63488 00:17:14.084 }, 00:17:14.084 { 00:17:14.084 "name": "pt2", 00:17:14.084 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:14.084 "is_configured": true, 00:17:14.084 "data_offset": 2048, 00:17:14.084 "data_size": 63488 00:17:14.084 }, 00:17:14.084 { 00:17:14.084 "name": "pt3", 00:17:14.084 "uuid": "00000000-0000-0000-0000-000000000003", 00:17:14.084 "is_configured": true, 00:17:14.084 "data_offset": 2048, 00:17:14.084 "data_size": 63488 00:17:14.084 } 00:17:14.084 ] 00:17:14.084 }' 00:17:14.084 11:27:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:14.084 11:27:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:17:14.653 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_properties raid_bdev1 00:17:14.653 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:17:14.653 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:17:14.653 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:17:14.653 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:17:14.653 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # local name 00:17:14.653 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:17:14.653 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:17:14.912 [2024-07-15 11:27:58.346306] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:14.912 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:17:14.912 "name": "raid_bdev1", 00:17:14.912 "aliases": [ 00:17:14.912 "356087b2-a385-4376-9bf5-a0fcd40d5fb1" 00:17:14.912 ], 00:17:14.912 "product_name": "Raid Volume", 00:17:14.912 "block_size": 512, 00:17:14.912 "num_blocks": 63488, 00:17:14.912 "uuid": "356087b2-a385-4376-9bf5-a0fcd40d5fb1", 00:17:14.912 "assigned_rate_limits": { 00:17:14.912 "rw_ios_per_sec": 0, 00:17:14.912 "rw_mbytes_per_sec": 0, 00:17:14.912 "r_mbytes_per_sec": 0, 00:17:14.912 "w_mbytes_per_sec": 0 00:17:14.912 }, 00:17:14.912 "claimed": false, 00:17:14.912 "zoned": false, 00:17:14.912 "supported_io_types": { 00:17:14.912 "read": true, 00:17:14.912 "write": true, 00:17:14.912 "unmap": false, 00:17:14.912 "flush": false, 00:17:14.912 "reset": true, 00:17:14.912 "nvme_admin": false, 00:17:14.912 "nvme_io": false, 00:17:14.912 "nvme_io_md": false, 00:17:14.912 "write_zeroes": true, 00:17:14.912 "zcopy": false, 00:17:14.912 "get_zone_info": false, 00:17:14.912 "zone_management": false, 00:17:14.912 "zone_append": false, 00:17:14.912 "compare": false, 00:17:14.912 "compare_and_write": false, 00:17:14.912 "abort": false, 00:17:14.912 "seek_hole": false, 00:17:14.912 "seek_data": false, 00:17:14.912 "copy": false, 00:17:14.912 "nvme_iov_md": false 00:17:14.912 }, 00:17:14.912 "memory_domains": [ 00:17:14.912 { 00:17:14.912 "dma_device_id": "system", 00:17:14.912 "dma_device_type": 1 00:17:14.912 }, 00:17:14.912 { 00:17:14.912 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:14.912 "dma_device_type": 2 00:17:14.912 }, 00:17:14.912 { 00:17:14.912 "dma_device_id": "system", 00:17:14.912 "dma_device_type": 1 00:17:14.912 }, 00:17:14.912 { 00:17:14.912 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:14.912 "dma_device_type": 2 00:17:14.912 }, 00:17:14.912 { 00:17:14.912 "dma_device_id": "system", 00:17:14.912 "dma_device_type": 1 00:17:14.912 }, 00:17:14.912 { 00:17:14.912 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:14.912 "dma_device_type": 2 00:17:14.912 } 00:17:14.912 ], 00:17:14.912 "driver_specific": { 00:17:14.912 "raid": { 00:17:14.912 "uuid": "356087b2-a385-4376-9bf5-a0fcd40d5fb1", 00:17:14.912 "strip_size_kb": 0, 00:17:14.912 "state": "online", 00:17:14.912 "raid_level": "raid1", 00:17:14.912 "superblock": true, 00:17:14.912 "num_base_bdevs": 3, 00:17:14.912 "num_base_bdevs_discovered": 3, 00:17:14.912 "num_base_bdevs_operational": 3, 00:17:14.912 "base_bdevs_list": [ 00:17:14.912 { 00:17:14.912 "name": "pt1", 00:17:14.912 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:14.912 "is_configured": true, 00:17:14.912 "data_offset": 2048, 00:17:14.912 "data_size": 63488 00:17:14.912 }, 00:17:14.912 { 00:17:14.912 "name": "pt2", 00:17:14.912 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:14.912 "is_configured": true, 00:17:14.912 "data_offset": 2048, 00:17:14.912 "data_size": 63488 00:17:14.912 }, 00:17:14.912 { 00:17:14.912 "name": "pt3", 00:17:14.912 "uuid": "00000000-0000-0000-0000-000000000003", 00:17:14.912 "is_configured": true, 00:17:14.912 "data_offset": 2048, 00:17:14.912 "data_size": 63488 00:17:14.912 } 00:17:14.912 ] 00:17:14.912 } 00:17:14.912 } 00:17:14.912 }' 00:17:14.912 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:17:14.912 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:17:14.912 pt2 00:17:14.912 pt3' 00:17:14.912 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:17:14.912 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:17:14.912 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:17:15.172 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:17:15.172 "name": "pt1", 00:17:15.172 "aliases": [ 00:17:15.172 "00000000-0000-0000-0000-000000000001" 00:17:15.172 ], 00:17:15.172 "product_name": "passthru", 00:17:15.172 "block_size": 512, 00:17:15.172 "num_blocks": 65536, 00:17:15.172 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:15.172 "assigned_rate_limits": { 00:17:15.172 "rw_ios_per_sec": 0, 00:17:15.172 "rw_mbytes_per_sec": 0, 00:17:15.172 "r_mbytes_per_sec": 0, 00:17:15.172 "w_mbytes_per_sec": 0 00:17:15.172 }, 00:17:15.172 "claimed": true, 00:17:15.172 "claim_type": "exclusive_write", 00:17:15.172 "zoned": false, 00:17:15.172 "supported_io_types": { 00:17:15.172 "read": true, 00:17:15.172 "write": true, 00:17:15.172 "unmap": true, 00:17:15.172 "flush": true, 00:17:15.172 "reset": true, 00:17:15.172 "nvme_admin": false, 00:17:15.172 "nvme_io": false, 00:17:15.172 "nvme_io_md": false, 00:17:15.172 "write_zeroes": true, 00:17:15.172 "zcopy": true, 00:17:15.172 "get_zone_info": false, 00:17:15.172 "zone_management": false, 00:17:15.172 "zone_append": false, 00:17:15.172 "compare": false, 00:17:15.172 "compare_and_write": false, 00:17:15.172 "abort": true, 00:17:15.172 "seek_hole": false, 00:17:15.172 "seek_data": false, 00:17:15.172 "copy": true, 00:17:15.172 "nvme_iov_md": false 00:17:15.172 }, 00:17:15.172 "memory_domains": [ 00:17:15.172 { 00:17:15.172 "dma_device_id": "system", 00:17:15.172 "dma_device_type": 1 00:17:15.172 }, 00:17:15.172 { 00:17:15.172 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:15.172 "dma_device_type": 2 00:17:15.172 } 00:17:15.172 ], 00:17:15.172 "driver_specific": { 00:17:15.172 "passthru": { 00:17:15.172 "name": "pt1", 00:17:15.172 "base_bdev_name": "malloc1" 00:17:15.172 } 00:17:15.172 } 00:17:15.172 }' 00:17:15.172 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:15.172 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:15.172 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:17:15.172 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:15.431 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:15.431 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:17:15.431 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:15.431 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:15.431 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:17:15.431 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:15.431 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:15.431 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:17:15.431 11:27:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:17:15.431 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:17:15.431 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:17:15.691 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:17:15.691 "name": "pt2", 00:17:15.691 "aliases": [ 00:17:15.691 "00000000-0000-0000-0000-000000000002" 00:17:15.691 ], 00:17:15.691 "product_name": "passthru", 00:17:15.691 "block_size": 512, 00:17:15.691 "num_blocks": 65536, 00:17:15.691 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:15.691 "assigned_rate_limits": { 00:17:15.691 "rw_ios_per_sec": 0, 00:17:15.691 "rw_mbytes_per_sec": 0, 00:17:15.691 "r_mbytes_per_sec": 0, 00:17:15.691 "w_mbytes_per_sec": 0 00:17:15.691 }, 00:17:15.691 "claimed": true, 00:17:15.691 "claim_type": "exclusive_write", 00:17:15.691 "zoned": false, 00:17:15.691 "supported_io_types": { 00:17:15.691 "read": true, 00:17:15.691 "write": true, 00:17:15.691 "unmap": true, 00:17:15.691 "flush": true, 00:17:15.691 "reset": true, 00:17:15.691 "nvme_admin": false, 00:17:15.691 "nvme_io": false, 00:17:15.691 "nvme_io_md": false, 00:17:15.691 "write_zeroes": true, 00:17:15.691 "zcopy": true, 00:17:15.691 "get_zone_info": false, 00:17:15.691 "zone_management": false, 00:17:15.691 "zone_append": false, 00:17:15.691 "compare": false, 00:17:15.691 "compare_and_write": false, 00:17:15.691 "abort": true, 00:17:15.691 "seek_hole": false, 00:17:15.691 "seek_data": false, 00:17:15.691 "copy": true, 00:17:15.691 "nvme_iov_md": false 00:17:15.691 }, 00:17:15.691 "memory_domains": [ 00:17:15.691 { 00:17:15.691 "dma_device_id": "system", 00:17:15.691 "dma_device_type": 1 00:17:15.691 }, 00:17:15.691 { 00:17:15.691 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:15.691 "dma_device_type": 2 00:17:15.691 } 00:17:15.691 ], 00:17:15.691 "driver_specific": { 00:17:15.691 "passthru": { 00:17:15.691 "name": "pt2", 00:17:15.691 "base_bdev_name": "malloc2" 00:17:15.691 } 00:17:15.691 } 00:17:15.691 }' 00:17:15.691 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:15.949 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:15.949 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:17:15.949 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:15.949 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:15.950 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:17:15.950 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:15.950 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:15.950 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:17:15.950 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:16.238 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:16.238 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:17:16.238 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:17:16.238 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt3 00:17:16.238 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:17:16.528 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:17:16.528 "name": "pt3", 00:17:16.528 "aliases": [ 00:17:16.528 "00000000-0000-0000-0000-000000000003" 00:17:16.528 ], 00:17:16.528 "product_name": "passthru", 00:17:16.528 "block_size": 512, 00:17:16.528 "num_blocks": 65536, 00:17:16.528 "uuid": "00000000-0000-0000-0000-000000000003", 00:17:16.528 "assigned_rate_limits": { 00:17:16.528 "rw_ios_per_sec": 0, 00:17:16.528 "rw_mbytes_per_sec": 0, 00:17:16.528 "r_mbytes_per_sec": 0, 00:17:16.528 "w_mbytes_per_sec": 0 00:17:16.528 }, 00:17:16.528 "claimed": true, 00:17:16.528 "claim_type": "exclusive_write", 00:17:16.528 "zoned": false, 00:17:16.528 "supported_io_types": { 00:17:16.528 "read": true, 00:17:16.528 "write": true, 00:17:16.528 "unmap": true, 00:17:16.528 "flush": true, 00:17:16.528 "reset": true, 00:17:16.528 "nvme_admin": false, 00:17:16.528 "nvme_io": false, 00:17:16.528 "nvme_io_md": false, 00:17:16.528 "write_zeroes": true, 00:17:16.528 "zcopy": true, 00:17:16.528 "get_zone_info": false, 00:17:16.528 "zone_management": false, 00:17:16.528 "zone_append": false, 00:17:16.528 "compare": false, 00:17:16.528 "compare_and_write": false, 00:17:16.528 "abort": true, 00:17:16.528 "seek_hole": false, 00:17:16.528 "seek_data": false, 00:17:16.528 "copy": true, 00:17:16.528 "nvme_iov_md": false 00:17:16.528 }, 00:17:16.528 "memory_domains": [ 00:17:16.528 { 00:17:16.528 "dma_device_id": "system", 00:17:16.528 "dma_device_type": 1 00:17:16.528 }, 00:17:16.528 { 00:17:16.528 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:16.528 "dma_device_type": 2 00:17:16.528 } 00:17:16.528 ], 00:17:16.528 "driver_specific": { 00:17:16.528 "passthru": { 00:17:16.528 "name": "pt3", 00:17:16.528 "base_bdev_name": "malloc3" 00:17:16.528 } 00:17:16.528 } 00:17:16.528 }' 00:17:16.528 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:16.528 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:16.528 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:17:16.528 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:16.528 11:27:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:16.528 11:28:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:17:16.528 11:28:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:16.528 11:28:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:16.787 11:28:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:17:16.787 11:28:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:16.787 11:28:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:16.787 11:28:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:17:16.787 11:28:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:17:16.787 11:28:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # jq -r '.[] | .uuid' 00:17:17.045 [2024-07-15 11:28:00.427840] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:17.045 11:28:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # raid_bdev_uuid=356087b2-a385-4376-9bf5-a0fcd40d5fb1 00:17:17.045 11:28:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # '[' -z 356087b2-a385-4376-9bf5-a0fcd40d5fb1 ']' 00:17:17.045 11:28:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@440 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:17:17.614 [2024-07-15 11:28:00.932902] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:17.614 [2024-07-15 11:28:00.932933] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:17.614 [2024-07-15 11:28:00.932984] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:17.614 [2024-07-15 11:28:00.933054] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:17.614 [2024-07-15 11:28:00.933066] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x210cea0 name raid_bdev1, state offline 00:17:17.614 11:28:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:17.614 11:28:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # jq -r '.[]' 00:17:17.873 11:28:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # raid_bdev= 00:17:17.873 11:28:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # '[' -n '' ']' 00:17:17.873 11:28:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:17:17.873 11:28:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt1 00:17:17.873 11:28:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:17:17.873 11:28:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:17:18.133 11:28:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:17:18.133 11:28:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt3 00:17:18.392 11:28:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs 00:17:18.392 11:28:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:17:18.651 11:28:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # '[' false == true ']' 00:17:18.651 11:28:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@456 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'malloc1 malloc2 malloc3' -n raid_bdev1 00:17:18.651 11:28:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@648 -- # local es=0 00:17:18.651 11:28:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'malloc1 malloc2 malloc3' -n raid_bdev1 00:17:18.651 11:28:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:17:18.651 11:28:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:17:18.651 11:28:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:17:18.651 11:28:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:17:18.651 11:28:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:17:18.651 11:28:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:17:18.651 11:28:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:17:18.651 11:28:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py ]] 00:17:18.651 11:28:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'malloc1 malloc2 malloc3' -n raid_bdev1 00:17:18.911 [2024-07-15 11:28:02.400720] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:17:18.911 [2024-07-15 11:28:02.402077] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:17:18.911 [2024-07-15 11:28:02.402121] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:17:18.911 [2024-07-15 11:28:02.402165] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:17:18.911 [2024-07-15 11:28:02.402207] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:17:18.911 [2024-07-15 11:28:02.402230] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:17:18.911 [2024-07-15 11:28:02.402255] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:18.911 [2024-07-15 11:28:02.402265] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x22b7ff0 name raid_bdev1, state configuring 00:17:18.911 request: 00:17:18.911 { 00:17:18.911 "name": "raid_bdev1", 00:17:18.911 "raid_level": "raid1", 00:17:18.911 "base_bdevs": [ 00:17:18.911 "malloc1", 00:17:18.911 "malloc2", 00:17:18.911 "malloc3" 00:17:18.911 ], 00:17:18.911 "superblock": false, 00:17:18.911 "method": "bdev_raid_create", 00:17:18.911 "req_id": 1 00:17:18.911 } 00:17:18.911 Got JSON-RPC error response 00:17:18.911 response: 00:17:18.911 { 00:17:18.911 "code": -17, 00:17:18.911 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:17:18.911 } 00:17:18.911 11:28:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@651 -- # es=1 00:17:18.911 11:28:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:17:18.911 11:28:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:17:18.911 11:28:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:17:18.911 11:28:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:18.911 11:28:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # jq -r '.[]' 00:17:19.170 11:28:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # raid_bdev= 00:17:19.170 11:28:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # '[' -n '' ']' 00:17:19.170 11:28:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@464 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:17:19.429 [2024-07-15 11:28:02.893966] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:17:19.429 [2024-07-15 11:28:02.894009] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:19.429 [2024-07-15 11:28:02.894031] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x21147a0 00:17:19.429 [2024-07-15 11:28:02.894043] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:19.429 [2024-07-15 11:28:02.895617] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:19.429 [2024-07-15 11:28:02.895646] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:17:19.429 [2024-07-15 11:28:02.895712] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:17:19.429 [2024-07-15 11:28:02.895739] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:17:19.429 pt1 00:17:19.429 11:28:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@467 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:17:19.429 11:28:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:17:19.429 11:28:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:17:19.429 11:28:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:17:19.429 11:28:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:17:19.429 11:28:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:17:19.429 11:28:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:19.429 11:28:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:19.429 11:28:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:19.429 11:28:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:19.429 11:28:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:19.429 11:28:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:19.688 11:28:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:19.688 "name": "raid_bdev1", 00:17:19.688 "uuid": "356087b2-a385-4376-9bf5-a0fcd40d5fb1", 00:17:19.688 "strip_size_kb": 0, 00:17:19.688 "state": "configuring", 00:17:19.688 "raid_level": "raid1", 00:17:19.688 "superblock": true, 00:17:19.688 "num_base_bdevs": 3, 00:17:19.688 "num_base_bdevs_discovered": 1, 00:17:19.688 "num_base_bdevs_operational": 3, 00:17:19.688 "base_bdevs_list": [ 00:17:19.688 { 00:17:19.688 "name": "pt1", 00:17:19.688 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:19.688 "is_configured": true, 00:17:19.688 "data_offset": 2048, 00:17:19.688 "data_size": 63488 00:17:19.688 }, 00:17:19.688 { 00:17:19.688 "name": null, 00:17:19.688 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:19.688 "is_configured": false, 00:17:19.688 "data_offset": 2048, 00:17:19.688 "data_size": 63488 00:17:19.688 }, 00:17:19.688 { 00:17:19.688 "name": null, 00:17:19.688 "uuid": "00000000-0000-0000-0000-000000000003", 00:17:19.688 "is_configured": false, 00:17:19.688 "data_offset": 2048, 00:17:19.688 "data_size": 63488 00:17:19.688 } 00:17:19.688 ] 00:17:19.688 }' 00:17:19.688 11:28:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:19.688 11:28:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:17:20.255 11:28:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@469 -- # '[' 3 -gt 2 ']' 00:17:20.255 11:28:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@471 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:17:20.514 [2024-07-15 11:28:03.960790] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:17:20.514 [2024-07-15 11:28:03.960841] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:20.514 [2024-07-15 11:28:03.960866] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x210ba10 00:17:20.514 [2024-07-15 11:28:03.960886] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:20.514 [2024-07-15 11:28:03.961246] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:20.514 [2024-07-15 11:28:03.961265] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:17:20.514 [2024-07-15 11:28:03.961328] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:17:20.514 [2024-07-15 11:28:03.961347] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:20.514 pt2 00:17:20.514 11:28:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:17:20.772 [2024-07-15 11:28:04.205450] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:17:20.772 11:28:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:17:20.772 11:28:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:17:20.772 11:28:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:17:20.772 11:28:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:17:20.772 11:28:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:17:20.772 11:28:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:17:20.772 11:28:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:20.772 11:28:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:20.772 11:28:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:20.772 11:28:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:20.772 11:28:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:20.772 11:28:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:21.031 11:28:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:21.031 "name": "raid_bdev1", 00:17:21.031 "uuid": "356087b2-a385-4376-9bf5-a0fcd40d5fb1", 00:17:21.031 "strip_size_kb": 0, 00:17:21.031 "state": "configuring", 00:17:21.031 "raid_level": "raid1", 00:17:21.031 "superblock": true, 00:17:21.031 "num_base_bdevs": 3, 00:17:21.031 "num_base_bdevs_discovered": 1, 00:17:21.031 "num_base_bdevs_operational": 3, 00:17:21.031 "base_bdevs_list": [ 00:17:21.031 { 00:17:21.031 "name": "pt1", 00:17:21.031 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:21.031 "is_configured": true, 00:17:21.031 "data_offset": 2048, 00:17:21.031 "data_size": 63488 00:17:21.031 }, 00:17:21.031 { 00:17:21.031 "name": null, 00:17:21.031 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:21.031 "is_configured": false, 00:17:21.031 "data_offset": 2048, 00:17:21.031 "data_size": 63488 00:17:21.031 }, 00:17:21.031 { 00:17:21.031 "name": null, 00:17:21.031 "uuid": "00000000-0000-0000-0000-000000000003", 00:17:21.031 "is_configured": false, 00:17:21.031 "data_offset": 2048, 00:17:21.031 "data_size": 63488 00:17:21.031 } 00:17:21.031 ] 00:17:21.031 }' 00:17:21.031 11:28:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:21.031 11:28:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:17:21.598 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i = 1 )) 00:17:21.598 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:17:21.598 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:17:21.856 [2024-07-15 11:28:05.248206] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:17:21.856 [2024-07-15 11:28:05.248254] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:21.856 [2024-07-15 11:28:05.248277] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2114a10 00:17:21.856 [2024-07-15 11:28:05.248290] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:21.856 [2024-07-15 11:28:05.248642] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:21.856 [2024-07-15 11:28:05.248662] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:17:21.856 [2024-07-15 11:28:05.248724] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:17:21.856 [2024-07-15 11:28:05.248743] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:21.856 pt2 00:17:21.856 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:17:21.856 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:17:21.856 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:17:21.856 [2024-07-15 11:28:05.424674] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:17:21.856 [2024-07-15 11:28:05.424711] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:21.856 [2024-07-15 11:28:05.424727] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x210b6c0 00:17:21.856 [2024-07-15 11:28:05.424739] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:21.856 [2024-07-15 11:28:05.425045] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:21.856 [2024-07-15 11:28:05.425064] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:17:21.856 [2024-07-15 11:28:05.425116] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:17:21.856 [2024-07-15 11:28:05.425133] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:17:21.856 [2024-07-15 11:28:05.425237] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x22aec00 00:17:21.856 [2024-07-15 11:28:05.425248] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:17:21.856 [2024-07-15 11:28:05.425416] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x210e610 00:17:21.856 [2024-07-15 11:28:05.425545] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x22aec00 00:17:21.856 [2024-07-15 11:28:05.425555] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x22aec00 00:17:21.856 [2024-07-15 11:28:05.425654] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:21.856 pt3 00:17:21.856 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:17:21.856 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:17:21.856 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@482 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:17:21.856 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:17:21.856 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:17:21.856 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:17:21.856 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:17:21.857 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:17:21.857 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:21.857 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:21.857 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:21.857 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:22.115 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:22.115 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:22.115 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:22.115 "name": "raid_bdev1", 00:17:22.115 "uuid": "356087b2-a385-4376-9bf5-a0fcd40d5fb1", 00:17:22.115 "strip_size_kb": 0, 00:17:22.115 "state": "online", 00:17:22.115 "raid_level": "raid1", 00:17:22.115 "superblock": true, 00:17:22.115 "num_base_bdevs": 3, 00:17:22.115 "num_base_bdevs_discovered": 3, 00:17:22.115 "num_base_bdevs_operational": 3, 00:17:22.115 "base_bdevs_list": [ 00:17:22.115 { 00:17:22.115 "name": "pt1", 00:17:22.115 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:22.115 "is_configured": true, 00:17:22.115 "data_offset": 2048, 00:17:22.115 "data_size": 63488 00:17:22.115 }, 00:17:22.115 { 00:17:22.115 "name": "pt2", 00:17:22.115 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:22.115 "is_configured": true, 00:17:22.115 "data_offset": 2048, 00:17:22.115 "data_size": 63488 00:17:22.115 }, 00:17:22.115 { 00:17:22.115 "name": "pt3", 00:17:22.115 "uuid": "00000000-0000-0000-0000-000000000003", 00:17:22.115 "is_configured": true, 00:17:22.115 "data_offset": 2048, 00:17:22.115 "data_size": 63488 00:17:22.115 } 00:17:22.115 ] 00:17:22.115 }' 00:17:22.115 11:28:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:22.115 11:28:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:17:22.682 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_properties raid_bdev1 00:17:22.682 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:17:22.682 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:17:22.682 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:17:22.682 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:17:22.682 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # local name 00:17:22.682 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:17:22.682 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:17:22.942 [2024-07-15 11:28:06.371460] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:22.942 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:17:22.942 "name": "raid_bdev1", 00:17:22.942 "aliases": [ 00:17:22.942 "356087b2-a385-4376-9bf5-a0fcd40d5fb1" 00:17:22.942 ], 00:17:22.942 "product_name": "Raid Volume", 00:17:22.942 "block_size": 512, 00:17:22.942 "num_blocks": 63488, 00:17:22.942 "uuid": "356087b2-a385-4376-9bf5-a0fcd40d5fb1", 00:17:22.942 "assigned_rate_limits": { 00:17:22.942 "rw_ios_per_sec": 0, 00:17:22.942 "rw_mbytes_per_sec": 0, 00:17:22.942 "r_mbytes_per_sec": 0, 00:17:22.942 "w_mbytes_per_sec": 0 00:17:22.942 }, 00:17:22.942 "claimed": false, 00:17:22.942 "zoned": false, 00:17:22.942 "supported_io_types": { 00:17:22.942 "read": true, 00:17:22.942 "write": true, 00:17:22.942 "unmap": false, 00:17:22.942 "flush": false, 00:17:22.942 "reset": true, 00:17:22.942 "nvme_admin": false, 00:17:22.942 "nvme_io": false, 00:17:22.942 "nvme_io_md": false, 00:17:22.942 "write_zeroes": true, 00:17:22.942 "zcopy": false, 00:17:22.942 "get_zone_info": false, 00:17:22.942 "zone_management": false, 00:17:22.942 "zone_append": false, 00:17:22.942 "compare": false, 00:17:22.942 "compare_and_write": false, 00:17:22.942 "abort": false, 00:17:22.942 "seek_hole": false, 00:17:22.942 "seek_data": false, 00:17:22.942 "copy": false, 00:17:22.942 "nvme_iov_md": false 00:17:22.942 }, 00:17:22.942 "memory_domains": [ 00:17:22.942 { 00:17:22.942 "dma_device_id": "system", 00:17:22.942 "dma_device_type": 1 00:17:22.942 }, 00:17:22.942 { 00:17:22.942 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:22.942 "dma_device_type": 2 00:17:22.942 }, 00:17:22.942 { 00:17:22.942 "dma_device_id": "system", 00:17:22.942 "dma_device_type": 1 00:17:22.942 }, 00:17:22.942 { 00:17:22.942 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:22.942 "dma_device_type": 2 00:17:22.942 }, 00:17:22.942 { 00:17:22.942 "dma_device_id": "system", 00:17:22.942 "dma_device_type": 1 00:17:22.942 }, 00:17:22.942 { 00:17:22.942 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:22.942 "dma_device_type": 2 00:17:22.942 } 00:17:22.942 ], 00:17:22.942 "driver_specific": { 00:17:22.942 "raid": { 00:17:22.942 "uuid": "356087b2-a385-4376-9bf5-a0fcd40d5fb1", 00:17:22.942 "strip_size_kb": 0, 00:17:22.942 "state": "online", 00:17:22.942 "raid_level": "raid1", 00:17:22.942 "superblock": true, 00:17:22.942 "num_base_bdevs": 3, 00:17:22.942 "num_base_bdevs_discovered": 3, 00:17:22.942 "num_base_bdevs_operational": 3, 00:17:22.942 "base_bdevs_list": [ 00:17:22.942 { 00:17:22.942 "name": "pt1", 00:17:22.942 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:22.942 "is_configured": true, 00:17:22.942 "data_offset": 2048, 00:17:22.942 "data_size": 63488 00:17:22.942 }, 00:17:22.942 { 00:17:22.942 "name": "pt2", 00:17:22.942 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:22.942 "is_configured": true, 00:17:22.942 "data_offset": 2048, 00:17:22.942 "data_size": 63488 00:17:22.942 }, 00:17:22.942 { 00:17:22.942 "name": "pt3", 00:17:22.942 "uuid": "00000000-0000-0000-0000-000000000003", 00:17:22.942 "is_configured": true, 00:17:22.942 "data_offset": 2048, 00:17:22.942 "data_size": 63488 00:17:22.942 } 00:17:22.942 ] 00:17:22.942 } 00:17:22.942 } 00:17:22.942 }' 00:17:22.942 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:17:22.942 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:17:22.942 pt2 00:17:22.942 pt3' 00:17:22.942 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:17:22.942 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:17:22.942 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:17:23.201 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:17:23.201 "name": "pt1", 00:17:23.201 "aliases": [ 00:17:23.201 "00000000-0000-0000-0000-000000000001" 00:17:23.201 ], 00:17:23.201 "product_name": "passthru", 00:17:23.201 "block_size": 512, 00:17:23.201 "num_blocks": 65536, 00:17:23.201 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:23.201 "assigned_rate_limits": { 00:17:23.201 "rw_ios_per_sec": 0, 00:17:23.201 "rw_mbytes_per_sec": 0, 00:17:23.201 "r_mbytes_per_sec": 0, 00:17:23.201 "w_mbytes_per_sec": 0 00:17:23.201 }, 00:17:23.201 "claimed": true, 00:17:23.201 "claim_type": "exclusive_write", 00:17:23.201 "zoned": false, 00:17:23.201 "supported_io_types": { 00:17:23.201 "read": true, 00:17:23.201 "write": true, 00:17:23.201 "unmap": true, 00:17:23.201 "flush": true, 00:17:23.201 "reset": true, 00:17:23.201 "nvme_admin": false, 00:17:23.201 "nvme_io": false, 00:17:23.201 "nvme_io_md": false, 00:17:23.201 "write_zeroes": true, 00:17:23.201 "zcopy": true, 00:17:23.201 "get_zone_info": false, 00:17:23.201 "zone_management": false, 00:17:23.201 "zone_append": false, 00:17:23.201 "compare": false, 00:17:23.201 "compare_and_write": false, 00:17:23.201 "abort": true, 00:17:23.201 "seek_hole": false, 00:17:23.201 "seek_data": false, 00:17:23.201 "copy": true, 00:17:23.201 "nvme_iov_md": false 00:17:23.201 }, 00:17:23.201 "memory_domains": [ 00:17:23.201 { 00:17:23.201 "dma_device_id": "system", 00:17:23.201 "dma_device_type": 1 00:17:23.201 }, 00:17:23.201 { 00:17:23.201 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:23.201 "dma_device_type": 2 00:17:23.201 } 00:17:23.201 ], 00:17:23.201 "driver_specific": { 00:17:23.201 "passthru": { 00:17:23.201 "name": "pt1", 00:17:23.201 "base_bdev_name": "malloc1" 00:17:23.201 } 00:17:23.201 } 00:17:23.201 }' 00:17:23.201 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:23.201 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:23.201 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:17:23.201 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:23.201 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:23.461 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:17:23.461 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:23.461 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:23.461 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:17:23.461 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:23.461 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:23.461 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:17:23.461 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:17:23.461 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:17:23.461 11:28:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:17:23.720 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:17:23.720 "name": "pt2", 00:17:23.720 "aliases": [ 00:17:23.720 "00000000-0000-0000-0000-000000000002" 00:17:23.720 ], 00:17:23.720 "product_name": "passthru", 00:17:23.720 "block_size": 512, 00:17:23.720 "num_blocks": 65536, 00:17:23.720 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:23.720 "assigned_rate_limits": { 00:17:23.720 "rw_ios_per_sec": 0, 00:17:23.720 "rw_mbytes_per_sec": 0, 00:17:23.720 "r_mbytes_per_sec": 0, 00:17:23.720 "w_mbytes_per_sec": 0 00:17:23.720 }, 00:17:23.720 "claimed": true, 00:17:23.720 "claim_type": "exclusive_write", 00:17:23.720 "zoned": false, 00:17:23.720 "supported_io_types": { 00:17:23.720 "read": true, 00:17:23.720 "write": true, 00:17:23.720 "unmap": true, 00:17:23.720 "flush": true, 00:17:23.720 "reset": true, 00:17:23.720 "nvme_admin": false, 00:17:23.720 "nvme_io": false, 00:17:23.720 "nvme_io_md": false, 00:17:23.720 "write_zeroes": true, 00:17:23.720 "zcopy": true, 00:17:23.720 "get_zone_info": false, 00:17:23.720 "zone_management": false, 00:17:23.721 "zone_append": false, 00:17:23.721 "compare": false, 00:17:23.721 "compare_and_write": false, 00:17:23.721 "abort": true, 00:17:23.721 "seek_hole": false, 00:17:23.721 "seek_data": false, 00:17:23.721 "copy": true, 00:17:23.721 "nvme_iov_md": false 00:17:23.721 }, 00:17:23.721 "memory_domains": [ 00:17:23.721 { 00:17:23.721 "dma_device_id": "system", 00:17:23.721 "dma_device_type": 1 00:17:23.721 }, 00:17:23.721 { 00:17:23.721 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:23.721 "dma_device_type": 2 00:17:23.721 } 00:17:23.721 ], 00:17:23.721 "driver_specific": { 00:17:23.721 "passthru": { 00:17:23.721 "name": "pt2", 00:17:23.721 "base_bdev_name": "malloc2" 00:17:23.721 } 00:17:23.721 } 00:17:23.721 }' 00:17:23.721 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:23.721 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:23.721 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:17:23.721 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:23.980 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:23.980 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:17:23.980 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:23.980 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:23.980 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:17:23.980 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:23.980 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:23.980 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:17:23.980 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:17:23.980 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt3 00:17:23.980 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:17:24.239 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:17:24.239 "name": "pt3", 00:17:24.239 "aliases": [ 00:17:24.239 "00000000-0000-0000-0000-000000000003" 00:17:24.239 ], 00:17:24.239 "product_name": "passthru", 00:17:24.239 "block_size": 512, 00:17:24.239 "num_blocks": 65536, 00:17:24.239 "uuid": "00000000-0000-0000-0000-000000000003", 00:17:24.239 "assigned_rate_limits": { 00:17:24.239 "rw_ios_per_sec": 0, 00:17:24.239 "rw_mbytes_per_sec": 0, 00:17:24.239 "r_mbytes_per_sec": 0, 00:17:24.239 "w_mbytes_per_sec": 0 00:17:24.239 }, 00:17:24.239 "claimed": true, 00:17:24.239 "claim_type": "exclusive_write", 00:17:24.239 "zoned": false, 00:17:24.239 "supported_io_types": { 00:17:24.239 "read": true, 00:17:24.239 "write": true, 00:17:24.239 "unmap": true, 00:17:24.239 "flush": true, 00:17:24.239 "reset": true, 00:17:24.239 "nvme_admin": false, 00:17:24.239 "nvme_io": false, 00:17:24.239 "nvme_io_md": false, 00:17:24.239 "write_zeroes": true, 00:17:24.239 "zcopy": true, 00:17:24.239 "get_zone_info": false, 00:17:24.239 "zone_management": false, 00:17:24.239 "zone_append": false, 00:17:24.239 "compare": false, 00:17:24.239 "compare_and_write": false, 00:17:24.239 "abort": true, 00:17:24.239 "seek_hole": false, 00:17:24.239 "seek_data": false, 00:17:24.239 "copy": true, 00:17:24.239 "nvme_iov_md": false 00:17:24.239 }, 00:17:24.239 "memory_domains": [ 00:17:24.239 { 00:17:24.239 "dma_device_id": "system", 00:17:24.239 "dma_device_type": 1 00:17:24.239 }, 00:17:24.239 { 00:17:24.239 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:24.239 "dma_device_type": 2 00:17:24.239 } 00:17:24.239 ], 00:17:24.239 "driver_specific": { 00:17:24.239 "passthru": { 00:17:24.239 "name": "pt3", 00:17:24.239 "base_bdev_name": "malloc3" 00:17:24.239 } 00:17:24.239 } 00:17:24.239 }' 00:17:24.239 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:24.239 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:24.497 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:17:24.497 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:24.497 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:24.497 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:17:24.497 11:28:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:24.497 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:24.497 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:17:24.497 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:24.755 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:24.755 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:17:24.755 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:17:24.755 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # jq -r '.[] | .uuid' 00:17:25.013 [2024-07-15 11:28:08.368871] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:25.013 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # '[' 356087b2-a385-4376-9bf5-a0fcd40d5fb1 '!=' 356087b2-a385-4376-9bf5-a0fcd40d5fb1 ']' 00:17:25.013 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@490 -- # has_redundancy raid1 00:17:25.013 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:17:25.013 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@214 -- # return 0 00:17:25.013 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@492 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt1 00:17:25.271 [2024-07-15 11:28:08.617287] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:17:25.271 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@495 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:25.271 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:17:25.271 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:17:25.271 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:17:25.271 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:17:25.271 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:17:25.271 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:25.271 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:25.271 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:25.272 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:25.272 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:25.272 11:28:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:25.838 11:28:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:25.838 "name": "raid_bdev1", 00:17:25.838 "uuid": "356087b2-a385-4376-9bf5-a0fcd40d5fb1", 00:17:25.838 "strip_size_kb": 0, 00:17:25.838 "state": "online", 00:17:25.838 "raid_level": "raid1", 00:17:25.838 "superblock": true, 00:17:25.838 "num_base_bdevs": 3, 00:17:25.838 "num_base_bdevs_discovered": 2, 00:17:25.838 "num_base_bdevs_operational": 2, 00:17:25.838 "base_bdevs_list": [ 00:17:25.838 { 00:17:25.838 "name": null, 00:17:25.838 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:25.838 "is_configured": false, 00:17:25.838 "data_offset": 2048, 00:17:25.838 "data_size": 63488 00:17:25.838 }, 00:17:25.838 { 00:17:25.838 "name": "pt2", 00:17:25.838 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:25.838 "is_configured": true, 00:17:25.838 "data_offset": 2048, 00:17:25.838 "data_size": 63488 00:17:25.838 }, 00:17:25.838 { 00:17:25.838 "name": "pt3", 00:17:25.838 "uuid": "00000000-0000-0000-0000-000000000003", 00:17:25.838 "is_configured": true, 00:17:25.838 "data_offset": 2048, 00:17:25.838 "data_size": 63488 00:17:25.838 } 00:17:25.838 ] 00:17:25.838 }' 00:17:25.838 11:28:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:25.838 11:28:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:17:26.404 11:28:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@498 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:17:26.404 [2024-07-15 11:28:09.960833] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:26.404 [2024-07-15 11:28:09.960864] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:26.404 [2024-07-15 11:28:09.960919] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:26.404 [2024-07-15 11:28:09.960979] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:26.404 [2024-07-15 11:28:09.960991] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x22aec00 name raid_bdev1, state offline 00:17:26.404 11:28:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:26.404 11:28:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # jq -r '.[]' 00:17:26.663 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # raid_bdev= 00:17:26.663 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # '[' -n '' ']' 00:17:26.663 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@505 -- # (( i = 1 )) 00:17:26.663 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@505 -- # (( i < num_base_bdevs )) 00:17:26.663 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:17:26.921 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@505 -- # (( i++ )) 00:17:26.921 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@505 -- # (( i < num_base_bdevs )) 00:17:26.921 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt3 00:17:27.178 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@505 -- # (( i++ )) 00:17:27.178 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@505 -- # (( i < num_base_bdevs )) 00:17:27.178 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@510 -- # (( i = 1 )) 00:17:27.178 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@510 -- # (( i < num_base_bdevs - 1 )) 00:17:27.178 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:17:27.436 [2024-07-15 11:28:10.951414] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:17:27.436 [2024-07-15 11:28:10.951461] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:27.436 [2024-07-15 11:28:10.951478] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x210c310 00:17:27.436 [2024-07-15 11:28:10.951491] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:27.436 [2024-07-15 11:28:10.953104] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:27.436 [2024-07-15 11:28:10.953135] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:17:27.436 [2024-07-15 11:28:10.953205] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:17:27.436 [2024-07-15 11:28:10.953233] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:27.436 pt2 00:17:27.436 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@514 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:17:27.436 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:17:27.436 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:17:27.436 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:17:27.436 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:17:27.436 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:17:27.436 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:27.436 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:27.436 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:27.436 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:27.436 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:27.436 11:28:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:27.694 11:28:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:27.694 "name": "raid_bdev1", 00:17:27.694 "uuid": "356087b2-a385-4376-9bf5-a0fcd40d5fb1", 00:17:27.694 "strip_size_kb": 0, 00:17:27.694 "state": "configuring", 00:17:27.694 "raid_level": "raid1", 00:17:27.694 "superblock": true, 00:17:27.694 "num_base_bdevs": 3, 00:17:27.694 "num_base_bdevs_discovered": 1, 00:17:27.694 "num_base_bdevs_operational": 2, 00:17:27.694 "base_bdevs_list": [ 00:17:27.694 { 00:17:27.694 "name": null, 00:17:27.694 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:27.694 "is_configured": false, 00:17:27.694 "data_offset": 2048, 00:17:27.694 "data_size": 63488 00:17:27.694 }, 00:17:27.694 { 00:17:27.694 "name": "pt2", 00:17:27.694 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:27.694 "is_configured": true, 00:17:27.695 "data_offset": 2048, 00:17:27.695 "data_size": 63488 00:17:27.695 }, 00:17:27.695 { 00:17:27.695 "name": null, 00:17:27.695 "uuid": "00000000-0000-0000-0000-000000000003", 00:17:27.695 "is_configured": false, 00:17:27.695 "data_offset": 2048, 00:17:27.695 "data_size": 63488 00:17:27.695 } 00:17:27.695 ] 00:17:27.695 }' 00:17:27.695 11:28:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:27.695 11:28:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:17:28.260 11:28:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@510 -- # (( i++ )) 00:17:28.260 11:28:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@510 -- # (( i < num_base_bdevs - 1 )) 00:17:28.260 11:28:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@518 -- # i=2 00:17:28.260 11:28:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:17:28.518 [2024-07-15 11:28:12.050437] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:17:28.518 [2024-07-15 11:28:12.050492] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:28.518 [2024-07-15 11:28:12.050512] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x210aec0 00:17:28.518 [2024-07-15 11:28:12.050525] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:28.518 [2024-07-15 11:28:12.050871] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:28.518 [2024-07-15 11:28:12.050890] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:17:28.518 [2024-07-15 11:28:12.050962] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:17:28.518 [2024-07-15 11:28:12.050984] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:17:28.518 [2024-07-15 11:28:12.051085] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x22accc0 00:17:28.518 [2024-07-15 11:28:12.051096] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:17:28.518 [2024-07-15 11:28:12.051265] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x22ad6d0 00:17:28.518 [2024-07-15 11:28:12.051391] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x22accc0 00:17:28.518 [2024-07-15 11:28:12.051401] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x22accc0 00:17:28.518 [2024-07-15 11:28:12.051499] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:28.518 pt3 00:17:28.518 11:28:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@522 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:28.518 11:28:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:17:28.518 11:28:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:17:28.518 11:28:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:17:28.518 11:28:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:17:28.518 11:28:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:17:28.518 11:28:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:28.518 11:28:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:28.518 11:28:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:28.518 11:28:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:28.518 11:28:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:28.518 11:28:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:28.776 11:28:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:28.776 "name": "raid_bdev1", 00:17:28.776 "uuid": "356087b2-a385-4376-9bf5-a0fcd40d5fb1", 00:17:28.776 "strip_size_kb": 0, 00:17:28.776 "state": "online", 00:17:28.776 "raid_level": "raid1", 00:17:28.776 "superblock": true, 00:17:28.776 "num_base_bdevs": 3, 00:17:28.776 "num_base_bdevs_discovered": 2, 00:17:28.776 "num_base_bdevs_operational": 2, 00:17:28.776 "base_bdevs_list": [ 00:17:28.776 { 00:17:28.776 "name": null, 00:17:28.776 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:28.776 "is_configured": false, 00:17:28.776 "data_offset": 2048, 00:17:28.776 "data_size": 63488 00:17:28.776 }, 00:17:28.776 { 00:17:28.776 "name": "pt2", 00:17:28.776 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:28.776 "is_configured": true, 00:17:28.776 "data_offset": 2048, 00:17:28.776 "data_size": 63488 00:17:28.776 }, 00:17:28.776 { 00:17:28.776 "name": "pt3", 00:17:28.776 "uuid": "00000000-0000-0000-0000-000000000003", 00:17:28.776 "is_configured": true, 00:17:28.776 "data_offset": 2048, 00:17:28.776 "data_size": 63488 00:17:28.776 } 00:17:28.776 ] 00:17:28.776 }' 00:17:28.776 11:28:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:28.776 11:28:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:17:29.339 11:28:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@525 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:17:29.596 [2024-07-15 11:28:13.133301] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:29.596 [2024-07-15 11:28:13.133328] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:29.596 [2024-07-15 11:28:13.133380] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:29.596 [2024-07-15 11:28:13.133435] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:29.596 [2024-07-15 11:28:13.133447] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x22accc0 name raid_bdev1, state offline 00:17:29.596 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:29.596 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # jq -r '.[]' 00:17:29.853 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # raid_bdev= 00:17:29.853 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # '[' -n '' ']' 00:17:29.853 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@531 -- # '[' 3 -gt 2 ']' 00:17:29.853 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@533 -- # i=2 00:17:29.853 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt3 00:17:30.111 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@539 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:17:30.371 [2024-07-15 11:28:13.875235] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:17:30.371 [2024-07-15 11:28:13.875278] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:30.371 [2024-07-15 11:28:13.875294] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x210aec0 00:17:30.371 [2024-07-15 11:28:13.875307] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:30.371 [2024-07-15 11:28:13.876908] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:30.371 [2024-07-15 11:28:13.876945] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:17:30.371 [2024-07-15 11:28:13.877011] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:17:30.371 [2024-07-15 11:28:13.877044] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:17:30.371 [2024-07-15 11:28:13.877143] bdev_raid.c:3547:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:17:30.371 [2024-07-15 11:28:13.877156] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:30.371 [2024-07-15 11:28:13.877171] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x22acf40 name raid_bdev1, state configuring 00:17:30.371 [2024-07-15 11:28:13.877194] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:30.371 pt1 00:17:30.371 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@541 -- # '[' 3 -gt 2 ']' 00:17:30.371 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@544 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:17:30.371 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:17:30.371 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:17:30.371 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:17:30.371 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:17:30.371 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:17:30.371 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:30.371 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:30.371 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:30.371 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:30.371 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:30.371 11:28:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:30.646 11:28:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:30.646 "name": "raid_bdev1", 00:17:30.646 "uuid": "356087b2-a385-4376-9bf5-a0fcd40d5fb1", 00:17:30.646 "strip_size_kb": 0, 00:17:30.646 "state": "configuring", 00:17:30.646 "raid_level": "raid1", 00:17:30.646 "superblock": true, 00:17:30.646 "num_base_bdevs": 3, 00:17:30.646 "num_base_bdevs_discovered": 1, 00:17:30.646 "num_base_bdevs_operational": 2, 00:17:30.646 "base_bdevs_list": [ 00:17:30.646 { 00:17:30.646 "name": null, 00:17:30.646 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:30.646 "is_configured": false, 00:17:30.646 "data_offset": 2048, 00:17:30.646 "data_size": 63488 00:17:30.646 }, 00:17:30.646 { 00:17:30.646 "name": "pt2", 00:17:30.646 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:30.646 "is_configured": true, 00:17:30.646 "data_offset": 2048, 00:17:30.646 "data_size": 63488 00:17:30.646 }, 00:17:30.646 { 00:17:30.646 "name": null, 00:17:30.646 "uuid": "00000000-0000-0000-0000-000000000003", 00:17:30.646 "is_configured": false, 00:17:30.646 "data_offset": 2048, 00:17:30.646 "data_size": 63488 00:17:30.647 } 00:17:30.647 ] 00:17:30.647 }' 00:17:30.647 11:28:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:30.647 11:28:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:17:31.225 11:28:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs configuring 00:17:31.225 11:28:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:17:31.482 11:28:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # [[ false == \f\a\l\s\e ]] 00:17:31.482 11:28:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@548 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:17:31.739 [2024-07-15 11:28:15.214795] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:17:31.739 [2024-07-15 11:28:15.214849] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:31.739 [2024-07-15 11:28:15.214870] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x210e0c0 00:17:31.739 [2024-07-15 11:28:15.214893] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:31.739 [2024-07-15 11:28:15.215271] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:31.739 [2024-07-15 11:28:15.215292] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:17:31.739 [2024-07-15 11:28:15.215360] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:17:31.739 [2024-07-15 11:28:15.215380] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:17:31.739 [2024-07-15 11:28:15.215485] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x210ea40 00:17:31.739 [2024-07-15 11:28:15.215496] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:17:31.739 [2024-07-15 11:28:15.215669] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x22ad6c0 00:17:31.739 [2024-07-15 11:28:15.215795] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x210ea40 00:17:31.739 [2024-07-15 11:28:15.215805] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x210ea40 00:17:31.739 [2024-07-15 11:28:15.215903] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:31.739 pt3 00:17:31.739 11:28:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@553 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:31.739 11:28:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:17:31.739 11:28:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:17:31.739 11:28:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:17:31.739 11:28:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:17:31.739 11:28:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:17:31.739 11:28:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:31.739 11:28:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:31.739 11:28:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:31.739 11:28:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:31.739 11:28:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:31.739 11:28:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:32.004 11:28:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:32.004 "name": "raid_bdev1", 00:17:32.004 "uuid": "356087b2-a385-4376-9bf5-a0fcd40d5fb1", 00:17:32.004 "strip_size_kb": 0, 00:17:32.004 "state": "online", 00:17:32.004 "raid_level": "raid1", 00:17:32.005 "superblock": true, 00:17:32.005 "num_base_bdevs": 3, 00:17:32.005 "num_base_bdevs_discovered": 2, 00:17:32.005 "num_base_bdevs_operational": 2, 00:17:32.005 "base_bdevs_list": [ 00:17:32.005 { 00:17:32.005 "name": null, 00:17:32.005 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:32.005 "is_configured": false, 00:17:32.005 "data_offset": 2048, 00:17:32.005 "data_size": 63488 00:17:32.005 }, 00:17:32.005 { 00:17:32.005 "name": "pt2", 00:17:32.005 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:32.005 "is_configured": true, 00:17:32.005 "data_offset": 2048, 00:17:32.005 "data_size": 63488 00:17:32.005 }, 00:17:32.005 { 00:17:32.005 "name": "pt3", 00:17:32.005 "uuid": "00000000-0000-0000-0000-000000000003", 00:17:32.005 "is_configured": true, 00:17:32.005 "data_offset": 2048, 00:17:32.005 "data_size": 63488 00:17:32.005 } 00:17:32.005 ] 00:17:32.005 }' 00:17:32.005 11:28:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:32.005 11:28:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:17:32.569 11:28:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs online 00:17:32.569 11:28:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:17:32.827 11:28:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # [[ false == \f\a\l\s\e ]] 00:17:32.827 11:28:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@557 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:17:32.827 11:28:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@557 -- # jq -r '.[] | .uuid' 00:17:33.085 [2024-07-15 11:28:16.542577] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:33.085 11:28:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@557 -- # '[' 356087b2-a385-4376-9bf5-a0fcd40d5fb1 '!=' 356087b2-a385-4376-9bf5-a0fcd40d5fb1 ']' 00:17:33.085 11:28:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@562 -- # killprocess 917573 00:17:33.085 11:28:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@948 -- # '[' -z 917573 ']' 00:17:33.085 11:28:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@952 -- # kill -0 917573 00:17:33.085 11:28:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@953 -- # uname 00:17:33.085 11:28:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:17:33.085 11:28:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 917573 00:17:33.085 11:28:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:17:33.085 11:28:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:17:33.085 11:28:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 917573' 00:17:33.085 killing process with pid 917573 00:17:33.085 11:28:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@967 -- # kill 917573 00:17:33.085 [2024-07-15 11:28:16.608606] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:33.085 [2024-07-15 11:28:16.608667] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:33.085 [2024-07-15 11:28:16.608724] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:33.085 [2024-07-15 11:28:16.608737] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x210ea40 name raid_bdev1, state offline 00:17:33.085 11:28:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # wait 917573 00:17:33.085 [2024-07-15 11:28:16.639133] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:33.342 11:28:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@564 -- # return 0 00:17:33.342 00:17:33.342 real 0m21.759s 00:17:33.342 user 0m39.735s 00:17:33.342 sys 0m3.978s 00:17:33.342 11:28:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:17:33.342 11:28:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:17:33.342 ************************************ 00:17:33.342 END TEST raid_superblock_test 00:17:33.342 ************************************ 00:17:33.342 11:28:16 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:17:33.342 11:28:16 bdev_raid -- bdev/bdev_raid.sh@870 -- # run_test raid_read_error_test raid_io_error_test raid1 3 read 00:17:33.342 11:28:16 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:17:33.342 11:28:16 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:17:33.342 11:28:16 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:33.600 ************************************ 00:17:33.600 START TEST raid_read_error_test 00:17:33.600 ************************************ 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1123 -- # raid_io_error_test raid1 3 read 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@788 -- # local raid_level=raid1 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@789 -- # local num_base_bdevs=3 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local error_io_type=read 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i = 1 )) 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev1 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev2 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev3 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local base_bdevs 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local raid_bdev_name=raid_bdev1 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local strip_size 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local create_arg 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local bdevperf_log 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local fail_per_s 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # '[' raid1 '!=' raid1 ']' 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # strip_size=0 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@805 -- # mktemp -p /raidtest 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@805 -- # bdevperf_log=/raidtest/tmp.9I034WwVWa 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@808 -- # raid_pid=920841 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # waitforlisten 920841 /var/tmp/spdk-raid.sock 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@829 -- # '[' -z 920841 ']' 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:17:33.600 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:17:33.600 11:28:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:17:33.600 [2024-07-15 11:28:17.022257] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:17:33.600 [2024-07-15 11:28:17.022327] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid920841 ] 00:17:33.600 [2024-07-15 11:28:17.151000] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:33.857 [2024-07-15 11:28:17.255877] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:17:33.857 [2024-07-15 11:28:17.321684] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:33.857 [2024-07-15 11:28:17.321721] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:34.422 11:28:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:17:34.422 11:28:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@862 -- # return 0 00:17:34.422 11:28:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:17:34.422 11:28:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:17:34.680 BaseBdev1_malloc 00:17:34.680 11:28:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev1_malloc 00:17:34.941 true 00:17:34.941 11:28:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:17:35.199 [2024-07-15 11:28:18.682867] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:17:35.199 [2024-07-15 11:28:18.682913] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:35.199 [2024-07-15 11:28:18.682944] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xa610d0 00:17:35.199 [2024-07-15 11:28:18.682958] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:35.199 [2024-07-15 11:28:18.684832] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:35.199 [2024-07-15 11:28:18.684862] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:17:35.199 BaseBdev1 00:17:35.200 11:28:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:17:35.200 11:28:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:17:35.457 BaseBdev2_malloc 00:17:35.457 11:28:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev2_malloc 00:17:35.715 true 00:17:35.715 11:28:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:17:35.973 [2024-07-15 11:28:19.401351] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:17:35.973 [2024-07-15 11:28:19.401391] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:35.973 [2024-07-15 11:28:19.401413] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xa65910 00:17:35.973 [2024-07-15 11:28:19.401425] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:35.973 [2024-07-15 11:28:19.402983] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:35.973 [2024-07-15 11:28:19.403010] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:17:35.973 BaseBdev2 00:17:35.973 11:28:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:17:35.973 11:28:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:17:36.231 BaseBdev3_malloc 00:17:36.231 11:28:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev3_malloc 00:17:36.488 true 00:17:36.488 11:28:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:17:36.746 [2024-07-15 11:28:20.133067] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:17:36.746 [2024-07-15 11:28:20.133116] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:36.746 [2024-07-15 11:28:20.133138] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xa67bd0 00:17:36.746 [2024-07-15 11:28:20.133151] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:36.746 [2024-07-15 11:28:20.134798] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:36.746 [2024-07-15 11:28:20.134826] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:17:36.746 BaseBdev3 00:17:36.746 11:28:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@819 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n raid_bdev1 -s 00:17:37.004 [2024-07-15 11:28:20.377728] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:37.004 [2024-07-15 11:28:20.379055] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:37.004 [2024-07-15 11:28:20.379130] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:17:37.004 [2024-07-15 11:28:20.379345] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xa69280 00:17:37.004 [2024-07-15 11:28:20.379357] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:17:37.004 [2024-07-15 11:28:20.379551] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xa68e20 00:17:37.004 [2024-07-15 11:28:20.379703] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xa69280 00:17:37.004 [2024-07-15 11:28:20.379714] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0xa69280 00:17:37.004 [2024-07-15 11:28:20.379821] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:37.004 11:28:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@820 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:17:37.004 11:28:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:17:37.004 11:28:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:17:37.004 11:28:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:17:37.004 11:28:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:17:37.004 11:28:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:17:37.004 11:28:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:37.004 11:28:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:37.004 11:28:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:37.004 11:28:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:37.004 11:28:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:37.004 11:28:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:37.262 11:28:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:37.262 "name": "raid_bdev1", 00:17:37.262 "uuid": "1de57c1f-3ac6-4557-87f5-ae182ab437c0", 00:17:37.262 "strip_size_kb": 0, 00:17:37.262 "state": "online", 00:17:37.262 "raid_level": "raid1", 00:17:37.262 "superblock": true, 00:17:37.262 "num_base_bdevs": 3, 00:17:37.262 "num_base_bdevs_discovered": 3, 00:17:37.262 "num_base_bdevs_operational": 3, 00:17:37.262 "base_bdevs_list": [ 00:17:37.262 { 00:17:37.262 "name": "BaseBdev1", 00:17:37.262 "uuid": "abdcce41-6ba4-5856-93a9-bbec235184ff", 00:17:37.262 "is_configured": true, 00:17:37.262 "data_offset": 2048, 00:17:37.262 "data_size": 63488 00:17:37.262 }, 00:17:37.262 { 00:17:37.262 "name": "BaseBdev2", 00:17:37.262 "uuid": "ecebbc00-8099-59c0-9fd7-4e2ee7872c5d", 00:17:37.263 "is_configured": true, 00:17:37.263 "data_offset": 2048, 00:17:37.263 "data_size": 63488 00:17:37.263 }, 00:17:37.263 { 00:17:37.263 "name": "BaseBdev3", 00:17:37.263 "uuid": "011262ff-ec92-5674-bfc7-8ab7ac9ab2d4", 00:17:37.263 "is_configured": true, 00:17:37.263 "data_offset": 2048, 00:17:37.263 "data_size": 63488 00:17:37.263 } 00:17:37.263 ] 00:17:37.263 }' 00:17:37.263 11:28:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:37.263 11:28:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:17:37.828 11:28:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@824 -- # sleep 1 00:17:37.828 11:28:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@823 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:17:37.828 [2024-07-15 11:28:21.328532] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x8b6e00 00:17:38.761 11:28:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@827 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:17:39.019 11:28:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # local expected_num_base_bdevs 00:17:39.019 11:28:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@830 -- # [[ raid1 = \r\a\i\d\1 ]] 00:17:39.019 11:28:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@830 -- # [[ read = \w\r\i\t\e ]] 00:17:39.019 11:28:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=3 00:17:39.019 11:28:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:17:39.019 11:28:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:17:39.019 11:28:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:17:39.019 11:28:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:17:39.019 11:28:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:17:39.019 11:28:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:17:39.019 11:28:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:39.019 11:28:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:39.019 11:28:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:39.019 11:28:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:39.019 11:28:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:39.019 11:28:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:39.277 11:28:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:39.277 "name": "raid_bdev1", 00:17:39.277 "uuid": "1de57c1f-3ac6-4557-87f5-ae182ab437c0", 00:17:39.277 "strip_size_kb": 0, 00:17:39.277 "state": "online", 00:17:39.277 "raid_level": "raid1", 00:17:39.278 "superblock": true, 00:17:39.278 "num_base_bdevs": 3, 00:17:39.278 "num_base_bdevs_discovered": 3, 00:17:39.278 "num_base_bdevs_operational": 3, 00:17:39.278 "base_bdevs_list": [ 00:17:39.278 { 00:17:39.278 "name": "BaseBdev1", 00:17:39.278 "uuid": "abdcce41-6ba4-5856-93a9-bbec235184ff", 00:17:39.278 "is_configured": true, 00:17:39.278 "data_offset": 2048, 00:17:39.278 "data_size": 63488 00:17:39.278 }, 00:17:39.278 { 00:17:39.278 "name": "BaseBdev2", 00:17:39.278 "uuid": "ecebbc00-8099-59c0-9fd7-4e2ee7872c5d", 00:17:39.278 "is_configured": true, 00:17:39.278 "data_offset": 2048, 00:17:39.278 "data_size": 63488 00:17:39.278 }, 00:17:39.278 { 00:17:39.278 "name": "BaseBdev3", 00:17:39.278 "uuid": "011262ff-ec92-5674-bfc7-8ab7ac9ab2d4", 00:17:39.278 "is_configured": true, 00:17:39.278 "data_offset": 2048, 00:17:39.278 "data_size": 63488 00:17:39.278 } 00:17:39.278 ] 00:17:39.278 }' 00:17:39.278 11:28:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:39.278 11:28:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:17:39.851 11:28:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:17:40.110 [2024-07-15 11:28:23.523899] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:40.110 [2024-07-15 11:28:23.523943] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:40.110 [2024-07-15 11:28:23.527156] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:40.110 [2024-07-15 11:28:23.527189] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:40.110 [2024-07-15 11:28:23.527286] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:40.110 [2024-07-15 11:28:23.527298] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xa69280 name raid_bdev1, state offline 00:17:40.110 0 00:17:40.110 11:28:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # killprocess 920841 00:17:40.110 11:28:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@948 -- # '[' -z 920841 ']' 00:17:40.110 11:28:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@952 -- # kill -0 920841 00:17:40.110 11:28:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@953 -- # uname 00:17:40.110 11:28:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:17:40.110 11:28:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 920841 00:17:40.110 11:28:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:17:40.110 11:28:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:17:40.110 11:28:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 920841' 00:17:40.110 killing process with pid 920841 00:17:40.110 11:28:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@967 -- # kill 920841 00:17:40.110 [2024-07-15 11:28:23.591033] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:40.110 11:28:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # wait 920841 00:17:40.110 [2024-07-15 11:28:23.611762] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:40.368 11:28:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # grep -v Job /raidtest/tmp.9I034WwVWa 00:17:40.368 11:28:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # grep raid_bdev1 00:17:40.368 11:28:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # awk '{print $6}' 00:17:40.368 11:28:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # fail_per_s=0.00 00:17:40.368 11:28:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@844 -- # has_redundancy raid1 00:17:40.368 11:28:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:17:40.368 11:28:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@214 -- # return 0 00:17:40.368 11:28:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # [[ 0.00 = \0\.\0\0 ]] 00:17:40.368 00:17:40.368 real 0m6.906s 00:17:40.368 user 0m10.884s 00:17:40.368 sys 0m1.227s 00:17:40.368 11:28:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:17:40.368 11:28:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:17:40.368 ************************************ 00:17:40.368 END TEST raid_read_error_test 00:17:40.368 ************************************ 00:17:40.368 11:28:23 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:17:40.368 11:28:23 bdev_raid -- bdev/bdev_raid.sh@871 -- # run_test raid_write_error_test raid_io_error_test raid1 3 write 00:17:40.368 11:28:23 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:17:40.368 11:28:23 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:17:40.368 11:28:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:40.368 ************************************ 00:17:40.368 START TEST raid_write_error_test 00:17:40.368 ************************************ 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1123 -- # raid_io_error_test raid1 3 write 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@788 -- # local raid_level=raid1 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@789 -- # local num_base_bdevs=3 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local error_io_type=write 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i = 1 )) 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev1 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev2 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev3 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local base_bdevs 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local raid_bdev_name=raid_bdev1 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local strip_size 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local create_arg 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local bdevperf_log 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local fail_per_s 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # '[' raid1 '!=' raid1 ']' 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # strip_size=0 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@805 -- # mktemp -p /raidtest 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@805 -- # bdevperf_log=/raidtest/tmp.6TiP8MTmJf 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@808 -- # raid_pid=921868 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # waitforlisten 921868 /var/tmp/spdk-raid.sock 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:17:40.368 11:28:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@829 -- # '[' -z 921868 ']' 00:17:40.626 11:28:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:17:40.626 11:28:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:17:40.626 11:28:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:17:40.626 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:17:40.626 11:28:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:17:40.626 11:28:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:17:40.626 [2024-07-15 11:28:24.024173] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:17:40.626 [2024-07-15 11:28:24.024246] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid921868 ] 00:17:40.626 [2024-07-15 11:28:24.154577] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:40.884 [2024-07-15 11:28:24.254384] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:17:40.884 [2024-07-15 11:28:24.319194] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:40.884 [2024-07-15 11:28:24.319244] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:41.448 11:28:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:17:41.448 11:28:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@862 -- # return 0 00:17:41.448 11:28:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:17:41.448 11:28:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:17:41.448 BaseBdev1_malloc 00:17:41.705 11:28:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev1_malloc 00:17:41.705 true 00:17:41.705 11:28:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:17:41.962 [2024-07-15 11:28:25.389070] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:17:41.962 [2024-07-15 11:28:25.389115] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:41.962 [2024-07-15 11:28:25.389136] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x10540d0 00:17:41.962 [2024-07-15 11:28:25.389148] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:41.962 [2024-07-15 11:28:25.390868] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:41.962 [2024-07-15 11:28:25.390895] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:17:41.962 BaseBdev1 00:17:41.962 11:28:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:17:41.962 11:28:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:17:42.219 BaseBdev2_malloc 00:17:42.219 11:28:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev2_malloc 00:17:42.219 true 00:17:42.219 11:28:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:17:42.476 [2024-07-15 11:28:25.919054] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:17:42.476 [2024-07-15 11:28:25.919098] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:42.476 [2024-07-15 11:28:25.919117] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1058910 00:17:42.476 [2024-07-15 11:28:25.919130] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:42.476 [2024-07-15 11:28:25.920508] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:42.476 [2024-07-15 11:28:25.920534] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:17:42.476 BaseBdev2 00:17:42.476 11:28:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:17:42.476 11:28:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:17:42.734 BaseBdev3_malloc 00:17:42.734 11:28:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev3_malloc 00:17:42.734 true 00:17:42.734 11:28:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:17:42.991 [2024-07-15 11:28:26.441075] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:17:42.992 [2024-07-15 11:28:26.441119] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:42.992 [2024-07-15 11:28:26.441137] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x105abd0 00:17:42.992 [2024-07-15 11:28:26.441149] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:42.992 [2024-07-15 11:28:26.442513] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:42.992 [2024-07-15 11:28:26.442539] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:17:42.992 BaseBdev3 00:17:42.992 11:28:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@819 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3' -n raid_bdev1 -s 00:17:43.248 [2024-07-15 11:28:26.621580] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:43.248 [2024-07-15 11:28:26.622818] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:43.248 [2024-07-15 11:28:26.622886] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:17:43.248 [2024-07-15 11:28:26.623101] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x105c280 00:17:43.248 [2024-07-15 11:28:26.623114] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:17:43.248 [2024-07-15 11:28:26.623301] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x105be20 00:17:43.248 [2024-07-15 11:28:26.623451] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x105c280 00:17:43.248 [2024-07-15 11:28:26.623461] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x105c280 00:17:43.248 [2024-07-15 11:28:26.623559] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:43.248 11:28:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@820 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:17:43.248 11:28:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:17:43.248 11:28:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:17:43.248 11:28:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:17:43.248 11:28:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:17:43.248 11:28:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:17:43.248 11:28:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:43.248 11:28:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:43.248 11:28:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:43.248 11:28:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:43.248 11:28:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:43.248 11:28:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:43.504 11:28:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:43.504 "name": "raid_bdev1", 00:17:43.504 "uuid": "a92260ac-80d6-4c84-8d2f-2c22e641cbfe", 00:17:43.504 "strip_size_kb": 0, 00:17:43.504 "state": "online", 00:17:43.504 "raid_level": "raid1", 00:17:43.504 "superblock": true, 00:17:43.504 "num_base_bdevs": 3, 00:17:43.504 "num_base_bdevs_discovered": 3, 00:17:43.504 "num_base_bdevs_operational": 3, 00:17:43.504 "base_bdevs_list": [ 00:17:43.504 { 00:17:43.504 "name": "BaseBdev1", 00:17:43.504 "uuid": "0f8644af-6066-5599-baf8-5c47d43b219b", 00:17:43.504 "is_configured": true, 00:17:43.504 "data_offset": 2048, 00:17:43.504 "data_size": 63488 00:17:43.504 }, 00:17:43.504 { 00:17:43.504 "name": "BaseBdev2", 00:17:43.504 "uuid": "c4070988-5f90-55bb-a0cd-1a388ad0aae5", 00:17:43.504 "is_configured": true, 00:17:43.504 "data_offset": 2048, 00:17:43.504 "data_size": 63488 00:17:43.504 }, 00:17:43.504 { 00:17:43.504 "name": "BaseBdev3", 00:17:43.504 "uuid": "443969f0-d7fd-5b61-b5f2-9c51eb3ae45f", 00:17:43.504 "is_configured": true, 00:17:43.504 "data_offset": 2048, 00:17:43.504 "data_size": 63488 00:17:43.504 } 00:17:43.504 ] 00:17:43.504 }' 00:17:43.504 11:28:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:43.504 11:28:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:17:44.069 11:28:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@824 -- # sleep 1 00:17:44.069 11:28:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@823 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:17:44.069 [2024-07-15 11:28:27.564361] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xea9e00 00:17:45.005 11:28:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@827 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:17:45.264 [2024-07-15 11:28:28.651414] bdev_raid.c:2221:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:17:45.264 [2024-07-15 11:28:28.651466] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:17:45.264 [2024-07-15 11:28:28.651662] bdev_raid.c:1919:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0xea9e00 00:17:45.264 11:28:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # local expected_num_base_bdevs 00:17:45.264 11:28:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@830 -- # [[ raid1 = \r\a\i\d\1 ]] 00:17:45.264 11:28:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@830 -- # [[ write = \w\r\i\t\e ]] 00:17:45.264 11:28:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # expected_num_base_bdevs=2 00:17:45.264 11:28:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:45.264 11:28:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:17:45.264 11:28:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:17:45.264 11:28:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:17:45.264 11:28:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:17:45.264 11:28:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:17:45.264 11:28:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:45.264 11:28:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:45.264 11:28:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:45.264 11:28:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:45.264 11:28:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:45.264 11:28:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:45.264 11:28:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:45.264 "name": "raid_bdev1", 00:17:45.264 "uuid": "a92260ac-80d6-4c84-8d2f-2c22e641cbfe", 00:17:45.264 "strip_size_kb": 0, 00:17:45.264 "state": "online", 00:17:45.264 "raid_level": "raid1", 00:17:45.264 "superblock": true, 00:17:45.264 "num_base_bdevs": 3, 00:17:45.264 "num_base_bdevs_discovered": 2, 00:17:45.264 "num_base_bdevs_operational": 2, 00:17:45.264 "base_bdevs_list": [ 00:17:45.264 { 00:17:45.264 "name": null, 00:17:45.264 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:45.264 "is_configured": false, 00:17:45.264 "data_offset": 2048, 00:17:45.264 "data_size": 63488 00:17:45.264 }, 00:17:45.264 { 00:17:45.264 "name": "BaseBdev2", 00:17:45.264 "uuid": "c4070988-5f90-55bb-a0cd-1a388ad0aae5", 00:17:45.264 "is_configured": true, 00:17:45.264 "data_offset": 2048, 00:17:45.264 "data_size": 63488 00:17:45.264 }, 00:17:45.264 { 00:17:45.264 "name": "BaseBdev3", 00:17:45.264 "uuid": "443969f0-d7fd-5b61-b5f2-9c51eb3ae45f", 00:17:45.264 "is_configured": true, 00:17:45.264 "data_offset": 2048, 00:17:45.264 "data_size": 63488 00:17:45.264 } 00:17:45.264 ] 00:17:45.264 }' 00:17:45.264 11:28:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:45.264 11:28:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:17:45.829 11:28:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:17:46.086 [2024-07-15 11:28:29.609582] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:46.086 [2024-07-15 11:28:29.609623] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:46.086 [2024-07-15 11:28:29.612803] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:46.086 [2024-07-15 11:28:29.612836] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:46.086 [2024-07-15 11:28:29.612909] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:46.086 [2024-07-15 11:28:29.612921] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x105c280 name raid_bdev1, state offline 00:17:46.086 0 00:17:46.086 11:28:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # killprocess 921868 00:17:46.086 11:28:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@948 -- # '[' -z 921868 ']' 00:17:46.086 11:28:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@952 -- # kill -0 921868 00:17:46.086 11:28:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@953 -- # uname 00:17:46.086 11:28:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:17:46.086 11:28:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 921868 00:17:46.086 11:28:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:17:46.086 11:28:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:17:46.086 11:28:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 921868' 00:17:46.086 killing process with pid 921868 00:17:46.086 11:28:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@967 -- # kill 921868 00:17:46.086 [2024-07-15 11:28:29.680348] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:46.086 11:28:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # wait 921868 00:17:46.343 [2024-07-15 11:28:29.704223] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:46.601 11:28:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # grep -v Job /raidtest/tmp.6TiP8MTmJf 00:17:46.601 11:28:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # grep raid_bdev1 00:17:46.601 11:28:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # awk '{print $6}' 00:17:46.601 11:28:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # fail_per_s=0.00 00:17:46.601 11:28:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@844 -- # has_redundancy raid1 00:17:46.601 11:28:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:17:46.601 11:28:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@214 -- # return 0 00:17:46.601 11:28:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # [[ 0.00 = \0\.\0\0 ]] 00:17:46.601 00:17:46.601 real 0m6.009s 00:17:46.601 user 0m9.222s 00:17:46.601 sys 0m1.099s 00:17:46.601 11:28:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:17:46.601 11:28:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:17:46.601 ************************************ 00:17:46.601 END TEST raid_write_error_test 00:17:46.601 ************************************ 00:17:46.601 11:28:29 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:17:46.601 11:28:29 bdev_raid -- bdev/bdev_raid.sh@865 -- # for n in {2..4} 00:17:46.601 11:28:29 bdev_raid -- bdev/bdev_raid.sh@866 -- # for level in raid0 concat raid1 00:17:46.601 11:28:29 bdev_raid -- bdev/bdev_raid.sh@867 -- # run_test raid_state_function_test raid_state_function_test raid0 4 false 00:17:46.601 11:28:29 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:17:46.601 11:28:29 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:17:46.601 11:28:29 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:46.601 ************************************ 00:17:46.601 START TEST raid_state_function_test 00:17:46.601 ************************************ 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1123 -- # raid_state_function_test raid0 4 false 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@220 -- # local raid_level=raid0 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=4 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # local superblock=false 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev3 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev4 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # local strip_size 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # '[' raid0 '!=' raid1 ']' 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # strip_size=64 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@232 -- # strip_size_create_arg='-z 64' 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # '[' false = true ']' 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@240 -- # superblock_create_arg= 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # raid_pid=922789 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 922789' 00:17:46.601 Process raid pid: 922789 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@246 -- # waitforlisten 922789 /var/tmp/spdk-raid.sock 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@829 -- # '[' -z 922789 ']' 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:17:46.601 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:17:46.601 11:28:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:17:46.601 [2024-07-15 11:28:30.150081] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:17:46.601 [2024-07-15 11:28:30.150217] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:17:46.859 [2024-07-15 11:28:30.339372] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:46.859 [2024-07-15 11:28:30.436344] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:17:47.116 [2024-07-15 11:28:30.496160] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:47.116 [2024-07-15 11:28:30.496196] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:47.682 11:28:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:17:47.682 11:28:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@862 -- # return 0 00:17:47.682 11:28:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:17:47.682 [2024-07-15 11:28:31.243307] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:17:47.682 [2024-07-15 11:28:31.243349] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:17:47.682 [2024-07-15 11:28:31.243360] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:17:47.682 [2024-07-15 11:28:31.243371] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:17:47.682 [2024-07-15 11:28:31.243380] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:17:47.682 [2024-07-15 11:28:31.243392] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:17:47.682 [2024-07-15 11:28:31.243400] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:17:47.682 [2024-07-15 11:28:31.243411] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:17:47.682 11:28:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:17:47.682 11:28:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:17:47.682 11:28:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:17:47.682 11:28:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:17:47.682 11:28:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:17:47.682 11:28:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:17:47.682 11:28:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:47.682 11:28:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:47.682 11:28:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:47.682 11:28:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:47.682 11:28:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:47.682 11:28:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:47.940 11:28:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:47.940 "name": "Existed_Raid", 00:17:47.940 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:47.940 "strip_size_kb": 64, 00:17:47.940 "state": "configuring", 00:17:47.940 "raid_level": "raid0", 00:17:47.940 "superblock": false, 00:17:47.940 "num_base_bdevs": 4, 00:17:47.940 "num_base_bdevs_discovered": 0, 00:17:47.940 "num_base_bdevs_operational": 4, 00:17:47.940 "base_bdevs_list": [ 00:17:47.940 { 00:17:47.940 "name": "BaseBdev1", 00:17:47.940 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:47.940 "is_configured": false, 00:17:47.940 "data_offset": 0, 00:17:47.940 "data_size": 0 00:17:47.940 }, 00:17:47.940 { 00:17:47.940 "name": "BaseBdev2", 00:17:47.940 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:47.940 "is_configured": false, 00:17:47.940 "data_offset": 0, 00:17:47.940 "data_size": 0 00:17:47.940 }, 00:17:47.940 { 00:17:47.940 "name": "BaseBdev3", 00:17:47.940 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:47.940 "is_configured": false, 00:17:47.940 "data_offset": 0, 00:17:47.940 "data_size": 0 00:17:47.940 }, 00:17:47.940 { 00:17:47.940 "name": "BaseBdev4", 00:17:47.940 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:47.940 "is_configured": false, 00:17:47.940 "data_offset": 0, 00:17:47.940 "data_size": 0 00:17:47.940 } 00:17:47.940 ] 00:17:47.940 }' 00:17:47.940 11:28:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:47.940 11:28:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:17:48.873 11:28:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:17:48.873 [2024-07-15 11:28:32.334079] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:17:48.873 [2024-07-15 11:28:32.334110] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1b2daa0 name Existed_Raid, state configuring 00:17:48.873 11:28:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:17:49.132 [2024-07-15 11:28:32.506554] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:17:49.132 [2024-07-15 11:28:32.506583] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:17:49.132 [2024-07-15 11:28:32.506592] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:17:49.132 [2024-07-15 11:28:32.506604] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:17:49.132 [2024-07-15 11:28:32.506612] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:17:49.132 [2024-07-15 11:28:32.506623] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:17:49.132 [2024-07-15 11:28:32.506632] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:17:49.132 [2024-07-15 11:28:32.506643] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:17:49.132 11:28:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:17:49.132 [2024-07-15 11:28:32.688901] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:49.132 BaseBdev1 00:17:49.132 11:28:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:17:49.132 11:28:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:17:49.132 11:28:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:17:49.132 11:28:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:17:49.132 11:28:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:17:49.132 11:28:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:17:49.132 11:28:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:17:49.389 11:28:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:17:49.646 [ 00:17:49.646 { 00:17:49.646 "name": "BaseBdev1", 00:17:49.646 "aliases": [ 00:17:49.646 "639694bd-d65a-4fd0-81e7-586f0647bf49" 00:17:49.646 ], 00:17:49.646 "product_name": "Malloc disk", 00:17:49.646 "block_size": 512, 00:17:49.646 "num_blocks": 65536, 00:17:49.646 "uuid": "639694bd-d65a-4fd0-81e7-586f0647bf49", 00:17:49.646 "assigned_rate_limits": { 00:17:49.646 "rw_ios_per_sec": 0, 00:17:49.646 "rw_mbytes_per_sec": 0, 00:17:49.646 "r_mbytes_per_sec": 0, 00:17:49.646 "w_mbytes_per_sec": 0 00:17:49.646 }, 00:17:49.646 "claimed": true, 00:17:49.646 "claim_type": "exclusive_write", 00:17:49.646 "zoned": false, 00:17:49.646 "supported_io_types": { 00:17:49.646 "read": true, 00:17:49.646 "write": true, 00:17:49.646 "unmap": true, 00:17:49.646 "flush": true, 00:17:49.646 "reset": true, 00:17:49.646 "nvme_admin": false, 00:17:49.646 "nvme_io": false, 00:17:49.646 "nvme_io_md": false, 00:17:49.646 "write_zeroes": true, 00:17:49.646 "zcopy": true, 00:17:49.646 "get_zone_info": false, 00:17:49.646 "zone_management": false, 00:17:49.646 "zone_append": false, 00:17:49.646 "compare": false, 00:17:49.646 "compare_and_write": false, 00:17:49.646 "abort": true, 00:17:49.646 "seek_hole": false, 00:17:49.646 "seek_data": false, 00:17:49.646 "copy": true, 00:17:49.646 "nvme_iov_md": false 00:17:49.646 }, 00:17:49.647 "memory_domains": [ 00:17:49.647 { 00:17:49.647 "dma_device_id": "system", 00:17:49.647 "dma_device_type": 1 00:17:49.647 }, 00:17:49.647 { 00:17:49.647 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:49.647 "dma_device_type": 2 00:17:49.647 } 00:17:49.647 ], 00:17:49.647 "driver_specific": {} 00:17:49.647 } 00:17:49.647 ] 00:17:49.647 11:28:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:17:49.647 11:28:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:17:49.647 11:28:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:17:49.647 11:28:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:17:49.647 11:28:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:17:49.647 11:28:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:17:49.647 11:28:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:17:49.647 11:28:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:49.647 11:28:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:49.647 11:28:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:49.647 11:28:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:49.647 11:28:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:49.647 11:28:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:49.904 11:28:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:49.904 "name": "Existed_Raid", 00:17:49.904 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:49.904 "strip_size_kb": 64, 00:17:49.904 "state": "configuring", 00:17:49.904 "raid_level": "raid0", 00:17:49.904 "superblock": false, 00:17:49.904 "num_base_bdevs": 4, 00:17:49.904 "num_base_bdevs_discovered": 1, 00:17:49.904 "num_base_bdevs_operational": 4, 00:17:49.904 "base_bdevs_list": [ 00:17:49.904 { 00:17:49.904 "name": "BaseBdev1", 00:17:49.904 "uuid": "639694bd-d65a-4fd0-81e7-586f0647bf49", 00:17:49.904 "is_configured": true, 00:17:49.904 "data_offset": 0, 00:17:49.904 "data_size": 65536 00:17:49.904 }, 00:17:49.904 { 00:17:49.904 "name": "BaseBdev2", 00:17:49.904 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:49.904 "is_configured": false, 00:17:49.904 "data_offset": 0, 00:17:49.904 "data_size": 0 00:17:49.904 }, 00:17:49.904 { 00:17:49.904 "name": "BaseBdev3", 00:17:49.904 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:49.904 "is_configured": false, 00:17:49.904 "data_offset": 0, 00:17:49.904 "data_size": 0 00:17:49.904 }, 00:17:49.904 { 00:17:49.904 "name": "BaseBdev4", 00:17:49.904 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:49.904 "is_configured": false, 00:17:49.904 "data_offset": 0, 00:17:49.904 "data_size": 0 00:17:49.904 } 00:17:49.904 ] 00:17:49.904 }' 00:17:49.904 11:28:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:49.904 11:28:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:17:50.469 11:28:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:17:51.035 [2024-07-15 11:28:34.373348] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:17:51.035 [2024-07-15 11:28:34.373388] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1b2d310 name Existed_Raid, state configuring 00:17:51.035 11:28:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:17:51.293 [2024-07-15 11:28:34.630067] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:51.293 [2024-07-15 11:28:34.631565] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:17:51.293 [2024-07-15 11:28:34.631598] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:17:51.293 [2024-07-15 11:28:34.631608] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:17:51.293 [2024-07-15 11:28:34.631620] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:17:51.293 [2024-07-15 11:28:34.631629] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:17:51.293 [2024-07-15 11:28:34.631640] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:17:51.293 11:28:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:17:51.293 11:28:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:17:51.293 11:28:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:17:51.293 11:28:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:17:51.293 11:28:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:17:51.293 11:28:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:17:51.293 11:28:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:17:51.293 11:28:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:17:51.293 11:28:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:51.293 11:28:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:51.293 11:28:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:51.293 11:28:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:51.293 11:28:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:51.293 11:28:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:51.551 11:28:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:51.551 "name": "Existed_Raid", 00:17:51.551 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:51.551 "strip_size_kb": 64, 00:17:51.551 "state": "configuring", 00:17:51.551 "raid_level": "raid0", 00:17:51.551 "superblock": false, 00:17:51.551 "num_base_bdevs": 4, 00:17:51.551 "num_base_bdevs_discovered": 1, 00:17:51.551 "num_base_bdevs_operational": 4, 00:17:51.551 "base_bdevs_list": [ 00:17:51.551 { 00:17:51.551 "name": "BaseBdev1", 00:17:51.551 "uuid": "639694bd-d65a-4fd0-81e7-586f0647bf49", 00:17:51.551 "is_configured": true, 00:17:51.551 "data_offset": 0, 00:17:51.551 "data_size": 65536 00:17:51.551 }, 00:17:51.551 { 00:17:51.551 "name": "BaseBdev2", 00:17:51.551 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:51.551 "is_configured": false, 00:17:51.551 "data_offset": 0, 00:17:51.551 "data_size": 0 00:17:51.551 }, 00:17:51.551 { 00:17:51.551 "name": "BaseBdev3", 00:17:51.551 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:51.551 "is_configured": false, 00:17:51.551 "data_offset": 0, 00:17:51.551 "data_size": 0 00:17:51.551 }, 00:17:51.551 { 00:17:51.551 "name": "BaseBdev4", 00:17:51.551 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:51.551 "is_configured": false, 00:17:51.551 "data_offset": 0, 00:17:51.551 "data_size": 0 00:17:51.551 } 00:17:51.551 ] 00:17:51.551 }' 00:17:51.551 11:28:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:51.551 11:28:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:17:52.114 11:28:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:17:52.372 [2024-07-15 11:28:35.720335] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:52.372 BaseBdev2 00:17:52.372 11:28:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:17:52.372 11:28:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:17:52.372 11:28:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:17:52.372 11:28:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:17:52.372 11:28:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:17:52.372 11:28:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:17:52.372 11:28:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:17:52.936 11:28:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:17:52.936 [ 00:17:52.936 { 00:17:52.936 "name": "BaseBdev2", 00:17:52.936 "aliases": [ 00:17:52.936 "6301c3e2-f974-4b66-9ec7-e663a2d0398f" 00:17:52.936 ], 00:17:52.936 "product_name": "Malloc disk", 00:17:52.936 "block_size": 512, 00:17:52.936 "num_blocks": 65536, 00:17:52.936 "uuid": "6301c3e2-f974-4b66-9ec7-e663a2d0398f", 00:17:52.936 "assigned_rate_limits": { 00:17:52.936 "rw_ios_per_sec": 0, 00:17:52.936 "rw_mbytes_per_sec": 0, 00:17:52.936 "r_mbytes_per_sec": 0, 00:17:52.936 "w_mbytes_per_sec": 0 00:17:52.936 }, 00:17:52.936 "claimed": true, 00:17:52.937 "claim_type": "exclusive_write", 00:17:52.937 "zoned": false, 00:17:52.937 "supported_io_types": { 00:17:52.937 "read": true, 00:17:52.937 "write": true, 00:17:52.937 "unmap": true, 00:17:52.937 "flush": true, 00:17:52.937 "reset": true, 00:17:52.937 "nvme_admin": false, 00:17:52.937 "nvme_io": false, 00:17:52.937 "nvme_io_md": false, 00:17:52.937 "write_zeroes": true, 00:17:52.937 "zcopy": true, 00:17:52.937 "get_zone_info": false, 00:17:52.937 "zone_management": false, 00:17:52.937 "zone_append": false, 00:17:52.937 "compare": false, 00:17:52.937 "compare_and_write": false, 00:17:52.937 "abort": true, 00:17:52.937 "seek_hole": false, 00:17:52.937 "seek_data": false, 00:17:52.937 "copy": true, 00:17:52.937 "nvme_iov_md": false 00:17:52.937 }, 00:17:52.937 "memory_domains": [ 00:17:52.937 { 00:17:52.937 "dma_device_id": "system", 00:17:52.937 "dma_device_type": 1 00:17:52.937 }, 00:17:52.937 { 00:17:52.937 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:52.937 "dma_device_type": 2 00:17:52.937 } 00:17:52.937 ], 00:17:52.937 "driver_specific": {} 00:17:52.937 } 00:17:52.937 ] 00:17:52.937 11:28:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:17:52.937 11:28:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:17:52.937 11:28:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:17:52.937 11:28:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:17:52.937 11:28:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:17:52.937 11:28:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:17:52.937 11:28:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:17:52.937 11:28:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:17:52.937 11:28:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:17:52.937 11:28:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:52.937 11:28:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:52.937 11:28:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:52.937 11:28:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:52.937 11:28:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:52.937 11:28:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:53.194 11:28:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:53.194 "name": "Existed_Raid", 00:17:53.194 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:53.194 "strip_size_kb": 64, 00:17:53.194 "state": "configuring", 00:17:53.194 "raid_level": "raid0", 00:17:53.194 "superblock": false, 00:17:53.194 "num_base_bdevs": 4, 00:17:53.194 "num_base_bdevs_discovered": 2, 00:17:53.194 "num_base_bdevs_operational": 4, 00:17:53.194 "base_bdevs_list": [ 00:17:53.194 { 00:17:53.194 "name": "BaseBdev1", 00:17:53.194 "uuid": "639694bd-d65a-4fd0-81e7-586f0647bf49", 00:17:53.194 "is_configured": true, 00:17:53.194 "data_offset": 0, 00:17:53.194 "data_size": 65536 00:17:53.194 }, 00:17:53.194 { 00:17:53.194 "name": "BaseBdev2", 00:17:53.194 "uuid": "6301c3e2-f974-4b66-9ec7-e663a2d0398f", 00:17:53.194 "is_configured": true, 00:17:53.194 "data_offset": 0, 00:17:53.194 "data_size": 65536 00:17:53.194 }, 00:17:53.194 { 00:17:53.194 "name": "BaseBdev3", 00:17:53.194 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:53.194 "is_configured": false, 00:17:53.194 "data_offset": 0, 00:17:53.194 "data_size": 0 00:17:53.194 }, 00:17:53.194 { 00:17:53.194 "name": "BaseBdev4", 00:17:53.194 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:53.194 "is_configured": false, 00:17:53.194 "data_offset": 0, 00:17:53.194 "data_size": 0 00:17:53.194 } 00:17:53.194 ] 00:17:53.194 }' 00:17:53.194 11:28:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:53.194 11:28:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:17:53.788 11:28:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:17:54.045 [2024-07-15 11:28:37.572628] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:17:54.045 BaseBdev3 00:17:54.045 11:28:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev3 00:17:54.045 11:28:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:17:54.045 11:28:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:17:54.045 11:28:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:17:54.045 11:28:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:17:54.045 11:28:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:17:54.045 11:28:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:17:54.302 11:28:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:17:54.560 [ 00:17:54.560 { 00:17:54.560 "name": "BaseBdev3", 00:17:54.560 "aliases": [ 00:17:54.560 "e3761302-a053-4633-8401-c2879c28e09f" 00:17:54.560 ], 00:17:54.560 "product_name": "Malloc disk", 00:17:54.560 "block_size": 512, 00:17:54.560 "num_blocks": 65536, 00:17:54.560 "uuid": "e3761302-a053-4633-8401-c2879c28e09f", 00:17:54.560 "assigned_rate_limits": { 00:17:54.560 "rw_ios_per_sec": 0, 00:17:54.560 "rw_mbytes_per_sec": 0, 00:17:54.560 "r_mbytes_per_sec": 0, 00:17:54.560 "w_mbytes_per_sec": 0 00:17:54.560 }, 00:17:54.560 "claimed": true, 00:17:54.560 "claim_type": "exclusive_write", 00:17:54.560 "zoned": false, 00:17:54.560 "supported_io_types": { 00:17:54.560 "read": true, 00:17:54.560 "write": true, 00:17:54.560 "unmap": true, 00:17:54.560 "flush": true, 00:17:54.560 "reset": true, 00:17:54.560 "nvme_admin": false, 00:17:54.560 "nvme_io": false, 00:17:54.560 "nvme_io_md": false, 00:17:54.560 "write_zeroes": true, 00:17:54.560 "zcopy": true, 00:17:54.560 "get_zone_info": false, 00:17:54.560 "zone_management": false, 00:17:54.560 "zone_append": false, 00:17:54.560 "compare": false, 00:17:54.560 "compare_and_write": false, 00:17:54.560 "abort": true, 00:17:54.560 "seek_hole": false, 00:17:54.560 "seek_data": false, 00:17:54.560 "copy": true, 00:17:54.560 "nvme_iov_md": false 00:17:54.560 }, 00:17:54.560 "memory_domains": [ 00:17:54.560 { 00:17:54.560 "dma_device_id": "system", 00:17:54.560 "dma_device_type": 1 00:17:54.560 }, 00:17:54.560 { 00:17:54.560 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:54.560 "dma_device_type": 2 00:17:54.560 } 00:17:54.560 ], 00:17:54.560 "driver_specific": {} 00:17:54.560 } 00:17:54.560 ] 00:17:54.560 11:28:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:17:54.560 11:28:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:17:54.560 11:28:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:17:54.560 11:28:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:17:54.560 11:28:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:17:54.560 11:28:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:17:54.560 11:28:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:17:54.560 11:28:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:17:54.560 11:28:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:17:54.560 11:28:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:54.560 11:28:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:54.560 11:28:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:54.560 11:28:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:54.560 11:28:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:54.560 11:28:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:54.818 11:28:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:54.818 "name": "Existed_Raid", 00:17:54.818 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:54.818 "strip_size_kb": 64, 00:17:54.818 "state": "configuring", 00:17:54.818 "raid_level": "raid0", 00:17:54.818 "superblock": false, 00:17:54.818 "num_base_bdevs": 4, 00:17:54.818 "num_base_bdevs_discovered": 3, 00:17:54.818 "num_base_bdevs_operational": 4, 00:17:54.818 "base_bdevs_list": [ 00:17:54.818 { 00:17:54.818 "name": "BaseBdev1", 00:17:54.818 "uuid": "639694bd-d65a-4fd0-81e7-586f0647bf49", 00:17:54.818 "is_configured": true, 00:17:54.818 "data_offset": 0, 00:17:54.818 "data_size": 65536 00:17:54.818 }, 00:17:54.818 { 00:17:54.818 "name": "BaseBdev2", 00:17:54.818 "uuid": "6301c3e2-f974-4b66-9ec7-e663a2d0398f", 00:17:54.818 "is_configured": true, 00:17:54.818 "data_offset": 0, 00:17:54.818 "data_size": 65536 00:17:54.818 }, 00:17:54.818 { 00:17:54.818 "name": "BaseBdev3", 00:17:54.818 "uuid": "e3761302-a053-4633-8401-c2879c28e09f", 00:17:54.818 "is_configured": true, 00:17:54.818 "data_offset": 0, 00:17:54.818 "data_size": 65536 00:17:54.818 }, 00:17:54.818 { 00:17:54.818 "name": "BaseBdev4", 00:17:54.818 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:54.818 "is_configured": false, 00:17:54.818 "data_offset": 0, 00:17:54.818 "data_size": 0 00:17:54.818 } 00:17:54.818 ] 00:17:54.818 }' 00:17:54.818 11:28:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:54.818 11:28:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:17:55.382 11:28:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4 00:17:55.640 [2024-07-15 11:28:39.084060] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:17:55.640 [2024-07-15 11:28:39.084099] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1b2e350 00:17:55.640 [2024-07-15 11:28:39.084107] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:17:55.640 [2024-07-15 11:28:39.084354] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1b2e020 00:17:55.640 [2024-07-15 11:28:39.084473] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1b2e350 00:17:55.640 [2024-07-15 11:28:39.084483] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x1b2e350 00:17:55.640 [2024-07-15 11:28:39.084641] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:55.640 BaseBdev4 00:17:55.640 11:28:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev4 00:17:55.640 11:28:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev4 00:17:55.640 11:28:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:17:55.640 11:28:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:17:55.640 11:28:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:17:55.640 11:28:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:17:55.640 11:28:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:17:55.897 11:28:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 -t 2000 00:17:56.155 [ 00:17:56.155 { 00:17:56.155 "name": "BaseBdev4", 00:17:56.155 "aliases": [ 00:17:56.155 "cbdc6a1e-c9e5-4232-a9f7-f2d1343aa6aa" 00:17:56.155 ], 00:17:56.155 "product_name": "Malloc disk", 00:17:56.155 "block_size": 512, 00:17:56.155 "num_blocks": 65536, 00:17:56.155 "uuid": "cbdc6a1e-c9e5-4232-a9f7-f2d1343aa6aa", 00:17:56.155 "assigned_rate_limits": { 00:17:56.155 "rw_ios_per_sec": 0, 00:17:56.155 "rw_mbytes_per_sec": 0, 00:17:56.155 "r_mbytes_per_sec": 0, 00:17:56.155 "w_mbytes_per_sec": 0 00:17:56.155 }, 00:17:56.155 "claimed": true, 00:17:56.155 "claim_type": "exclusive_write", 00:17:56.155 "zoned": false, 00:17:56.155 "supported_io_types": { 00:17:56.155 "read": true, 00:17:56.155 "write": true, 00:17:56.155 "unmap": true, 00:17:56.155 "flush": true, 00:17:56.155 "reset": true, 00:17:56.155 "nvme_admin": false, 00:17:56.155 "nvme_io": false, 00:17:56.155 "nvme_io_md": false, 00:17:56.155 "write_zeroes": true, 00:17:56.155 "zcopy": true, 00:17:56.155 "get_zone_info": false, 00:17:56.155 "zone_management": false, 00:17:56.155 "zone_append": false, 00:17:56.155 "compare": false, 00:17:56.155 "compare_and_write": false, 00:17:56.155 "abort": true, 00:17:56.155 "seek_hole": false, 00:17:56.155 "seek_data": false, 00:17:56.155 "copy": true, 00:17:56.155 "nvme_iov_md": false 00:17:56.155 }, 00:17:56.155 "memory_domains": [ 00:17:56.155 { 00:17:56.155 "dma_device_id": "system", 00:17:56.155 "dma_device_type": 1 00:17:56.155 }, 00:17:56.155 { 00:17:56.155 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:56.155 "dma_device_type": 2 00:17:56.155 } 00:17:56.155 ], 00:17:56.155 "driver_specific": {} 00:17:56.155 } 00:17:56.155 ] 00:17:56.155 11:28:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:17:56.155 11:28:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:17:56.155 11:28:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:17:56.155 11:28:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:17:56.155 11:28:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:17:56.155 11:28:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:17:56.155 11:28:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:17:56.155 11:28:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:17:56.155 11:28:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:17:56.155 11:28:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:56.155 11:28:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:56.155 11:28:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:56.155 11:28:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:56.155 11:28:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:56.155 11:28:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:56.412 11:28:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:17:56.412 "name": "Existed_Raid", 00:17:56.412 "uuid": "2285c7ae-edd9-48e8-b4bd-dd2d2c3701ca", 00:17:56.412 "strip_size_kb": 64, 00:17:56.412 "state": "online", 00:17:56.412 "raid_level": "raid0", 00:17:56.412 "superblock": false, 00:17:56.412 "num_base_bdevs": 4, 00:17:56.412 "num_base_bdevs_discovered": 4, 00:17:56.412 "num_base_bdevs_operational": 4, 00:17:56.412 "base_bdevs_list": [ 00:17:56.412 { 00:17:56.412 "name": "BaseBdev1", 00:17:56.412 "uuid": "639694bd-d65a-4fd0-81e7-586f0647bf49", 00:17:56.412 "is_configured": true, 00:17:56.412 "data_offset": 0, 00:17:56.412 "data_size": 65536 00:17:56.412 }, 00:17:56.412 { 00:17:56.412 "name": "BaseBdev2", 00:17:56.412 "uuid": "6301c3e2-f974-4b66-9ec7-e663a2d0398f", 00:17:56.412 "is_configured": true, 00:17:56.412 "data_offset": 0, 00:17:56.412 "data_size": 65536 00:17:56.412 }, 00:17:56.412 { 00:17:56.412 "name": "BaseBdev3", 00:17:56.412 "uuid": "e3761302-a053-4633-8401-c2879c28e09f", 00:17:56.412 "is_configured": true, 00:17:56.412 "data_offset": 0, 00:17:56.412 "data_size": 65536 00:17:56.412 }, 00:17:56.412 { 00:17:56.412 "name": "BaseBdev4", 00:17:56.412 "uuid": "cbdc6a1e-c9e5-4232-a9f7-f2d1343aa6aa", 00:17:56.412 "is_configured": true, 00:17:56.412 "data_offset": 0, 00:17:56.412 "data_size": 65536 00:17:56.412 } 00:17:56.412 ] 00:17:56.412 }' 00:17:56.412 11:28:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:17:56.412 11:28:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:17:56.977 11:28:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:17:56.977 11:28:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:17:56.977 11:28:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:17:56.977 11:28:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:17:56.977 11:28:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:17:56.977 11:28:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # local name 00:17:56.977 11:28:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:17:56.977 11:28:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:17:57.234 [2024-07-15 11:28:40.656604] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:57.234 11:28:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:17:57.234 "name": "Existed_Raid", 00:17:57.234 "aliases": [ 00:17:57.234 "2285c7ae-edd9-48e8-b4bd-dd2d2c3701ca" 00:17:57.234 ], 00:17:57.234 "product_name": "Raid Volume", 00:17:57.234 "block_size": 512, 00:17:57.234 "num_blocks": 262144, 00:17:57.234 "uuid": "2285c7ae-edd9-48e8-b4bd-dd2d2c3701ca", 00:17:57.234 "assigned_rate_limits": { 00:17:57.234 "rw_ios_per_sec": 0, 00:17:57.234 "rw_mbytes_per_sec": 0, 00:17:57.234 "r_mbytes_per_sec": 0, 00:17:57.234 "w_mbytes_per_sec": 0 00:17:57.234 }, 00:17:57.234 "claimed": false, 00:17:57.234 "zoned": false, 00:17:57.234 "supported_io_types": { 00:17:57.234 "read": true, 00:17:57.234 "write": true, 00:17:57.234 "unmap": true, 00:17:57.234 "flush": true, 00:17:57.234 "reset": true, 00:17:57.234 "nvme_admin": false, 00:17:57.234 "nvme_io": false, 00:17:57.234 "nvme_io_md": false, 00:17:57.234 "write_zeroes": true, 00:17:57.234 "zcopy": false, 00:17:57.234 "get_zone_info": false, 00:17:57.234 "zone_management": false, 00:17:57.234 "zone_append": false, 00:17:57.234 "compare": false, 00:17:57.234 "compare_and_write": false, 00:17:57.234 "abort": false, 00:17:57.234 "seek_hole": false, 00:17:57.234 "seek_data": false, 00:17:57.234 "copy": false, 00:17:57.234 "nvme_iov_md": false 00:17:57.234 }, 00:17:57.234 "memory_domains": [ 00:17:57.234 { 00:17:57.234 "dma_device_id": "system", 00:17:57.234 "dma_device_type": 1 00:17:57.234 }, 00:17:57.234 { 00:17:57.234 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:57.234 "dma_device_type": 2 00:17:57.234 }, 00:17:57.234 { 00:17:57.234 "dma_device_id": "system", 00:17:57.234 "dma_device_type": 1 00:17:57.234 }, 00:17:57.234 { 00:17:57.234 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:57.234 "dma_device_type": 2 00:17:57.234 }, 00:17:57.234 { 00:17:57.234 "dma_device_id": "system", 00:17:57.234 "dma_device_type": 1 00:17:57.234 }, 00:17:57.234 { 00:17:57.234 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:57.234 "dma_device_type": 2 00:17:57.234 }, 00:17:57.234 { 00:17:57.234 "dma_device_id": "system", 00:17:57.234 "dma_device_type": 1 00:17:57.234 }, 00:17:57.234 { 00:17:57.234 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:57.234 "dma_device_type": 2 00:17:57.234 } 00:17:57.234 ], 00:17:57.234 "driver_specific": { 00:17:57.234 "raid": { 00:17:57.234 "uuid": "2285c7ae-edd9-48e8-b4bd-dd2d2c3701ca", 00:17:57.234 "strip_size_kb": 64, 00:17:57.234 "state": "online", 00:17:57.234 "raid_level": "raid0", 00:17:57.234 "superblock": false, 00:17:57.234 "num_base_bdevs": 4, 00:17:57.234 "num_base_bdevs_discovered": 4, 00:17:57.234 "num_base_bdevs_operational": 4, 00:17:57.234 "base_bdevs_list": [ 00:17:57.234 { 00:17:57.234 "name": "BaseBdev1", 00:17:57.234 "uuid": "639694bd-d65a-4fd0-81e7-586f0647bf49", 00:17:57.235 "is_configured": true, 00:17:57.235 "data_offset": 0, 00:17:57.235 "data_size": 65536 00:17:57.235 }, 00:17:57.235 { 00:17:57.235 "name": "BaseBdev2", 00:17:57.235 "uuid": "6301c3e2-f974-4b66-9ec7-e663a2d0398f", 00:17:57.235 "is_configured": true, 00:17:57.235 "data_offset": 0, 00:17:57.235 "data_size": 65536 00:17:57.235 }, 00:17:57.235 { 00:17:57.235 "name": "BaseBdev3", 00:17:57.235 "uuid": "e3761302-a053-4633-8401-c2879c28e09f", 00:17:57.235 "is_configured": true, 00:17:57.235 "data_offset": 0, 00:17:57.235 "data_size": 65536 00:17:57.235 }, 00:17:57.235 { 00:17:57.235 "name": "BaseBdev4", 00:17:57.235 "uuid": "cbdc6a1e-c9e5-4232-a9f7-f2d1343aa6aa", 00:17:57.235 "is_configured": true, 00:17:57.235 "data_offset": 0, 00:17:57.235 "data_size": 65536 00:17:57.235 } 00:17:57.235 ] 00:17:57.235 } 00:17:57.235 } 00:17:57.235 }' 00:17:57.235 11:28:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:17:57.235 11:28:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:17:57.235 BaseBdev2 00:17:57.235 BaseBdev3 00:17:57.235 BaseBdev4' 00:17:57.235 11:28:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:17:57.235 11:28:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:17:57.235 11:28:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:17:57.492 11:28:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:17:57.492 "name": "BaseBdev1", 00:17:57.492 "aliases": [ 00:17:57.492 "639694bd-d65a-4fd0-81e7-586f0647bf49" 00:17:57.492 ], 00:17:57.492 "product_name": "Malloc disk", 00:17:57.492 "block_size": 512, 00:17:57.492 "num_blocks": 65536, 00:17:57.492 "uuid": "639694bd-d65a-4fd0-81e7-586f0647bf49", 00:17:57.492 "assigned_rate_limits": { 00:17:57.492 "rw_ios_per_sec": 0, 00:17:57.492 "rw_mbytes_per_sec": 0, 00:17:57.492 "r_mbytes_per_sec": 0, 00:17:57.492 "w_mbytes_per_sec": 0 00:17:57.492 }, 00:17:57.492 "claimed": true, 00:17:57.492 "claim_type": "exclusive_write", 00:17:57.492 "zoned": false, 00:17:57.492 "supported_io_types": { 00:17:57.492 "read": true, 00:17:57.492 "write": true, 00:17:57.492 "unmap": true, 00:17:57.492 "flush": true, 00:17:57.492 "reset": true, 00:17:57.492 "nvme_admin": false, 00:17:57.492 "nvme_io": false, 00:17:57.492 "nvme_io_md": false, 00:17:57.492 "write_zeroes": true, 00:17:57.492 "zcopy": true, 00:17:57.492 "get_zone_info": false, 00:17:57.492 "zone_management": false, 00:17:57.492 "zone_append": false, 00:17:57.492 "compare": false, 00:17:57.492 "compare_and_write": false, 00:17:57.492 "abort": true, 00:17:57.492 "seek_hole": false, 00:17:57.492 "seek_data": false, 00:17:57.492 "copy": true, 00:17:57.492 "nvme_iov_md": false 00:17:57.492 }, 00:17:57.492 "memory_domains": [ 00:17:57.492 { 00:17:57.492 "dma_device_id": "system", 00:17:57.492 "dma_device_type": 1 00:17:57.492 }, 00:17:57.492 { 00:17:57.492 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:57.492 "dma_device_type": 2 00:17:57.492 } 00:17:57.492 ], 00:17:57.492 "driver_specific": {} 00:17:57.492 }' 00:17:57.492 11:28:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:57.492 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:57.492 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:17:57.492 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:57.750 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:57.750 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:17:57.750 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:57.750 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:57.750 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:17:57.750 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:57.750 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:57.750 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:17:57.750 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:17:57.750 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:17:57.750 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:17:58.007 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:17:58.008 "name": "BaseBdev2", 00:17:58.008 "aliases": [ 00:17:58.008 "6301c3e2-f974-4b66-9ec7-e663a2d0398f" 00:17:58.008 ], 00:17:58.008 "product_name": "Malloc disk", 00:17:58.008 "block_size": 512, 00:17:58.008 "num_blocks": 65536, 00:17:58.008 "uuid": "6301c3e2-f974-4b66-9ec7-e663a2d0398f", 00:17:58.008 "assigned_rate_limits": { 00:17:58.008 "rw_ios_per_sec": 0, 00:17:58.008 "rw_mbytes_per_sec": 0, 00:17:58.008 "r_mbytes_per_sec": 0, 00:17:58.008 "w_mbytes_per_sec": 0 00:17:58.008 }, 00:17:58.008 "claimed": true, 00:17:58.008 "claim_type": "exclusive_write", 00:17:58.008 "zoned": false, 00:17:58.008 "supported_io_types": { 00:17:58.008 "read": true, 00:17:58.008 "write": true, 00:17:58.008 "unmap": true, 00:17:58.008 "flush": true, 00:17:58.008 "reset": true, 00:17:58.008 "nvme_admin": false, 00:17:58.008 "nvme_io": false, 00:17:58.008 "nvme_io_md": false, 00:17:58.008 "write_zeroes": true, 00:17:58.008 "zcopy": true, 00:17:58.008 "get_zone_info": false, 00:17:58.008 "zone_management": false, 00:17:58.008 "zone_append": false, 00:17:58.008 "compare": false, 00:17:58.008 "compare_and_write": false, 00:17:58.008 "abort": true, 00:17:58.008 "seek_hole": false, 00:17:58.008 "seek_data": false, 00:17:58.008 "copy": true, 00:17:58.008 "nvme_iov_md": false 00:17:58.008 }, 00:17:58.008 "memory_domains": [ 00:17:58.008 { 00:17:58.008 "dma_device_id": "system", 00:17:58.008 "dma_device_type": 1 00:17:58.008 }, 00:17:58.008 { 00:17:58.008 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:58.008 "dma_device_type": 2 00:17:58.008 } 00:17:58.008 ], 00:17:58.008 "driver_specific": {} 00:17:58.008 }' 00:17:58.008 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:58.266 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:58.266 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:17:58.266 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:58.266 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:58.266 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:17:58.266 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:58.266 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:58.266 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:17:58.266 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:58.523 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:58.524 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:17:58.524 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:17:58.524 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:17:58.524 11:28:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:17:58.780 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:17:58.780 "name": "BaseBdev3", 00:17:58.780 "aliases": [ 00:17:58.780 "e3761302-a053-4633-8401-c2879c28e09f" 00:17:58.780 ], 00:17:58.780 "product_name": "Malloc disk", 00:17:58.780 "block_size": 512, 00:17:58.780 "num_blocks": 65536, 00:17:58.780 "uuid": "e3761302-a053-4633-8401-c2879c28e09f", 00:17:58.780 "assigned_rate_limits": { 00:17:58.780 "rw_ios_per_sec": 0, 00:17:58.780 "rw_mbytes_per_sec": 0, 00:17:58.780 "r_mbytes_per_sec": 0, 00:17:58.780 "w_mbytes_per_sec": 0 00:17:58.780 }, 00:17:58.780 "claimed": true, 00:17:58.780 "claim_type": "exclusive_write", 00:17:58.780 "zoned": false, 00:17:58.780 "supported_io_types": { 00:17:58.780 "read": true, 00:17:58.780 "write": true, 00:17:58.780 "unmap": true, 00:17:58.780 "flush": true, 00:17:58.780 "reset": true, 00:17:58.780 "nvme_admin": false, 00:17:58.780 "nvme_io": false, 00:17:58.780 "nvme_io_md": false, 00:17:58.780 "write_zeroes": true, 00:17:58.780 "zcopy": true, 00:17:58.780 "get_zone_info": false, 00:17:58.780 "zone_management": false, 00:17:58.780 "zone_append": false, 00:17:58.780 "compare": false, 00:17:58.780 "compare_and_write": false, 00:17:58.780 "abort": true, 00:17:58.780 "seek_hole": false, 00:17:58.780 "seek_data": false, 00:17:58.781 "copy": true, 00:17:58.781 "nvme_iov_md": false 00:17:58.781 }, 00:17:58.781 "memory_domains": [ 00:17:58.781 { 00:17:58.781 "dma_device_id": "system", 00:17:58.781 "dma_device_type": 1 00:17:58.781 }, 00:17:58.781 { 00:17:58.781 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:58.781 "dma_device_type": 2 00:17:58.781 } 00:17:58.781 ], 00:17:58.781 "driver_specific": {} 00:17:58.781 }' 00:17:58.781 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:58.781 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:58.781 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:17:58.781 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:58.781 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:58.781 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:17:58.781 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:59.037 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:59.037 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:17:59.037 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:59.037 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:59.037 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:17:59.037 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:17:59.037 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:17:59.037 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 00:17:59.337 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:17:59.337 "name": "BaseBdev4", 00:17:59.337 "aliases": [ 00:17:59.337 "cbdc6a1e-c9e5-4232-a9f7-f2d1343aa6aa" 00:17:59.337 ], 00:17:59.337 "product_name": "Malloc disk", 00:17:59.337 "block_size": 512, 00:17:59.337 "num_blocks": 65536, 00:17:59.337 "uuid": "cbdc6a1e-c9e5-4232-a9f7-f2d1343aa6aa", 00:17:59.337 "assigned_rate_limits": { 00:17:59.337 "rw_ios_per_sec": 0, 00:17:59.337 "rw_mbytes_per_sec": 0, 00:17:59.337 "r_mbytes_per_sec": 0, 00:17:59.337 "w_mbytes_per_sec": 0 00:17:59.337 }, 00:17:59.337 "claimed": true, 00:17:59.337 "claim_type": "exclusive_write", 00:17:59.337 "zoned": false, 00:17:59.337 "supported_io_types": { 00:17:59.337 "read": true, 00:17:59.337 "write": true, 00:17:59.337 "unmap": true, 00:17:59.337 "flush": true, 00:17:59.337 "reset": true, 00:17:59.337 "nvme_admin": false, 00:17:59.337 "nvme_io": false, 00:17:59.337 "nvme_io_md": false, 00:17:59.337 "write_zeroes": true, 00:17:59.337 "zcopy": true, 00:17:59.337 "get_zone_info": false, 00:17:59.337 "zone_management": false, 00:17:59.337 "zone_append": false, 00:17:59.337 "compare": false, 00:17:59.337 "compare_and_write": false, 00:17:59.337 "abort": true, 00:17:59.337 "seek_hole": false, 00:17:59.338 "seek_data": false, 00:17:59.338 "copy": true, 00:17:59.338 "nvme_iov_md": false 00:17:59.338 }, 00:17:59.338 "memory_domains": [ 00:17:59.338 { 00:17:59.338 "dma_device_id": "system", 00:17:59.338 "dma_device_type": 1 00:17:59.338 }, 00:17:59.338 { 00:17:59.338 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:59.338 "dma_device_type": 2 00:17:59.338 } 00:17:59.338 ], 00:17:59.338 "driver_specific": {} 00:17:59.338 }' 00:17:59.338 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:59.338 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:17:59.338 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:17:59.338 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:59.338 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:17:59.611 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:17:59.611 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:59.611 11:28:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:17:59.611 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:17:59.611 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:59.611 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:17:59.611 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:17:59.611 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:17:59.867 [2024-07-15 11:28:43.347460] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:17:59.867 [2024-07-15 11:28:43.347488] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:59.867 [2024-07-15 11:28:43.347536] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:59.867 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@275 -- # local expected_state 00:17:59.867 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # has_redundancy raid0 00:17:59.867 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:17:59.867 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # return 1 00:17:59.867 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@277 -- # expected_state=offline 00:17:59.867 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:17:59.867 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:17:59.867 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=offline 00:17:59.867 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:17:59.867 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:17:59.867 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:17:59.867 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:17:59.867 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:17:59.867 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:17:59.867 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:17:59.867 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:17:59.867 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:00.125 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:00.125 "name": "Existed_Raid", 00:18:00.125 "uuid": "2285c7ae-edd9-48e8-b4bd-dd2d2c3701ca", 00:18:00.125 "strip_size_kb": 64, 00:18:00.125 "state": "offline", 00:18:00.125 "raid_level": "raid0", 00:18:00.125 "superblock": false, 00:18:00.125 "num_base_bdevs": 4, 00:18:00.125 "num_base_bdevs_discovered": 3, 00:18:00.125 "num_base_bdevs_operational": 3, 00:18:00.125 "base_bdevs_list": [ 00:18:00.125 { 00:18:00.125 "name": null, 00:18:00.125 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:00.125 "is_configured": false, 00:18:00.125 "data_offset": 0, 00:18:00.125 "data_size": 65536 00:18:00.125 }, 00:18:00.125 { 00:18:00.125 "name": "BaseBdev2", 00:18:00.125 "uuid": "6301c3e2-f974-4b66-9ec7-e663a2d0398f", 00:18:00.125 "is_configured": true, 00:18:00.125 "data_offset": 0, 00:18:00.125 "data_size": 65536 00:18:00.125 }, 00:18:00.125 { 00:18:00.125 "name": "BaseBdev3", 00:18:00.125 "uuid": "e3761302-a053-4633-8401-c2879c28e09f", 00:18:00.125 "is_configured": true, 00:18:00.125 "data_offset": 0, 00:18:00.125 "data_size": 65536 00:18:00.125 }, 00:18:00.125 { 00:18:00.125 "name": "BaseBdev4", 00:18:00.125 "uuid": "cbdc6a1e-c9e5-4232-a9f7-f2d1343aa6aa", 00:18:00.125 "is_configured": true, 00:18:00.125 "data_offset": 0, 00:18:00.125 "data_size": 65536 00:18:00.125 } 00:18:00.125 ] 00:18:00.125 }' 00:18:00.125 11:28:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:00.125 11:28:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:18:00.690 11:28:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:18:00.690 11:28:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:18:00.690 11:28:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:00.690 11:28:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:18:00.948 11:28:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:18:00.948 11:28:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:18:00.948 11:28:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:18:01.206 [2024-07-15 11:28:44.696945] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:18:01.206 11:28:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:18:01.206 11:28:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:18:01.206 11:28:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:01.206 11:28:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:18:01.463 11:28:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:18:01.463 11:28:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:18:01.463 11:28:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev3 00:18:01.721 [2024-07-15 11:28:45.198650] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:18:01.721 11:28:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:18:01.721 11:28:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:18:01.721 11:28:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:01.721 11:28:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:18:01.979 11:28:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:18:01.979 11:28:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:18:01.979 11:28:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev4 00:18:02.237 [2024-07-15 11:28:45.700340] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:18:02.237 [2024-07-15 11:28:45.700380] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1b2e350 name Existed_Raid, state offline 00:18:02.237 11:28:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:18:02.237 11:28:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:18:02.237 11:28:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:02.237 11:28:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:18:02.494 11:28:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:18:02.494 11:28:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:18:02.494 11:28:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # '[' 4 -gt 2 ']' 00:18:02.494 11:28:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i = 1 )) 00:18:02.494 11:28:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:18:02.494 11:28:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:18:02.752 BaseBdev2 00:18:02.752 11:28:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev2 00:18:02.752 11:28:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:18:02.752 11:28:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:18:02.752 11:28:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:18:02.752 11:28:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:18:02.752 11:28:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:18:02.752 11:28:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:18:03.011 11:28:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:18:03.269 [ 00:18:03.269 { 00:18:03.269 "name": "BaseBdev2", 00:18:03.269 "aliases": [ 00:18:03.269 "aa0fd5ff-176d-4f83-a000-c2e8baa60bb8" 00:18:03.269 ], 00:18:03.269 "product_name": "Malloc disk", 00:18:03.269 "block_size": 512, 00:18:03.269 "num_blocks": 65536, 00:18:03.269 "uuid": "aa0fd5ff-176d-4f83-a000-c2e8baa60bb8", 00:18:03.269 "assigned_rate_limits": { 00:18:03.269 "rw_ios_per_sec": 0, 00:18:03.269 "rw_mbytes_per_sec": 0, 00:18:03.269 "r_mbytes_per_sec": 0, 00:18:03.269 "w_mbytes_per_sec": 0 00:18:03.269 }, 00:18:03.269 "claimed": false, 00:18:03.269 "zoned": false, 00:18:03.269 "supported_io_types": { 00:18:03.269 "read": true, 00:18:03.269 "write": true, 00:18:03.269 "unmap": true, 00:18:03.269 "flush": true, 00:18:03.269 "reset": true, 00:18:03.269 "nvme_admin": false, 00:18:03.269 "nvme_io": false, 00:18:03.269 "nvme_io_md": false, 00:18:03.269 "write_zeroes": true, 00:18:03.269 "zcopy": true, 00:18:03.269 "get_zone_info": false, 00:18:03.269 "zone_management": false, 00:18:03.269 "zone_append": false, 00:18:03.269 "compare": false, 00:18:03.269 "compare_and_write": false, 00:18:03.269 "abort": true, 00:18:03.269 "seek_hole": false, 00:18:03.269 "seek_data": false, 00:18:03.269 "copy": true, 00:18:03.269 "nvme_iov_md": false 00:18:03.269 }, 00:18:03.269 "memory_domains": [ 00:18:03.269 { 00:18:03.269 "dma_device_id": "system", 00:18:03.269 "dma_device_type": 1 00:18:03.269 }, 00:18:03.269 { 00:18:03.269 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:03.269 "dma_device_type": 2 00:18:03.269 } 00:18:03.269 ], 00:18:03.269 "driver_specific": {} 00:18:03.269 } 00:18:03.269 ] 00:18:03.269 11:28:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:18:03.269 11:28:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:18:03.269 11:28:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:18:03.269 11:28:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:18:03.527 BaseBdev3 00:18:03.527 11:28:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev3 00:18:03.527 11:28:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:18:03.527 11:28:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:18:03.527 11:28:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:18:03.527 11:28:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:18:03.527 11:28:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:18:03.527 11:28:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:18:03.784 11:28:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:18:04.041 [ 00:18:04.041 { 00:18:04.041 "name": "BaseBdev3", 00:18:04.041 "aliases": [ 00:18:04.041 "7b11f59a-e862-4e6f-8a0f-49a374d923d8" 00:18:04.041 ], 00:18:04.041 "product_name": "Malloc disk", 00:18:04.041 "block_size": 512, 00:18:04.041 "num_blocks": 65536, 00:18:04.041 "uuid": "7b11f59a-e862-4e6f-8a0f-49a374d923d8", 00:18:04.041 "assigned_rate_limits": { 00:18:04.041 "rw_ios_per_sec": 0, 00:18:04.041 "rw_mbytes_per_sec": 0, 00:18:04.041 "r_mbytes_per_sec": 0, 00:18:04.041 "w_mbytes_per_sec": 0 00:18:04.041 }, 00:18:04.041 "claimed": false, 00:18:04.041 "zoned": false, 00:18:04.041 "supported_io_types": { 00:18:04.041 "read": true, 00:18:04.041 "write": true, 00:18:04.041 "unmap": true, 00:18:04.041 "flush": true, 00:18:04.041 "reset": true, 00:18:04.041 "nvme_admin": false, 00:18:04.041 "nvme_io": false, 00:18:04.041 "nvme_io_md": false, 00:18:04.041 "write_zeroes": true, 00:18:04.041 "zcopy": true, 00:18:04.041 "get_zone_info": false, 00:18:04.041 "zone_management": false, 00:18:04.041 "zone_append": false, 00:18:04.041 "compare": false, 00:18:04.041 "compare_and_write": false, 00:18:04.041 "abort": true, 00:18:04.041 "seek_hole": false, 00:18:04.041 "seek_data": false, 00:18:04.041 "copy": true, 00:18:04.041 "nvme_iov_md": false 00:18:04.041 }, 00:18:04.041 "memory_domains": [ 00:18:04.041 { 00:18:04.041 "dma_device_id": "system", 00:18:04.041 "dma_device_type": 1 00:18:04.041 }, 00:18:04.041 { 00:18:04.041 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:04.041 "dma_device_type": 2 00:18:04.041 } 00:18:04.041 ], 00:18:04.041 "driver_specific": {} 00:18:04.041 } 00:18:04.041 ] 00:18:04.041 11:28:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:18:04.041 11:28:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:18:04.041 11:28:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:18:04.041 11:28:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4 00:18:04.298 BaseBdev4 00:18:04.298 11:28:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev4 00:18:04.298 11:28:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev4 00:18:04.298 11:28:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:18:04.298 11:28:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:18:04.298 11:28:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:18:04.298 11:28:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:18:04.298 11:28:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:18:04.555 11:28:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 -t 2000 00:18:04.555 [ 00:18:04.555 { 00:18:04.555 "name": "BaseBdev4", 00:18:04.555 "aliases": [ 00:18:04.555 "ce7487de-abd9-458b-90a4-bca08d769e8f" 00:18:04.555 ], 00:18:04.555 "product_name": "Malloc disk", 00:18:04.555 "block_size": 512, 00:18:04.555 "num_blocks": 65536, 00:18:04.555 "uuid": "ce7487de-abd9-458b-90a4-bca08d769e8f", 00:18:04.555 "assigned_rate_limits": { 00:18:04.555 "rw_ios_per_sec": 0, 00:18:04.555 "rw_mbytes_per_sec": 0, 00:18:04.555 "r_mbytes_per_sec": 0, 00:18:04.555 "w_mbytes_per_sec": 0 00:18:04.555 }, 00:18:04.555 "claimed": false, 00:18:04.555 "zoned": false, 00:18:04.555 "supported_io_types": { 00:18:04.555 "read": true, 00:18:04.555 "write": true, 00:18:04.555 "unmap": true, 00:18:04.555 "flush": true, 00:18:04.555 "reset": true, 00:18:04.555 "nvme_admin": false, 00:18:04.555 "nvme_io": false, 00:18:04.555 "nvme_io_md": false, 00:18:04.555 "write_zeroes": true, 00:18:04.555 "zcopy": true, 00:18:04.555 "get_zone_info": false, 00:18:04.555 "zone_management": false, 00:18:04.555 "zone_append": false, 00:18:04.555 "compare": false, 00:18:04.555 "compare_and_write": false, 00:18:04.555 "abort": true, 00:18:04.555 "seek_hole": false, 00:18:04.555 "seek_data": false, 00:18:04.555 "copy": true, 00:18:04.555 "nvme_iov_md": false 00:18:04.555 }, 00:18:04.555 "memory_domains": [ 00:18:04.555 { 00:18:04.555 "dma_device_id": "system", 00:18:04.555 "dma_device_type": 1 00:18:04.555 }, 00:18:04.555 { 00:18:04.555 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:04.555 "dma_device_type": 2 00:18:04.555 } 00:18:04.555 ], 00:18:04.555 "driver_specific": {} 00:18:04.555 } 00:18:04.555 ] 00:18:04.813 11:28:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:18:04.813 11:28:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:18:04.813 11:28:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:18:04.813 11:28:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@305 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:18:04.813 [2024-07-15 11:28:48.385777] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:18:04.813 [2024-07-15 11:28:48.385819] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:18:04.813 [2024-07-15 11:28:48.385840] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:18:04.813 [2024-07-15 11:28:48.387212] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:18:04.813 [2024-07-15 11:28:48.387254] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:18:04.813 11:28:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:18:04.813 11:28:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:04.813 11:28:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:18:04.814 11:28:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:04.814 11:28:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:04.814 11:28:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:04.814 11:28:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:04.814 11:28:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:04.814 11:28:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:04.814 11:28:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:05.071 11:28:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:05.071 11:28:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:05.071 11:28:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:05.071 "name": "Existed_Raid", 00:18:05.071 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:05.071 "strip_size_kb": 64, 00:18:05.071 "state": "configuring", 00:18:05.071 "raid_level": "raid0", 00:18:05.071 "superblock": false, 00:18:05.071 "num_base_bdevs": 4, 00:18:05.071 "num_base_bdevs_discovered": 3, 00:18:05.071 "num_base_bdevs_operational": 4, 00:18:05.071 "base_bdevs_list": [ 00:18:05.071 { 00:18:05.071 "name": "BaseBdev1", 00:18:05.071 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:05.071 "is_configured": false, 00:18:05.071 "data_offset": 0, 00:18:05.071 "data_size": 0 00:18:05.071 }, 00:18:05.071 { 00:18:05.071 "name": "BaseBdev2", 00:18:05.071 "uuid": "aa0fd5ff-176d-4f83-a000-c2e8baa60bb8", 00:18:05.071 "is_configured": true, 00:18:05.071 "data_offset": 0, 00:18:05.071 "data_size": 65536 00:18:05.071 }, 00:18:05.071 { 00:18:05.071 "name": "BaseBdev3", 00:18:05.071 "uuid": "7b11f59a-e862-4e6f-8a0f-49a374d923d8", 00:18:05.071 "is_configured": true, 00:18:05.071 "data_offset": 0, 00:18:05.071 "data_size": 65536 00:18:05.072 }, 00:18:05.072 { 00:18:05.072 "name": "BaseBdev4", 00:18:05.072 "uuid": "ce7487de-abd9-458b-90a4-bca08d769e8f", 00:18:05.072 "is_configured": true, 00:18:05.072 "data_offset": 0, 00:18:05.072 "data_size": 65536 00:18:05.072 } 00:18:05.072 ] 00:18:05.072 }' 00:18:05.072 11:28:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:05.072 11:28:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:18:06.004 11:28:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev2 00:18:06.004 [2024-07-15 11:28:49.476635] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:18:06.004 11:28:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@309 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:18:06.004 11:28:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:06.004 11:28:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:18:06.004 11:28:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:06.004 11:28:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:06.004 11:28:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:06.004 11:28:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:06.004 11:28:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:06.004 11:28:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:06.004 11:28:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:06.004 11:28:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:06.004 11:28:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:06.262 11:28:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:06.262 "name": "Existed_Raid", 00:18:06.262 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:06.262 "strip_size_kb": 64, 00:18:06.262 "state": "configuring", 00:18:06.262 "raid_level": "raid0", 00:18:06.262 "superblock": false, 00:18:06.262 "num_base_bdevs": 4, 00:18:06.262 "num_base_bdevs_discovered": 2, 00:18:06.262 "num_base_bdevs_operational": 4, 00:18:06.262 "base_bdevs_list": [ 00:18:06.262 { 00:18:06.262 "name": "BaseBdev1", 00:18:06.262 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:06.262 "is_configured": false, 00:18:06.262 "data_offset": 0, 00:18:06.262 "data_size": 0 00:18:06.262 }, 00:18:06.262 { 00:18:06.262 "name": null, 00:18:06.262 "uuid": "aa0fd5ff-176d-4f83-a000-c2e8baa60bb8", 00:18:06.262 "is_configured": false, 00:18:06.262 "data_offset": 0, 00:18:06.262 "data_size": 65536 00:18:06.262 }, 00:18:06.262 { 00:18:06.262 "name": "BaseBdev3", 00:18:06.262 "uuid": "7b11f59a-e862-4e6f-8a0f-49a374d923d8", 00:18:06.262 "is_configured": true, 00:18:06.262 "data_offset": 0, 00:18:06.262 "data_size": 65536 00:18:06.262 }, 00:18:06.262 { 00:18:06.262 "name": "BaseBdev4", 00:18:06.262 "uuid": "ce7487de-abd9-458b-90a4-bca08d769e8f", 00:18:06.262 "is_configured": true, 00:18:06.262 "data_offset": 0, 00:18:06.262 "data_size": 65536 00:18:06.262 } 00:18:06.262 ] 00:18:06.262 }' 00:18:06.262 11:28:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:06.262 11:28:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:18:06.827 11:28:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:06.827 11:28:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:18:07.084 11:28:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # [[ false == \f\a\l\s\e ]] 00:18:07.084 11:28:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:18:07.341 [2024-07-15 11:28:50.795474] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:18:07.341 BaseBdev1 00:18:07.341 11:28:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@313 -- # waitforbdev BaseBdev1 00:18:07.341 11:28:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:18:07.341 11:28:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:18:07.341 11:28:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:18:07.341 11:28:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:18:07.341 11:28:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:18:07.341 11:28:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:18:07.599 11:28:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:18:07.856 [ 00:18:07.856 { 00:18:07.856 "name": "BaseBdev1", 00:18:07.856 "aliases": [ 00:18:07.856 "16b9fd03-45a3-4cad-89b7-3beea2623805" 00:18:07.856 ], 00:18:07.856 "product_name": "Malloc disk", 00:18:07.856 "block_size": 512, 00:18:07.856 "num_blocks": 65536, 00:18:07.856 "uuid": "16b9fd03-45a3-4cad-89b7-3beea2623805", 00:18:07.856 "assigned_rate_limits": { 00:18:07.856 "rw_ios_per_sec": 0, 00:18:07.856 "rw_mbytes_per_sec": 0, 00:18:07.856 "r_mbytes_per_sec": 0, 00:18:07.856 "w_mbytes_per_sec": 0 00:18:07.856 }, 00:18:07.856 "claimed": true, 00:18:07.856 "claim_type": "exclusive_write", 00:18:07.856 "zoned": false, 00:18:07.856 "supported_io_types": { 00:18:07.856 "read": true, 00:18:07.856 "write": true, 00:18:07.856 "unmap": true, 00:18:07.856 "flush": true, 00:18:07.856 "reset": true, 00:18:07.856 "nvme_admin": false, 00:18:07.856 "nvme_io": false, 00:18:07.856 "nvme_io_md": false, 00:18:07.856 "write_zeroes": true, 00:18:07.856 "zcopy": true, 00:18:07.856 "get_zone_info": false, 00:18:07.856 "zone_management": false, 00:18:07.856 "zone_append": false, 00:18:07.856 "compare": false, 00:18:07.856 "compare_and_write": false, 00:18:07.856 "abort": true, 00:18:07.856 "seek_hole": false, 00:18:07.856 "seek_data": false, 00:18:07.856 "copy": true, 00:18:07.856 "nvme_iov_md": false 00:18:07.856 }, 00:18:07.856 "memory_domains": [ 00:18:07.856 { 00:18:07.856 "dma_device_id": "system", 00:18:07.856 "dma_device_type": 1 00:18:07.856 }, 00:18:07.856 { 00:18:07.856 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:07.856 "dma_device_type": 2 00:18:07.856 } 00:18:07.856 ], 00:18:07.856 "driver_specific": {} 00:18:07.856 } 00:18:07.856 ] 00:18:07.856 11:28:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:18:07.856 11:28:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:18:07.856 11:28:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:07.856 11:28:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:18:07.856 11:28:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:07.856 11:28:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:07.856 11:28:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:07.856 11:28:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:07.856 11:28:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:07.856 11:28:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:07.856 11:28:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:07.856 11:28:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:07.856 11:28:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:08.113 11:28:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:08.113 "name": "Existed_Raid", 00:18:08.113 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:08.113 "strip_size_kb": 64, 00:18:08.113 "state": "configuring", 00:18:08.113 "raid_level": "raid0", 00:18:08.113 "superblock": false, 00:18:08.113 "num_base_bdevs": 4, 00:18:08.113 "num_base_bdevs_discovered": 3, 00:18:08.113 "num_base_bdevs_operational": 4, 00:18:08.113 "base_bdevs_list": [ 00:18:08.113 { 00:18:08.113 "name": "BaseBdev1", 00:18:08.113 "uuid": "16b9fd03-45a3-4cad-89b7-3beea2623805", 00:18:08.113 "is_configured": true, 00:18:08.113 "data_offset": 0, 00:18:08.113 "data_size": 65536 00:18:08.113 }, 00:18:08.113 { 00:18:08.113 "name": null, 00:18:08.113 "uuid": "aa0fd5ff-176d-4f83-a000-c2e8baa60bb8", 00:18:08.113 "is_configured": false, 00:18:08.113 "data_offset": 0, 00:18:08.113 "data_size": 65536 00:18:08.113 }, 00:18:08.113 { 00:18:08.113 "name": "BaseBdev3", 00:18:08.113 "uuid": "7b11f59a-e862-4e6f-8a0f-49a374d923d8", 00:18:08.113 "is_configured": true, 00:18:08.113 "data_offset": 0, 00:18:08.113 "data_size": 65536 00:18:08.113 }, 00:18:08.113 { 00:18:08.113 "name": "BaseBdev4", 00:18:08.113 "uuid": "ce7487de-abd9-458b-90a4-bca08d769e8f", 00:18:08.113 "is_configured": true, 00:18:08.113 "data_offset": 0, 00:18:08.113 "data_size": 65536 00:18:08.113 } 00:18:08.113 ] 00:18:08.113 }' 00:18:08.113 11:28:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:08.113 11:28:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:18:08.677 11:28:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:08.677 11:28:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:18:08.935 11:28:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # [[ true == \t\r\u\e ]] 00:18:08.935 11:28:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@317 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev3 00:18:09.191 [2024-07-15 11:28:52.620337] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:18:09.191 11:28:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:18:09.191 11:28:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:09.191 11:28:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:18:09.191 11:28:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:09.191 11:28:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:09.191 11:28:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:09.191 11:28:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:09.191 11:28:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:09.191 11:28:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:09.192 11:28:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:09.192 11:28:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:09.192 11:28:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:09.448 11:28:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:09.448 "name": "Existed_Raid", 00:18:09.448 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:09.448 "strip_size_kb": 64, 00:18:09.448 "state": "configuring", 00:18:09.448 "raid_level": "raid0", 00:18:09.448 "superblock": false, 00:18:09.448 "num_base_bdevs": 4, 00:18:09.448 "num_base_bdevs_discovered": 2, 00:18:09.448 "num_base_bdevs_operational": 4, 00:18:09.448 "base_bdevs_list": [ 00:18:09.448 { 00:18:09.448 "name": "BaseBdev1", 00:18:09.448 "uuid": "16b9fd03-45a3-4cad-89b7-3beea2623805", 00:18:09.448 "is_configured": true, 00:18:09.448 "data_offset": 0, 00:18:09.448 "data_size": 65536 00:18:09.448 }, 00:18:09.448 { 00:18:09.448 "name": null, 00:18:09.448 "uuid": "aa0fd5ff-176d-4f83-a000-c2e8baa60bb8", 00:18:09.448 "is_configured": false, 00:18:09.448 "data_offset": 0, 00:18:09.448 "data_size": 65536 00:18:09.448 }, 00:18:09.448 { 00:18:09.448 "name": null, 00:18:09.448 "uuid": "7b11f59a-e862-4e6f-8a0f-49a374d923d8", 00:18:09.448 "is_configured": false, 00:18:09.448 "data_offset": 0, 00:18:09.448 "data_size": 65536 00:18:09.448 }, 00:18:09.448 { 00:18:09.448 "name": "BaseBdev4", 00:18:09.448 "uuid": "ce7487de-abd9-458b-90a4-bca08d769e8f", 00:18:09.448 "is_configured": true, 00:18:09.448 "data_offset": 0, 00:18:09.448 "data_size": 65536 00:18:09.448 } 00:18:09.448 ] 00:18:09.448 }' 00:18:09.448 11:28:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:09.448 11:28:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:18:10.015 11:28:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:10.015 11:28:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:18:10.272 11:28:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # [[ false == \f\a\l\s\e ]] 00:18:10.272 11:28:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:18:10.530 [2024-07-15 11:28:53.955912] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:18:10.530 11:28:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@322 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:18:10.530 11:28:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:10.530 11:28:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:18:10.530 11:28:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:10.530 11:28:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:10.530 11:28:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:10.530 11:28:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:10.530 11:28:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:10.530 11:28:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:10.530 11:28:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:10.530 11:28:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:10.530 11:28:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:10.788 11:28:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:10.788 "name": "Existed_Raid", 00:18:10.788 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:10.788 "strip_size_kb": 64, 00:18:10.788 "state": "configuring", 00:18:10.788 "raid_level": "raid0", 00:18:10.788 "superblock": false, 00:18:10.788 "num_base_bdevs": 4, 00:18:10.788 "num_base_bdevs_discovered": 3, 00:18:10.788 "num_base_bdevs_operational": 4, 00:18:10.788 "base_bdevs_list": [ 00:18:10.788 { 00:18:10.788 "name": "BaseBdev1", 00:18:10.788 "uuid": "16b9fd03-45a3-4cad-89b7-3beea2623805", 00:18:10.788 "is_configured": true, 00:18:10.788 "data_offset": 0, 00:18:10.788 "data_size": 65536 00:18:10.788 }, 00:18:10.788 { 00:18:10.788 "name": null, 00:18:10.788 "uuid": "aa0fd5ff-176d-4f83-a000-c2e8baa60bb8", 00:18:10.788 "is_configured": false, 00:18:10.788 "data_offset": 0, 00:18:10.788 "data_size": 65536 00:18:10.788 }, 00:18:10.788 { 00:18:10.788 "name": "BaseBdev3", 00:18:10.788 "uuid": "7b11f59a-e862-4e6f-8a0f-49a374d923d8", 00:18:10.788 "is_configured": true, 00:18:10.788 "data_offset": 0, 00:18:10.788 "data_size": 65536 00:18:10.788 }, 00:18:10.788 { 00:18:10.788 "name": "BaseBdev4", 00:18:10.788 "uuid": "ce7487de-abd9-458b-90a4-bca08d769e8f", 00:18:10.788 "is_configured": true, 00:18:10.788 "data_offset": 0, 00:18:10.788 "data_size": 65536 00:18:10.788 } 00:18:10.788 ] 00:18:10.788 }' 00:18:10.788 11:28:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:10.788 11:28:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:18:11.352 11:28:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:18:11.352 11:28:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:11.610 11:28:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # [[ true == \t\r\u\e ]] 00:18:11.610 11:28:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@325 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:18:11.867 [2024-07-15 11:28:55.219288] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:18:11.867 11:28:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:18:11.867 11:28:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:11.867 11:28:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:18:11.867 11:28:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:11.867 11:28:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:11.867 11:28:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:11.867 11:28:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:11.867 11:28:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:11.867 11:28:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:11.868 11:28:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:11.868 11:28:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:11.868 11:28:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:12.125 11:28:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:12.125 "name": "Existed_Raid", 00:18:12.125 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:12.125 "strip_size_kb": 64, 00:18:12.125 "state": "configuring", 00:18:12.125 "raid_level": "raid0", 00:18:12.125 "superblock": false, 00:18:12.125 "num_base_bdevs": 4, 00:18:12.125 "num_base_bdevs_discovered": 2, 00:18:12.125 "num_base_bdevs_operational": 4, 00:18:12.125 "base_bdevs_list": [ 00:18:12.125 { 00:18:12.125 "name": null, 00:18:12.125 "uuid": "16b9fd03-45a3-4cad-89b7-3beea2623805", 00:18:12.125 "is_configured": false, 00:18:12.125 "data_offset": 0, 00:18:12.125 "data_size": 65536 00:18:12.125 }, 00:18:12.125 { 00:18:12.125 "name": null, 00:18:12.125 "uuid": "aa0fd5ff-176d-4f83-a000-c2e8baa60bb8", 00:18:12.125 "is_configured": false, 00:18:12.125 "data_offset": 0, 00:18:12.125 "data_size": 65536 00:18:12.125 }, 00:18:12.125 { 00:18:12.125 "name": "BaseBdev3", 00:18:12.125 "uuid": "7b11f59a-e862-4e6f-8a0f-49a374d923d8", 00:18:12.125 "is_configured": true, 00:18:12.125 "data_offset": 0, 00:18:12.125 "data_size": 65536 00:18:12.125 }, 00:18:12.125 { 00:18:12.125 "name": "BaseBdev4", 00:18:12.125 "uuid": "ce7487de-abd9-458b-90a4-bca08d769e8f", 00:18:12.125 "is_configured": true, 00:18:12.125 "data_offset": 0, 00:18:12.125 "data_size": 65536 00:18:12.125 } 00:18:12.125 ] 00:18:12.125 }' 00:18:12.125 11:28:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:12.125 11:28:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:18:12.690 11:28:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@327 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:12.690 11:28:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@327 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:18:12.948 11:28:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@327 -- # [[ false == \f\a\l\s\e ]] 00:18:12.948 11:28:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@329 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:18:13.240 [2024-07-15 11:28:56.581695] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:18:13.240 11:28:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@330 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:18:13.240 11:28:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:13.240 11:28:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:18:13.240 11:28:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:13.240 11:28:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:13.240 11:28:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:13.240 11:28:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:13.240 11:28:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:13.240 11:28:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:13.240 11:28:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:13.240 11:28:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:13.240 11:28:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:13.499 11:28:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:13.499 "name": "Existed_Raid", 00:18:13.499 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:13.499 "strip_size_kb": 64, 00:18:13.499 "state": "configuring", 00:18:13.499 "raid_level": "raid0", 00:18:13.499 "superblock": false, 00:18:13.499 "num_base_bdevs": 4, 00:18:13.499 "num_base_bdevs_discovered": 3, 00:18:13.499 "num_base_bdevs_operational": 4, 00:18:13.499 "base_bdevs_list": [ 00:18:13.499 { 00:18:13.499 "name": null, 00:18:13.499 "uuid": "16b9fd03-45a3-4cad-89b7-3beea2623805", 00:18:13.499 "is_configured": false, 00:18:13.499 "data_offset": 0, 00:18:13.499 "data_size": 65536 00:18:13.499 }, 00:18:13.499 { 00:18:13.499 "name": "BaseBdev2", 00:18:13.499 "uuid": "aa0fd5ff-176d-4f83-a000-c2e8baa60bb8", 00:18:13.499 "is_configured": true, 00:18:13.499 "data_offset": 0, 00:18:13.499 "data_size": 65536 00:18:13.499 }, 00:18:13.499 { 00:18:13.499 "name": "BaseBdev3", 00:18:13.499 "uuid": "7b11f59a-e862-4e6f-8a0f-49a374d923d8", 00:18:13.499 "is_configured": true, 00:18:13.499 "data_offset": 0, 00:18:13.499 "data_size": 65536 00:18:13.499 }, 00:18:13.499 { 00:18:13.499 "name": "BaseBdev4", 00:18:13.499 "uuid": "ce7487de-abd9-458b-90a4-bca08d769e8f", 00:18:13.499 "is_configured": true, 00:18:13.499 "data_offset": 0, 00:18:13.499 "data_size": 65536 00:18:13.499 } 00:18:13.499 ] 00:18:13.499 }' 00:18:13.499 11:28:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:13.499 11:28:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:18:14.068 11:28:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@331 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:14.068 11:28:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@331 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:18:14.327 11:28:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@331 -- # [[ true == \t\r\u\e ]] 00:18:14.327 11:28:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:14.327 11:28:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@333 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:18:14.586 11:28:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b NewBaseBdev -u 16b9fd03-45a3-4cad-89b7-3beea2623805 00:18:14.586 [2024-07-15 11:28:58.158442] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:18:14.586 [2024-07-15 11:28:58.158481] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1b32040 00:18:14.586 [2024-07-15 11:28:58.158490] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:18:14.586 [2024-07-15 11:28:58.158685] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1b2da70 00:18:14.586 [2024-07-15 11:28:58.158800] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1b32040 00:18:14.586 [2024-07-15 11:28:58.158809] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x1b32040 00:18:14.586 [2024-07-15 11:28:58.158985] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:18:14.586 NewBaseBdev 00:18:14.586 11:28:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@334 -- # waitforbdev NewBaseBdev 00:18:14.586 11:28:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=NewBaseBdev 00:18:14.586 11:28:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:18:14.586 11:28:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:18:14.586 11:28:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:18:14.586 11:28:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:18:14.586 11:28:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:18:15.154 11:28:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev -t 2000 00:18:15.721 [ 00:18:15.721 { 00:18:15.721 "name": "NewBaseBdev", 00:18:15.721 "aliases": [ 00:18:15.721 "16b9fd03-45a3-4cad-89b7-3beea2623805" 00:18:15.721 ], 00:18:15.721 "product_name": "Malloc disk", 00:18:15.721 "block_size": 512, 00:18:15.721 "num_blocks": 65536, 00:18:15.721 "uuid": "16b9fd03-45a3-4cad-89b7-3beea2623805", 00:18:15.721 "assigned_rate_limits": { 00:18:15.721 "rw_ios_per_sec": 0, 00:18:15.721 "rw_mbytes_per_sec": 0, 00:18:15.721 "r_mbytes_per_sec": 0, 00:18:15.721 "w_mbytes_per_sec": 0 00:18:15.721 }, 00:18:15.721 "claimed": true, 00:18:15.721 "claim_type": "exclusive_write", 00:18:15.721 "zoned": false, 00:18:15.721 "supported_io_types": { 00:18:15.721 "read": true, 00:18:15.721 "write": true, 00:18:15.721 "unmap": true, 00:18:15.721 "flush": true, 00:18:15.721 "reset": true, 00:18:15.721 "nvme_admin": false, 00:18:15.721 "nvme_io": false, 00:18:15.721 "nvme_io_md": false, 00:18:15.721 "write_zeroes": true, 00:18:15.721 "zcopy": true, 00:18:15.721 "get_zone_info": false, 00:18:15.721 "zone_management": false, 00:18:15.721 "zone_append": false, 00:18:15.721 "compare": false, 00:18:15.721 "compare_and_write": false, 00:18:15.721 "abort": true, 00:18:15.721 "seek_hole": false, 00:18:15.721 "seek_data": false, 00:18:15.721 "copy": true, 00:18:15.721 "nvme_iov_md": false 00:18:15.721 }, 00:18:15.721 "memory_domains": [ 00:18:15.721 { 00:18:15.721 "dma_device_id": "system", 00:18:15.721 "dma_device_type": 1 00:18:15.721 }, 00:18:15.721 { 00:18:15.721 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:15.721 "dma_device_type": 2 00:18:15.721 } 00:18:15.721 ], 00:18:15.721 "driver_specific": {} 00:18:15.721 } 00:18:15.721 ] 00:18:15.721 11:28:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:18:15.721 11:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@335 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:18:15.721 11:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:15.721 11:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:18:15.721 11:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:15.721 11:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:15.721 11:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:15.721 11:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:15.721 11:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:15.721 11:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:15.721 11:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:15.721 11:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:15.722 11:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:15.980 11:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:15.980 "name": "Existed_Raid", 00:18:15.980 "uuid": "ed94918f-9c72-4956-a510-e6315aacae53", 00:18:15.980 "strip_size_kb": 64, 00:18:15.980 "state": "online", 00:18:15.980 "raid_level": "raid0", 00:18:15.980 "superblock": false, 00:18:15.980 "num_base_bdevs": 4, 00:18:15.980 "num_base_bdevs_discovered": 4, 00:18:15.980 "num_base_bdevs_operational": 4, 00:18:15.980 "base_bdevs_list": [ 00:18:15.980 { 00:18:15.980 "name": "NewBaseBdev", 00:18:15.980 "uuid": "16b9fd03-45a3-4cad-89b7-3beea2623805", 00:18:15.980 "is_configured": true, 00:18:15.980 "data_offset": 0, 00:18:15.980 "data_size": 65536 00:18:15.980 }, 00:18:15.980 { 00:18:15.980 "name": "BaseBdev2", 00:18:15.980 "uuid": "aa0fd5ff-176d-4f83-a000-c2e8baa60bb8", 00:18:15.980 "is_configured": true, 00:18:15.980 "data_offset": 0, 00:18:15.980 "data_size": 65536 00:18:15.980 }, 00:18:15.980 { 00:18:15.980 "name": "BaseBdev3", 00:18:15.980 "uuid": "7b11f59a-e862-4e6f-8a0f-49a374d923d8", 00:18:15.980 "is_configured": true, 00:18:15.980 "data_offset": 0, 00:18:15.980 "data_size": 65536 00:18:15.980 }, 00:18:15.980 { 00:18:15.980 "name": "BaseBdev4", 00:18:15.980 "uuid": "ce7487de-abd9-458b-90a4-bca08d769e8f", 00:18:15.980 "is_configured": true, 00:18:15.980 "data_offset": 0, 00:18:15.980 "data_size": 65536 00:18:15.980 } 00:18:15.980 ] 00:18:15.980 }' 00:18:15.980 11:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:15.980 11:28:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:18:16.546 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@336 -- # verify_raid_bdev_properties Existed_Raid 00:18:16.546 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:18:16.546 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:18:16.546 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:18:16.546 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:18:16.546 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # local name 00:18:16.546 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:18:16.546 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:18:16.804 [2024-07-15 11:29:00.256375] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:18:16.804 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:18:16.804 "name": "Existed_Raid", 00:18:16.804 "aliases": [ 00:18:16.804 "ed94918f-9c72-4956-a510-e6315aacae53" 00:18:16.804 ], 00:18:16.804 "product_name": "Raid Volume", 00:18:16.804 "block_size": 512, 00:18:16.804 "num_blocks": 262144, 00:18:16.804 "uuid": "ed94918f-9c72-4956-a510-e6315aacae53", 00:18:16.804 "assigned_rate_limits": { 00:18:16.804 "rw_ios_per_sec": 0, 00:18:16.804 "rw_mbytes_per_sec": 0, 00:18:16.804 "r_mbytes_per_sec": 0, 00:18:16.804 "w_mbytes_per_sec": 0 00:18:16.804 }, 00:18:16.804 "claimed": false, 00:18:16.804 "zoned": false, 00:18:16.804 "supported_io_types": { 00:18:16.804 "read": true, 00:18:16.804 "write": true, 00:18:16.804 "unmap": true, 00:18:16.804 "flush": true, 00:18:16.804 "reset": true, 00:18:16.804 "nvme_admin": false, 00:18:16.804 "nvme_io": false, 00:18:16.804 "nvme_io_md": false, 00:18:16.804 "write_zeroes": true, 00:18:16.804 "zcopy": false, 00:18:16.804 "get_zone_info": false, 00:18:16.804 "zone_management": false, 00:18:16.804 "zone_append": false, 00:18:16.804 "compare": false, 00:18:16.804 "compare_and_write": false, 00:18:16.804 "abort": false, 00:18:16.804 "seek_hole": false, 00:18:16.804 "seek_data": false, 00:18:16.804 "copy": false, 00:18:16.804 "nvme_iov_md": false 00:18:16.804 }, 00:18:16.804 "memory_domains": [ 00:18:16.804 { 00:18:16.804 "dma_device_id": "system", 00:18:16.804 "dma_device_type": 1 00:18:16.804 }, 00:18:16.804 { 00:18:16.804 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:16.804 "dma_device_type": 2 00:18:16.804 }, 00:18:16.804 { 00:18:16.804 "dma_device_id": "system", 00:18:16.804 "dma_device_type": 1 00:18:16.804 }, 00:18:16.804 { 00:18:16.804 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:16.804 "dma_device_type": 2 00:18:16.804 }, 00:18:16.804 { 00:18:16.804 "dma_device_id": "system", 00:18:16.804 "dma_device_type": 1 00:18:16.804 }, 00:18:16.804 { 00:18:16.804 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:16.804 "dma_device_type": 2 00:18:16.804 }, 00:18:16.804 { 00:18:16.804 "dma_device_id": "system", 00:18:16.804 "dma_device_type": 1 00:18:16.804 }, 00:18:16.804 { 00:18:16.804 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:16.804 "dma_device_type": 2 00:18:16.804 } 00:18:16.804 ], 00:18:16.804 "driver_specific": { 00:18:16.804 "raid": { 00:18:16.804 "uuid": "ed94918f-9c72-4956-a510-e6315aacae53", 00:18:16.804 "strip_size_kb": 64, 00:18:16.804 "state": "online", 00:18:16.804 "raid_level": "raid0", 00:18:16.804 "superblock": false, 00:18:16.804 "num_base_bdevs": 4, 00:18:16.804 "num_base_bdevs_discovered": 4, 00:18:16.804 "num_base_bdevs_operational": 4, 00:18:16.804 "base_bdevs_list": [ 00:18:16.804 { 00:18:16.804 "name": "NewBaseBdev", 00:18:16.804 "uuid": "16b9fd03-45a3-4cad-89b7-3beea2623805", 00:18:16.804 "is_configured": true, 00:18:16.804 "data_offset": 0, 00:18:16.804 "data_size": 65536 00:18:16.804 }, 00:18:16.804 { 00:18:16.804 "name": "BaseBdev2", 00:18:16.804 "uuid": "aa0fd5ff-176d-4f83-a000-c2e8baa60bb8", 00:18:16.804 "is_configured": true, 00:18:16.804 "data_offset": 0, 00:18:16.804 "data_size": 65536 00:18:16.804 }, 00:18:16.804 { 00:18:16.804 "name": "BaseBdev3", 00:18:16.804 "uuid": "7b11f59a-e862-4e6f-8a0f-49a374d923d8", 00:18:16.804 "is_configured": true, 00:18:16.804 "data_offset": 0, 00:18:16.805 "data_size": 65536 00:18:16.805 }, 00:18:16.805 { 00:18:16.805 "name": "BaseBdev4", 00:18:16.805 "uuid": "ce7487de-abd9-458b-90a4-bca08d769e8f", 00:18:16.805 "is_configured": true, 00:18:16.805 "data_offset": 0, 00:18:16.805 "data_size": 65536 00:18:16.805 } 00:18:16.805 ] 00:18:16.805 } 00:18:16.805 } 00:18:16.805 }' 00:18:16.805 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:18:16.805 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='NewBaseBdev 00:18:16.805 BaseBdev2 00:18:16.805 BaseBdev3 00:18:16.805 BaseBdev4' 00:18:16.805 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:18:16.805 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev 00:18:16.805 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:18:17.063 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:18:17.063 "name": "NewBaseBdev", 00:18:17.063 "aliases": [ 00:18:17.063 "16b9fd03-45a3-4cad-89b7-3beea2623805" 00:18:17.063 ], 00:18:17.063 "product_name": "Malloc disk", 00:18:17.063 "block_size": 512, 00:18:17.063 "num_blocks": 65536, 00:18:17.063 "uuid": "16b9fd03-45a3-4cad-89b7-3beea2623805", 00:18:17.063 "assigned_rate_limits": { 00:18:17.063 "rw_ios_per_sec": 0, 00:18:17.063 "rw_mbytes_per_sec": 0, 00:18:17.063 "r_mbytes_per_sec": 0, 00:18:17.063 "w_mbytes_per_sec": 0 00:18:17.063 }, 00:18:17.063 "claimed": true, 00:18:17.063 "claim_type": "exclusive_write", 00:18:17.063 "zoned": false, 00:18:17.063 "supported_io_types": { 00:18:17.063 "read": true, 00:18:17.063 "write": true, 00:18:17.063 "unmap": true, 00:18:17.063 "flush": true, 00:18:17.063 "reset": true, 00:18:17.063 "nvme_admin": false, 00:18:17.063 "nvme_io": false, 00:18:17.063 "nvme_io_md": false, 00:18:17.063 "write_zeroes": true, 00:18:17.063 "zcopy": true, 00:18:17.063 "get_zone_info": false, 00:18:17.063 "zone_management": false, 00:18:17.063 "zone_append": false, 00:18:17.063 "compare": false, 00:18:17.063 "compare_and_write": false, 00:18:17.063 "abort": true, 00:18:17.063 "seek_hole": false, 00:18:17.063 "seek_data": false, 00:18:17.063 "copy": true, 00:18:17.063 "nvme_iov_md": false 00:18:17.063 }, 00:18:17.063 "memory_domains": [ 00:18:17.063 { 00:18:17.063 "dma_device_id": "system", 00:18:17.063 "dma_device_type": 1 00:18:17.063 }, 00:18:17.063 { 00:18:17.063 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:17.063 "dma_device_type": 2 00:18:17.063 } 00:18:17.063 ], 00:18:17.063 "driver_specific": {} 00:18:17.063 }' 00:18:17.063 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:17.063 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:17.321 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:18:17.321 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:17.321 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:17.321 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:18:17.321 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:17.321 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:17.321 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:18:17.321 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:17.321 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:17.578 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:18:17.578 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:18:17.578 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:18:17.578 11:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:18:17.835 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:18:17.835 "name": "BaseBdev2", 00:18:17.835 "aliases": [ 00:18:17.835 "aa0fd5ff-176d-4f83-a000-c2e8baa60bb8" 00:18:17.835 ], 00:18:17.835 "product_name": "Malloc disk", 00:18:17.835 "block_size": 512, 00:18:17.835 "num_blocks": 65536, 00:18:17.835 "uuid": "aa0fd5ff-176d-4f83-a000-c2e8baa60bb8", 00:18:17.835 "assigned_rate_limits": { 00:18:17.835 "rw_ios_per_sec": 0, 00:18:17.835 "rw_mbytes_per_sec": 0, 00:18:17.835 "r_mbytes_per_sec": 0, 00:18:17.835 "w_mbytes_per_sec": 0 00:18:17.835 }, 00:18:17.835 "claimed": true, 00:18:17.835 "claim_type": "exclusive_write", 00:18:17.835 "zoned": false, 00:18:17.835 "supported_io_types": { 00:18:17.835 "read": true, 00:18:17.835 "write": true, 00:18:17.835 "unmap": true, 00:18:17.835 "flush": true, 00:18:17.835 "reset": true, 00:18:17.835 "nvme_admin": false, 00:18:17.836 "nvme_io": false, 00:18:17.836 "nvme_io_md": false, 00:18:17.836 "write_zeroes": true, 00:18:17.836 "zcopy": true, 00:18:17.836 "get_zone_info": false, 00:18:17.836 "zone_management": false, 00:18:17.836 "zone_append": false, 00:18:17.836 "compare": false, 00:18:17.836 "compare_and_write": false, 00:18:17.836 "abort": true, 00:18:17.836 "seek_hole": false, 00:18:17.836 "seek_data": false, 00:18:17.836 "copy": true, 00:18:17.836 "nvme_iov_md": false 00:18:17.836 }, 00:18:17.836 "memory_domains": [ 00:18:17.836 { 00:18:17.836 "dma_device_id": "system", 00:18:17.836 "dma_device_type": 1 00:18:17.836 }, 00:18:17.836 { 00:18:17.836 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:17.836 "dma_device_type": 2 00:18:17.836 } 00:18:17.836 ], 00:18:17.836 "driver_specific": {} 00:18:17.836 }' 00:18:17.836 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:17.836 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:17.836 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:18:17.836 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:17.836 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:17.836 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:18:17.836 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:17.836 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:18.093 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:18:18.093 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:18.093 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:18.093 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:18:18.093 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:18:18.093 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:18:18.093 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:18:18.351 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:18:18.351 "name": "BaseBdev3", 00:18:18.351 "aliases": [ 00:18:18.351 "7b11f59a-e862-4e6f-8a0f-49a374d923d8" 00:18:18.351 ], 00:18:18.351 "product_name": "Malloc disk", 00:18:18.351 "block_size": 512, 00:18:18.351 "num_blocks": 65536, 00:18:18.351 "uuid": "7b11f59a-e862-4e6f-8a0f-49a374d923d8", 00:18:18.351 "assigned_rate_limits": { 00:18:18.351 "rw_ios_per_sec": 0, 00:18:18.351 "rw_mbytes_per_sec": 0, 00:18:18.351 "r_mbytes_per_sec": 0, 00:18:18.351 "w_mbytes_per_sec": 0 00:18:18.351 }, 00:18:18.351 "claimed": true, 00:18:18.351 "claim_type": "exclusive_write", 00:18:18.351 "zoned": false, 00:18:18.351 "supported_io_types": { 00:18:18.351 "read": true, 00:18:18.351 "write": true, 00:18:18.351 "unmap": true, 00:18:18.351 "flush": true, 00:18:18.351 "reset": true, 00:18:18.351 "nvme_admin": false, 00:18:18.351 "nvme_io": false, 00:18:18.351 "nvme_io_md": false, 00:18:18.351 "write_zeroes": true, 00:18:18.351 "zcopy": true, 00:18:18.351 "get_zone_info": false, 00:18:18.351 "zone_management": false, 00:18:18.351 "zone_append": false, 00:18:18.351 "compare": false, 00:18:18.351 "compare_and_write": false, 00:18:18.351 "abort": true, 00:18:18.351 "seek_hole": false, 00:18:18.351 "seek_data": false, 00:18:18.351 "copy": true, 00:18:18.351 "nvme_iov_md": false 00:18:18.351 }, 00:18:18.351 "memory_domains": [ 00:18:18.351 { 00:18:18.351 "dma_device_id": "system", 00:18:18.351 "dma_device_type": 1 00:18:18.351 }, 00:18:18.351 { 00:18:18.351 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:18.351 "dma_device_type": 2 00:18:18.351 } 00:18:18.351 ], 00:18:18.351 "driver_specific": {} 00:18:18.351 }' 00:18:18.351 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:18.351 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:18.351 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:18:18.351 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:18.351 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:18.351 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:18:18.351 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:18.609 11:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:18.609 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:18:18.609 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:18.609 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:18.609 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:18:18.609 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:18:18.609 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:18:18.609 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 00:18:18.865 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:18:18.865 "name": "BaseBdev4", 00:18:18.865 "aliases": [ 00:18:18.865 "ce7487de-abd9-458b-90a4-bca08d769e8f" 00:18:18.865 ], 00:18:18.865 "product_name": "Malloc disk", 00:18:18.865 "block_size": 512, 00:18:18.865 "num_blocks": 65536, 00:18:18.865 "uuid": "ce7487de-abd9-458b-90a4-bca08d769e8f", 00:18:18.865 "assigned_rate_limits": { 00:18:18.865 "rw_ios_per_sec": 0, 00:18:18.865 "rw_mbytes_per_sec": 0, 00:18:18.865 "r_mbytes_per_sec": 0, 00:18:18.865 "w_mbytes_per_sec": 0 00:18:18.865 }, 00:18:18.865 "claimed": true, 00:18:18.865 "claim_type": "exclusive_write", 00:18:18.865 "zoned": false, 00:18:18.865 "supported_io_types": { 00:18:18.865 "read": true, 00:18:18.865 "write": true, 00:18:18.865 "unmap": true, 00:18:18.865 "flush": true, 00:18:18.865 "reset": true, 00:18:18.865 "nvme_admin": false, 00:18:18.865 "nvme_io": false, 00:18:18.865 "nvme_io_md": false, 00:18:18.865 "write_zeroes": true, 00:18:18.865 "zcopy": true, 00:18:18.865 "get_zone_info": false, 00:18:18.865 "zone_management": false, 00:18:18.865 "zone_append": false, 00:18:18.865 "compare": false, 00:18:18.865 "compare_and_write": false, 00:18:18.865 "abort": true, 00:18:18.865 "seek_hole": false, 00:18:18.865 "seek_data": false, 00:18:18.865 "copy": true, 00:18:18.865 "nvme_iov_md": false 00:18:18.865 }, 00:18:18.865 "memory_domains": [ 00:18:18.865 { 00:18:18.865 "dma_device_id": "system", 00:18:18.865 "dma_device_type": 1 00:18:18.865 }, 00:18:18.865 { 00:18:18.865 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:18.865 "dma_device_type": 2 00:18:18.865 } 00:18:18.865 ], 00:18:18.865 "driver_specific": {} 00:18:18.865 }' 00:18:18.865 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:18.865 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:18.865 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:18:19.122 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:19.122 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:19.122 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:18:19.122 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:19.122 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:19.122 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:18:19.122 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:19.122 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:19.379 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:18:19.379 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@338 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:18:19.379 [2024-07-15 11:29:02.947190] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:18:19.379 [2024-07-15 11:29:02.947216] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:18:19.379 [2024-07-15 11:29:02.947267] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:18:19.379 [2024-07-15 11:29:02.947324] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:18:19.379 [2024-07-15 11:29:02.947336] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1b32040 name Existed_Raid, state offline 00:18:19.379 11:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@341 -- # killprocess 922789 00:18:19.379 11:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@948 -- # '[' -z 922789 ']' 00:18:19.379 11:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@952 -- # kill -0 922789 00:18:19.379 11:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@953 -- # uname 00:18:19.636 11:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:18:19.636 11:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 922789 00:18:19.636 11:29:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:18:19.636 11:29:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:18:19.636 11:29:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 922789' 00:18:19.636 killing process with pid 922789 00:18:19.636 11:29:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@967 -- # kill 922789 00:18:19.636 [2024-07-15 11:29:03.019648] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:18:19.637 11:29:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # wait 922789 00:18:19.637 [2024-07-15 11:29:03.058511] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@343 -- # return 0 00:18:19.895 00:18:19.895 real 0m33.252s 00:18:19.895 user 1m0.922s 00:18:19.895 sys 0m5.989s 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:18:19.895 ************************************ 00:18:19.895 END TEST raid_state_function_test 00:18:19.895 ************************************ 00:18:19.895 11:29:03 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:18:19.895 11:29:03 bdev_raid -- bdev/bdev_raid.sh@868 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 4 true 00:18:19.895 11:29:03 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:18:19.895 11:29:03 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:18:19.895 11:29:03 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:18:19.895 ************************************ 00:18:19.895 START TEST raid_state_function_test_sb 00:18:19.895 ************************************ 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1123 -- # raid_state_function_test raid0 4 true 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@220 -- # local raid_level=raid0 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=4 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # local superblock=true 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev3 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev4 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # local strip_size 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # '[' raid0 '!=' raid1 ']' 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # strip_size=64 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@232 -- # strip_size_create_arg='-z 64' 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # '[' true = true ']' 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@238 -- # superblock_create_arg=-s 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # raid_pid=927846 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 927846' 00:18:19.895 Process raid pid: 927846 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@246 -- # waitforlisten 927846 /var/tmp/spdk-raid.sock 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@829 -- # '[' -z 927846 ']' 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@834 -- # local max_retries=100 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:18:19.895 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # xtrace_disable 00:18:19.895 11:29:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:18:19.895 [2024-07-15 11:29:03.437806] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:18:19.895 [2024-07-15 11:29:03.437875] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:18:20.153 [2024-07-15 11:29:03.568171] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:20.153 [2024-07-15 11:29:03.671051] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:18:20.153 [2024-07-15 11:29:03.730322] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:18:20.153 [2024-07-15 11:29:03.730359] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:18:21.088 11:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:18:21.088 11:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@862 -- # return 0 00:18:21.088 11:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r raid0 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:18:21.088 [2024-07-15 11:29:04.602848] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:18:21.088 [2024-07-15 11:29:04.602891] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:18:21.088 [2024-07-15 11:29:04.602902] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:18:21.088 [2024-07-15 11:29:04.602915] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:18:21.088 [2024-07-15 11:29:04.602924] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:18:21.088 [2024-07-15 11:29:04.602943] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:18:21.088 [2024-07-15 11:29:04.602953] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:18:21.088 [2024-07-15 11:29:04.602963] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:18:21.088 11:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:18:21.088 11:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:21.088 11:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:18:21.088 11:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:21.088 11:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:21.088 11:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:21.088 11:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:21.088 11:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:21.088 11:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:21.088 11:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:21.088 11:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:21.088 11:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:21.347 11:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:21.347 "name": "Existed_Raid", 00:18:21.347 "uuid": "d7b8e7e1-5a13-405d-8ad6-048acec3a927", 00:18:21.347 "strip_size_kb": 64, 00:18:21.347 "state": "configuring", 00:18:21.347 "raid_level": "raid0", 00:18:21.347 "superblock": true, 00:18:21.347 "num_base_bdevs": 4, 00:18:21.347 "num_base_bdevs_discovered": 0, 00:18:21.347 "num_base_bdevs_operational": 4, 00:18:21.347 "base_bdevs_list": [ 00:18:21.347 { 00:18:21.347 "name": "BaseBdev1", 00:18:21.347 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:21.347 "is_configured": false, 00:18:21.347 "data_offset": 0, 00:18:21.347 "data_size": 0 00:18:21.347 }, 00:18:21.347 { 00:18:21.347 "name": "BaseBdev2", 00:18:21.347 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:21.347 "is_configured": false, 00:18:21.347 "data_offset": 0, 00:18:21.347 "data_size": 0 00:18:21.347 }, 00:18:21.347 { 00:18:21.347 "name": "BaseBdev3", 00:18:21.347 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:21.347 "is_configured": false, 00:18:21.347 "data_offset": 0, 00:18:21.347 "data_size": 0 00:18:21.347 }, 00:18:21.347 { 00:18:21.347 "name": "BaseBdev4", 00:18:21.347 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:21.347 "is_configured": false, 00:18:21.347 "data_offset": 0, 00:18:21.347 "data_size": 0 00:18:21.347 } 00:18:21.347 ] 00:18:21.347 }' 00:18:21.347 11:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:21.347 11:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:18:21.914 11:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:18:22.172 [2024-07-15 11:29:05.673520] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:18:22.172 [2024-07-15 11:29:05.673552] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xe82aa0 name Existed_Raid, state configuring 00:18:22.172 11:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r raid0 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:18:22.430 [2024-07-15 11:29:05.906158] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:18:22.430 [2024-07-15 11:29:05.906188] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:18:22.430 [2024-07-15 11:29:05.906198] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:18:22.430 [2024-07-15 11:29:05.906209] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:18:22.430 [2024-07-15 11:29:05.906218] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:18:22.430 [2024-07-15 11:29:05.906229] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:18:22.430 [2024-07-15 11:29:05.906238] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:18:22.430 [2024-07-15 11:29:05.906248] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:18:22.430 11:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:18:22.689 [2024-07-15 11:29:06.161860] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:18:22.689 BaseBdev1 00:18:22.689 11:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:18:22.689 11:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:18:22.689 11:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:18:22.689 11:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:18:22.689 11:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:18:22.689 11:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:18:22.689 11:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:18:22.948 11:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:18:23.207 [ 00:18:23.207 { 00:18:23.207 "name": "BaseBdev1", 00:18:23.207 "aliases": [ 00:18:23.207 "32a7dcbe-d5ad-41c5-be5b-82a3c01d26a4" 00:18:23.207 ], 00:18:23.207 "product_name": "Malloc disk", 00:18:23.207 "block_size": 512, 00:18:23.207 "num_blocks": 65536, 00:18:23.207 "uuid": "32a7dcbe-d5ad-41c5-be5b-82a3c01d26a4", 00:18:23.207 "assigned_rate_limits": { 00:18:23.207 "rw_ios_per_sec": 0, 00:18:23.207 "rw_mbytes_per_sec": 0, 00:18:23.207 "r_mbytes_per_sec": 0, 00:18:23.207 "w_mbytes_per_sec": 0 00:18:23.207 }, 00:18:23.207 "claimed": true, 00:18:23.207 "claim_type": "exclusive_write", 00:18:23.207 "zoned": false, 00:18:23.207 "supported_io_types": { 00:18:23.207 "read": true, 00:18:23.207 "write": true, 00:18:23.207 "unmap": true, 00:18:23.207 "flush": true, 00:18:23.207 "reset": true, 00:18:23.207 "nvme_admin": false, 00:18:23.207 "nvme_io": false, 00:18:23.207 "nvme_io_md": false, 00:18:23.207 "write_zeroes": true, 00:18:23.207 "zcopy": true, 00:18:23.207 "get_zone_info": false, 00:18:23.207 "zone_management": false, 00:18:23.207 "zone_append": false, 00:18:23.207 "compare": false, 00:18:23.207 "compare_and_write": false, 00:18:23.207 "abort": true, 00:18:23.207 "seek_hole": false, 00:18:23.207 "seek_data": false, 00:18:23.207 "copy": true, 00:18:23.207 "nvme_iov_md": false 00:18:23.207 }, 00:18:23.207 "memory_domains": [ 00:18:23.207 { 00:18:23.207 "dma_device_id": "system", 00:18:23.207 "dma_device_type": 1 00:18:23.207 }, 00:18:23.207 { 00:18:23.207 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:23.207 "dma_device_type": 2 00:18:23.207 } 00:18:23.207 ], 00:18:23.207 "driver_specific": {} 00:18:23.207 } 00:18:23.207 ] 00:18:23.207 11:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:18:23.207 11:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:18:23.207 11:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:23.207 11:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:18:23.207 11:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:23.207 11:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:23.207 11:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:23.207 11:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:23.207 11:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:23.207 11:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:23.207 11:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:23.207 11:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:23.207 11:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:23.465 11:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:23.465 "name": "Existed_Raid", 00:18:23.465 "uuid": "6fc68515-8c63-4f0e-9272-a5000aa1dc7b", 00:18:23.465 "strip_size_kb": 64, 00:18:23.465 "state": "configuring", 00:18:23.465 "raid_level": "raid0", 00:18:23.465 "superblock": true, 00:18:23.465 "num_base_bdevs": 4, 00:18:23.465 "num_base_bdevs_discovered": 1, 00:18:23.465 "num_base_bdevs_operational": 4, 00:18:23.465 "base_bdevs_list": [ 00:18:23.465 { 00:18:23.465 "name": "BaseBdev1", 00:18:23.465 "uuid": "32a7dcbe-d5ad-41c5-be5b-82a3c01d26a4", 00:18:23.465 "is_configured": true, 00:18:23.465 "data_offset": 2048, 00:18:23.465 "data_size": 63488 00:18:23.465 }, 00:18:23.465 { 00:18:23.465 "name": "BaseBdev2", 00:18:23.465 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:23.465 "is_configured": false, 00:18:23.465 "data_offset": 0, 00:18:23.465 "data_size": 0 00:18:23.465 }, 00:18:23.465 { 00:18:23.465 "name": "BaseBdev3", 00:18:23.465 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:23.465 "is_configured": false, 00:18:23.465 "data_offset": 0, 00:18:23.465 "data_size": 0 00:18:23.465 }, 00:18:23.465 { 00:18:23.465 "name": "BaseBdev4", 00:18:23.465 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:23.465 "is_configured": false, 00:18:23.465 "data_offset": 0, 00:18:23.465 "data_size": 0 00:18:23.465 } 00:18:23.465 ] 00:18:23.465 }' 00:18:23.465 11:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:23.465 11:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:18:24.030 11:29:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:18:24.288 [2024-07-15 11:29:07.697936] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:18:24.288 [2024-07-15 11:29:07.697974] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xe82310 name Existed_Raid, state configuring 00:18:24.288 11:29:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r raid0 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:18:24.288 [2024-07-15 11:29:07.874445] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:18:24.288 [2024-07-15 11:29:07.875878] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:18:24.288 [2024-07-15 11:29:07.875910] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:18:24.288 [2024-07-15 11:29:07.875920] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:18:24.288 [2024-07-15 11:29:07.875940] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:18:24.288 [2024-07-15 11:29:07.875949] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:18:24.288 [2024-07-15 11:29:07.875960] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:18:24.547 11:29:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:18:24.547 11:29:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:18:24.547 11:29:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:18:24.547 11:29:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:24.547 11:29:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:18:24.547 11:29:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:24.547 11:29:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:24.547 11:29:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:24.547 11:29:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:24.547 11:29:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:24.547 11:29:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:24.547 11:29:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:24.547 11:29:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:24.547 11:29:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:24.547 11:29:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:24.547 "name": "Existed_Raid", 00:18:24.547 "uuid": "88cfb57e-f525-4bc6-abcd-1d53e1b1ed94", 00:18:24.547 "strip_size_kb": 64, 00:18:24.547 "state": "configuring", 00:18:24.547 "raid_level": "raid0", 00:18:24.547 "superblock": true, 00:18:24.547 "num_base_bdevs": 4, 00:18:24.547 "num_base_bdevs_discovered": 1, 00:18:24.547 "num_base_bdevs_operational": 4, 00:18:24.547 "base_bdevs_list": [ 00:18:24.547 { 00:18:24.547 "name": "BaseBdev1", 00:18:24.547 "uuid": "32a7dcbe-d5ad-41c5-be5b-82a3c01d26a4", 00:18:24.547 "is_configured": true, 00:18:24.547 "data_offset": 2048, 00:18:24.547 "data_size": 63488 00:18:24.547 }, 00:18:24.547 { 00:18:24.547 "name": "BaseBdev2", 00:18:24.547 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:24.547 "is_configured": false, 00:18:24.547 "data_offset": 0, 00:18:24.547 "data_size": 0 00:18:24.547 }, 00:18:24.547 { 00:18:24.547 "name": "BaseBdev3", 00:18:24.547 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:24.547 "is_configured": false, 00:18:24.547 "data_offset": 0, 00:18:24.547 "data_size": 0 00:18:24.547 }, 00:18:24.547 { 00:18:24.547 "name": "BaseBdev4", 00:18:24.547 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:24.547 "is_configured": false, 00:18:24.547 "data_offset": 0, 00:18:24.547 "data_size": 0 00:18:24.547 } 00:18:24.547 ] 00:18:24.547 }' 00:18:24.547 11:29:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:24.547 11:29:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:18:25.113 11:29:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:18:25.372 [2024-07-15 11:29:08.884452] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:18:25.372 BaseBdev2 00:18:25.372 11:29:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:18:25.372 11:29:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:18:25.372 11:29:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:18:25.372 11:29:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:18:25.372 11:29:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:18:25.372 11:29:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:18:25.372 11:29:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:18:25.630 11:29:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:18:25.888 [ 00:18:25.888 { 00:18:25.888 "name": "BaseBdev2", 00:18:25.888 "aliases": [ 00:18:25.888 "3827b760-0b14-4f31-b7b8-6b47cd1ac326" 00:18:25.888 ], 00:18:25.888 "product_name": "Malloc disk", 00:18:25.888 "block_size": 512, 00:18:25.888 "num_blocks": 65536, 00:18:25.888 "uuid": "3827b760-0b14-4f31-b7b8-6b47cd1ac326", 00:18:25.888 "assigned_rate_limits": { 00:18:25.888 "rw_ios_per_sec": 0, 00:18:25.888 "rw_mbytes_per_sec": 0, 00:18:25.888 "r_mbytes_per_sec": 0, 00:18:25.889 "w_mbytes_per_sec": 0 00:18:25.889 }, 00:18:25.889 "claimed": true, 00:18:25.889 "claim_type": "exclusive_write", 00:18:25.889 "zoned": false, 00:18:25.889 "supported_io_types": { 00:18:25.889 "read": true, 00:18:25.889 "write": true, 00:18:25.889 "unmap": true, 00:18:25.889 "flush": true, 00:18:25.889 "reset": true, 00:18:25.889 "nvme_admin": false, 00:18:25.889 "nvme_io": false, 00:18:25.889 "nvme_io_md": false, 00:18:25.889 "write_zeroes": true, 00:18:25.889 "zcopy": true, 00:18:25.889 "get_zone_info": false, 00:18:25.889 "zone_management": false, 00:18:25.889 "zone_append": false, 00:18:25.889 "compare": false, 00:18:25.889 "compare_and_write": false, 00:18:25.889 "abort": true, 00:18:25.889 "seek_hole": false, 00:18:25.889 "seek_data": false, 00:18:25.889 "copy": true, 00:18:25.889 "nvme_iov_md": false 00:18:25.889 }, 00:18:25.889 "memory_domains": [ 00:18:25.889 { 00:18:25.889 "dma_device_id": "system", 00:18:25.889 "dma_device_type": 1 00:18:25.889 }, 00:18:25.889 { 00:18:25.889 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:25.889 "dma_device_type": 2 00:18:25.889 } 00:18:25.889 ], 00:18:25.889 "driver_specific": {} 00:18:25.889 } 00:18:25.889 ] 00:18:25.889 11:29:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:18:25.889 11:29:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:18:25.889 11:29:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:18:25.889 11:29:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:18:25.889 11:29:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:25.889 11:29:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:18:25.889 11:29:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:25.889 11:29:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:25.889 11:29:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:25.889 11:29:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:25.889 11:29:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:25.889 11:29:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:25.889 11:29:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:25.889 11:29:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:25.889 11:29:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:25.889 11:29:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:25.889 "name": "Existed_Raid", 00:18:25.889 "uuid": "88cfb57e-f525-4bc6-abcd-1d53e1b1ed94", 00:18:25.889 "strip_size_kb": 64, 00:18:25.889 "state": "configuring", 00:18:25.889 "raid_level": "raid0", 00:18:25.889 "superblock": true, 00:18:25.889 "num_base_bdevs": 4, 00:18:25.889 "num_base_bdevs_discovered": 2, 00:18:25.889 "num_base_bdevs_operational": 4, 00:18:25.889 "base_bdevs_list": [ 00:18:25.889 { 00:18:25.889 "name": "BaseBdev1", 00:18:25.889 "uuid": "32a7dcbe-d5ad-41c5-be5b-82a3c01d26a4", 00:18:25.889 "is_configured": true, 00:18:25.889 "data_offset": 2048, 00:18:25.889 "data_size": 63488 00:18:25.889 }, 00:18:25.889 { 00:18:25.889 "name": "BaseBdev2", 00:18:25.889 "uuid": "3827b760-0b14-4f31-b7b8-6b47cd1ac326", 00:18:25.889 "is_configured": true, 00:18:25.889 "data_offset": 2048, 00:18:25.889 "data_size": 63488 00:18:25.889 }, 00:18:25.889 { 00:18:25.889 "name": "BaseBdev3", 00:18:25.889 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:25.889 "is_configured": false, 00:18:25.889 "data_offset": 0, 00:18:25.889 "data_size": 0 00:18:25.889 }, 00:18:25.889 { 00:18:25.889 "name": "BaseBdev4", 00:18:25.889 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:25.889 "is_configured": false, 00:18:25.889 "data_offset": 0, 00:18:25.889 "data_size": 0 00:18:25.889 } 00:18:25.889 ] 00:18:25.889 }' 00:18:25.889 11:29:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:25.889 11:29:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:18:26.456 11:29:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:18:26.714 [2024-07-15 11:29:10.271619] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:18:26.714 BaseBdev3 00:18:26.714 11:29:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev3 00:18:26.714 11:29:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:18:26.714 11:29:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:18:26.714 11:29:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:18:26.714 11:29:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:18:26.714 11:29:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:18:26.714 11:29:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:18:26.972 11:29:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:18:27.230 [ 00:18:27.230 { 00:18:27.230 "name": "BaseBdev3", 00:18:27.230 "aliases": [ 00:18:27.230 "02aa1a62-8717-41dd-b4c4-6cb58c3c41d6" 00:18:27.230 ], 00:18:27.230 "product_name": "Malloc disk", 00:18:27.230 "block_size": 512, 00:18:27.230 "num_blocks": 65536, 00:18:27.230 "uuid": "02aa1a62-8717-41dd-b4c4-6cb58c3c41d6", 00:18:27.230 "assigned_rate_limits": { 00:18:27.230 "rw_ios_per_sec": 0, 00:18:27.230 "rw_mbytes_per_sec": 0, 00:18:27.230 "r_mbytes_per_sec": 0, 00:18:27.230 "w_mbytes_per_sec": 0 00:18:27.230 }, 00:18:27.230 "claimed": true, 00:18:27.230 "claim_type": "exclusive_write", 00:18:27.230 "zoned": false, 00:18:27.230 "supported_io_types": { 00:18:27.230 "read": true, 00:18:27.230 "write": true, 00:18:27.230 "unmap": true, 00:18:27.230 "flush": true, 00:18:27.230 "reset": true, 00:18:27.230 "nvme_admin": false, 00:18:27.230 "nvme_io": false, 00:18:27.230 "nvme_io_md": false, 00:18:27.230 "write_zeroes": true, 00:18:27.230 "zcopy": true, 00:18:27.230 "get_zone_info": false, 00:18:27.230 "zone_management": false, 00:18:27.230 "zone_append": false, 00:18:27.230 "compare": false, 00:18:27.230 "compare_and_write": false, 00:18:27.230 "abort": true, 00:18:27.230 "seek_hole": false, 00:18:27.230 "seek_data": false, 00:18:27.230 "copy": true, 00:18:27.230 "nvme_iov_md": false 00:18:27.230 }, 00:18:27.230 "memory_domains": [ 00:18:27.230 { 00:18:27.230 "dma_device_id": "system", 00:18:27.230 "dma_device_type": 1 00:18:27.230 }, 00:18:27.230 { 00:18:27.230 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:27.230 "dma_device_type": 2 00:18:27.230 } 00:18:27.230 ], 00:18:27.230 "driver_specific": {} 00:18:27.230 } 00:18:27.230 ] 00:18:27.230 11:29:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:18:27.230 11:29:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:18:27.230 11:29:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:18:27.230 11:29:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:18:27.230 11:29:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:27.230 11:29:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:18:27.230 11:29:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:27.230 11:29:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:27.230 11:29:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:27.230 11:29:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:27.230 11:29:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:27.230 11:29:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:27.230 11:29:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:27.230 11:29:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:27.230 11:29:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:27.490 11:29:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:27.490 "name": "Existed_Raid", 00:18:27.490 "uuid": "88cfb57e-f525-4bc6-abcd-1d53e1b1ed94", 00:18:27.490 "strip_size_kb": 64, 00:18:27.490 "state": "configuring", 00:18:27.490 "raid_level": "raid0", 00:18:27.490 "superblock": true, 00:18:27.490 "num_base_bdevs": 4, 00:18:27.490 "num_base_bdevs_discovered": 3, 00:18:27.490 "num_base_bdevs_operational": 4, 00:18:27.490 "base_bdevs_list": [ 00:18:27.490 { 00:18:27.490 "name": "BaseBdev1", 00:18:27.490 "uuid": "32a7dcbe-d5ad-41c5-be5b-82a3c01d26a4", 00:18:27.490 "is_configured": true, 00:18:27.490 "data_offset": 2048, 00:18:27.490 "data_size": 63488 00:18:27.490 }, 00:18:27.490 { 00:18:27.490 "name": "BaseBdev2", 00:18:27.490 "uuid": "3827b760-0b14-4f31-b7b8-6b47cd1ac326", 00:18:27.490 "is_configured": true, 00:18:27.490 "data_offset": 2048, 00:18:27.490 "data_size": 63488 00:18:27.490 }, 00:18:27.490 { 00:18:27.490 "name": "BaseBdev3", 00:18:27.490 "uuid": "02aa1a62-8717-41dd-b4c4-6cb58c3c41d6", 00:18:27.490 "is_configured": true, 00:18:27.490 "data_offset": 2048, 00:18:27.490 "data_size": 63488 00:18:27.490 }, 00:18:27.490 { 00:18:27.490 "name": "BaseBdev4", 00:18:27.490 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:27.490 "is_configured": false, 00:18:27.490 "data_offset": 0, 00:18:27.490 "data_size": 0 00:18:27.490 } 00:18:27.490 ] 00:18:27.490 }' 00:18:27.490 11:29:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:27.490 11:29:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:18:28.054 11:29:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4 00:18:28.311 [2024-07-15 11:29:11.835170] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:18:28.311 [2024-07-15 11:29:11.835340] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xe83350 00:18:28.311 [2024-07-15 11:29:11.835354] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:18:28.311 [2024-07-15 11:29:11.835529] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xe83020 00:18:28.311 [2024-07-15 11:29:11.835645] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xe83350 00:18:28.311 [2024-07-15 11:29:11.835660] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0xe83350 00:18:28.312 [2024-07-15 11:29:11.835748] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:18:28.312 BaseBdev4 00:18:28.312 11:29:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev4 00:18:28.312 11:29:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev4 00:18:28.312 11:29:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:18:28.312 11:29:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:18:28.312 11:29:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:18:28.312 11:29:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:18:28.312 11:29:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:18:28.570 11:29:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 -t 2000 00:18:28.828 [ 00:18:28.828 { 00:18:28.828 "name": "BaseBdev4", 00:18:28.828 "aliases": [ 00:18:28.828 "15b70ca5-b0c3-4f5c-b8b7-dd0fb0e712c6" 00:18:28.828 ], 00:18:28.828 "product_name": "Malloc disk", 00:18:28.828 "block_size": 512, 00:18:28.828 "num_blocks": 65536, 00:18:28.828 "uuid": "15b70ca5-b0c3-4f5c-b8b7-dd0fb0e712c6", 00:18:28.828 "assigned_rate_limits": { 00:18:28.828 "rw_ios_per_sec": 0, 00:18:28.828 "rw_mbytes_per_sec": 0, 00:18:28.828 "r_mbytes_per_sec": 0, 00:18:28.828 "w_mbytes_per_sec": 0 00:18:28.828 }, 00:18:28.828 "claimed": true, 00:18:28.828 "claim_type": "exclusive_write", 00:18:28.828 "zoned": false, 00:18:28.828 "supported_io_types": { 00:18:28.828 "read": true, 00:18:28.828 "write": true, 00:18:28.828 "unmap": true, 00:18:28.828 "flush": true, 00:18:28.828 "reset": true, 00:18:28.828 "nvme_admin": false, 00:18:28.828 "nvme_io": false, 00:18:28.828 "nvme_io_md": false, 00:18:28.828 "write_zeroes": true, 00:18:28.828 "zcopy": true, 00:18:28.828 "get_zone_info": false, 00:18:28.828 "zone_management": false, 00:18:28.828 "zone_append": false, 00:18:28.828 "compare": false, 00:18:28.828 "compare_and_write": false, 00:18:28.828 "abort": true, 00:18:28.828 "seek_hole": false, 00:18:28.828 "seek_data": false, 00:18:28.828 "copy": true, 00:18:28.828 "nvme_iov_md": false 00:18:28.828 }, 00:18:28.828 "memory_domains": [ 00:18:28.828 { 00:18:28.828 "dma_device_id": "system", 00:18:28.828 "dma_device_type": 1 00:18:28.828 }, 00:18:28.828 { 00:18:28.828 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:28.828 "dma_device_type": 2 00:18:28.828 } 00:18:28.828 ], 00:18:28.828 "driver_specific": {} 00:18:28.828 } 00:18:28.828 ] 00:18:28.828 11:29:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:18:28.828 11:29:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:18:28.828 11:29:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:18:28.828 11:29:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:18:28.828 11:29:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:28.828 11:29:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:18:28.828 11:29:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:28.828 11:29:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:28.828 11:29:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:28.828 11:29:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:28.828 11:29:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:28.828 11:29:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:28.828 11:29:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:28.828 11:29:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:28.828 11:29:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:29.087 11:29:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:29.087 "name": "Existed_Raid", 00:18:29.087 "uuid": "88cfb57e-f525-4bc6-abcd-1d53e1b1ed94", 00:18:29.087 "strip_size_kb": 64, 00:18:29.087 "state": "online", 00:18:29.087 "raid_level": "raid0", 00:18:29.087 "superblock": true, 00:18:29.087 "num_base_bdevs": 4, 00:18:29.087 "num_base_bdevs_discovered": 4, 00:18:29.087 "num_base_bdevs_operational": 4, 00:18:29.087 "base_bdevs_list": [ 00:18:29.087 { 00:18:29.087 "name": "BaseBdev1", 00:18:29.087 "uuid": "32a7dcbe-d5ad-41c5-be5b-82a3c01d26a4", 00:18:29.087 "is_configured": true, 00:18:29.087 "data_offset": 2048, 00:18:29.087 "data_size": 63488 00:18:29.087 }, 00:18:29.087 { 00:18:29.087 "name": "BaseBdev2", 00:18:29.087 "uuid": "3827b760-0b14-4f31-b7b8-6b47cd1ac326", 00:18:29.087 "is_configured": true, 00:18:29.087 "data_offset": 2048, 00:18:29.087 "data_size": 63488 00:18:29.087 }, 00:18:29.087 { 00:18:29.087 "name": "BaseBdev3", 00:18:29.087 "uuid": "02aa1a62-8717-41dd-b4c4-6cb58c3c41d6", 00:18:29.087 "is_configured": true, 00:18:29.087 "data_offset": 2048, 00:18:29.087 "data_size": 63488 00:18:29.087 }, 00:18:29.087 { 00:18:29.087 "name": "BaseBdev4", 00:18:29.087 "uuid": "15b70ca5-b0c3-4f5c-b8b7-dd0fb0e712c6", 00:18:29.087 "is_configured": true, 00:18:29.087 "data_offset": 2048, 00:18:29.087 "data_size": 63488 00:18:29.087 } 00:18:29.087 ] 00:18:29.087 }' 00:18:29.087 11:29:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:29.087 11:29:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:18:30.020 11:29:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:18:30.020 11:29:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:18:30.020 11:29:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:18:30.020 11:29:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:18:30.020 11:29:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:18:30.020 11:29:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # local name 00:18:30.020 11:29:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:18:30.020 11:29:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:18:30.278 [2024-07-15 11:29:13.680397] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:18:30.278 11:29:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:18:30.278 "name": "Existed_Raid", 00:18:30.278 "aliases": [ 00:18:30.278 "88cfb57e-f525-4bc6-abcd-1d53e1b1ed94" 00:18:30.278 ], 00:18:30.278 "product_name": "Raid Volume", 00:18:30.278 "block_size": 512, 00:18:30.278 "num_blocks": 253952, 00:18:30.278 "uuid": "88cfb57e-f525-4bc6-abcd-1d53e1b1ed94", 00:18:30.278 "assigned_rate_limits": { 00:18:30.278 "rw_ios_per_sec": 0, 00:18:30.278 "rw_mbytes_per_sec": 0, 00:18:30.278 "r_mbytes_per_sec": 0, 00:18:30.278 "w_mbytes_per_sec": 0 00:18:30.278 }, 00:18:30.278 "claimed": false, 00:18:30.278 "zoned": false, 00:18:30.278 "supported_io_types": { 00:18:30.278 "read": true, 00:18:30.278 "write": true, 00:18:30.278 "unmap": true, 00:18:30.278 "flush": true, 00:18:30.278 "reset": true, 00:18:30.278 "nvme_admin": false, 00:18:30.278 "nvme_io": false, 00:18:30.278 "nvme_io_md": false, 00:18:30.278 "write_zeroes": true, 00:18:30.278 "zcopy": false, 00:18:30.278 "get_zone_info": false, 00:18:30.278 "zone_management": false, 00:18:30.278 "zone_append": false, 00:18:30.278 "compare": false, 00:18:30.278 "compare_and_write": false, 00:18:30.278 "abort": false, 00:18:30.278 "seek_hole": false, 00:18:30.278 "seek_data": false, 00:18:30.278 "copy": false, 00:18:30.278 "nvme_iov_md": false 00:18:30.278 }, 00:18:30.278 "memory_domains": [ 00:18:30.278 { 00:18:30.278 "dma_device_id": "system", 00:18:30.278 "dma_device_type": 1 00:18:30.278 }, 00:18:30.278 { 00:18:30.278 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:30.278 "dma_device_type": 2 00:18:30.278 }, 00:18:30.278 { 00:18:30.278 "dma_device_id": "system", 00:18:30.278 "dma_device_type": 1 00:18:30.278 }, 00:18:30.278 { 00:18:30.278 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:30.278 "dma_device_type": 2 00:18:30.278 }, 00:18:30.278 { 00:18:30.278 "dma_device_id": "system", 00:18:30.278 "dma_device_type": 1 00:18:30.278 }, 00:18:30.278 { 00:18:30.278 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:30.278 "dma_device_type": 2 00:18:30.278 }, 00:18:30.278 { 00:18:30.278 "dma_device_id": "system", 00:18:30.278 "dma_device_type": 1 00:18:30.278 }, 00:18:30.278 { 00:18:30.278 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:30.278 "dma_device_type": 2 00:18:30.278 } 00:18:30.278 ], 00:18:30.278 "driver_specific": { 00:18:30.278 "raid": { 00:18:30.278 "uuid": "88cfb57e-f525-4bc6-abcd-1d53e1b1ed94", 00:18:30.278 "strip_size_kb": 64, 00:18:30.278 "state": "online", 00:18:30.278 "raid_level": "raid0", 00:18:30.278 "superblock": true, 00:18:30.278 "num_base_bdevs": 4, 00:18:30.278 "num_base_bdevs_discovered": 4, 00:18:30.278 "num_base_bdevs_operational": 4, 00:18:30.278 "base_bdevs_list": [ 00:18:30.278 { 00:18:30.278 "name": "BaseBdev1", 00:18:30.278 "uuid": "32a7dcbe-d5ad-41c5-be5b-82a3c01d26a4", 00:18:30.278 "is_configured": true, 00:18:30.278 "data_offset": 2048, 00:18:30.278 "data_size": 63488 00:18:30.278 }, 00:18:30.278 { 00:18:30.278 "name": "BaseBdev2", 00:18:30.278 "uuid": "3827b760-0b14-4f31-b7b8-6b47cd1ac326", 00:18:30.278 "is_configured": true, 00:18:30.278 "data_offset": 2048, 00:18:30.278 "data_size": 63488 00:18:30.278 }, 00:18:30.278 { 00:18:30.278 "name": "BaseBdev3", 00:18:30.278 "uuid": "02aa1a62-8717-41dd-b4c4-6cb58c3c41d6", 00:18:30.278 "is_configured": true, 00:18:30.278 "data_offset": 2048, 00:18:30.278 "data_size": 63488 00:18:30.278 }, 00:18:30.278 { 00:18:30.278 "name": "BaseBdev4", 00:18:30.278 "uuid": "15b70ca5-b0c3-4f5c-b8b7-dd0fb0e712c6", 00:18:30.278 "is_configured": true, 00:18:30.278 "data_offset": 2048, 00:18:30.278 "data_size": 63488 00:18:30.278 } 00:18:30.278 ] 00:18:30.278 } 00:18:30.278 } 00:18:30.278 }' 00:18:30.278 11:29:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:18:30.278 11:29:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:18:30.278 BaseBdev2 00:18:30.278 BaseBdev3 00:18:30.278 BaseBdev4' 00:18:30.278 11:29:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:18:30.278 11:29:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:18:30.278 11:29:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:18:30.536 11:29:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:18:30.536 "name": "BaseBdev1", 00:18:30.536 "aliases": [ 00:18:30.536 "32a7dcbe-d5ad-41c5-be5b-82a3c01d26a4" 00:18:30.536 ], 00:18:30.536 "product_name": "Malloc disk", 00:18:30.536 "block_size": 512, 00:18:30.536 "num_blocks": 65536, 00:18:30.536 "uuid": "32a7dcbe-d5ad-41c5-be5b-82a3c01d26a4", 00:18:30.536 "assigned_rate_limits": { 00:18:30.536 "rw_ios_per_sec": 0, 00:18:30.536 "rw_mbytes_per_sec": 0, 00:18:30.536 "r_mbytes_per_sec": 0, 00:18:30.536 "w_mbytes_per_sec": 0 00:18:30.536 }, 00:18:30.536 "claimed": true, 00:18:30.536 "claim_type": "exclusive_write", 00:18:30.536 "zoned": false, 00:18:30.536 "supported_io_types": { 00:18:30.536 "read": true, 00:18:30.536 "write": true, 00:18:30.536 "unmap": true, 00:18:30.536 "flush": true, 00:18:30.536 "reset": true, 00:18:30.536 "nvme_admin": false, 00:18:30.536 "nvme_io": false, 00:18:30.536 "nvme_io_md": false, 00:18:30.536 "write_zeroes": true, 00:18:30.536 "zcopy": true, 00:18:30.536 "get_zone_info": false, 00:18:30.536 "zone_management": false, 00:18:30.536 "zone_append": false, 00:18:30.536 "compare": false, 00:18:30.536 "compare_and_write": false, 00:18:30.536 "abort": true, 00:18:30.536 "seek_hole": false, 00:18:30.536 "seek_data": false, 00:18:30.536 "copy": true, 00:18:30.536 "nvme_iov_md": false 00:18:30.536 }, 00:18:30.536 "memory_domains": [ 00:18:30.536 { 00:18:30.536 "dma_device_id": "system", 00:18:30.536 "dma_device_type": 1 00:18:30.536 }, 00:18:30.536 { 00:18:30.536 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:30.536 "dma_device_type": 2 00:18:30.536 } 00:18:30.536 ], 00:18:30.536 "driver_specific": {} 00:18:30.536 }' 00:18:30.537 11:29:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:30.537 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:30.537 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:18:30.537 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:30.794 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:30.794 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:18:30.794 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:30.794 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:30.794 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:18:30.794 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:30.794 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:30.794 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:18:30.794 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:18:30.794 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:18:30.794 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:18:31.052 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:18:31.052 "name": "BaseBdev2", 00:18:31.052 "aliases": [ 00:18:31.052 "3827b760-0b14-4f31-b7b8-6b47cd1ac326" 00:18:31.052 ], 00:18:31.052 "product_name": "Malloc disk", 00:18:31.052 "block_size": 512, 00:18:31.052 "num_blocks": 65536, 00:18:31.052 "uuid": "3827b760-0b14-4f31-b7b8-6b47cd1ac326", 00:18:31.052 "assigned_rate_limits": { 00:18:31.052 "rw_ios_per_sec": 0, 00:18:31.052 "rw_mbytes_per_sec": 0, 00:18:31.052 "r_mbytes_per_sec": 0, 00:18:31.052 "w_mbytes_per_sec": 0 00:18:31.052 }, 00:18:31.052 "claimed": true, 00:18:31.052 "claim_type": "exclusive_write", 00:18:31.053 "zoned": false, 00:18:31.053 "supported_io_types": { 00:18:31.053 "read": true, 00:18:31.053 "write": true, 00:18:31.053 "unmap": true, 00:18:31.053 "flush": true, 00:18:31.053 "reset": true, 00:18:31.053 "nvme_admin": false, 00:18:31.053 "nvme_io": false, 00:18:31.053 "nvme_io_md": false, 00:18:31.053 "write_zeroes": true, 00:18:31.053 "zcopy": true, 00:18:31.053 "get_zone_info": false, 00:18:31.053 "zone_management": false, 00:18:31.053 "zone_append": false, 00:18:31.053 "compare": false, 00:18:31.053 "compare_and_write": false, 00:18:31.053 "abort": true, 00:18:31.053 "seek_hole": false, 00:18:31.053 "seek_data": false, 00:18:31.053 "copy": true, 00:18:31.053 "nvme_iov_md": false 00:18:31.053 }, 00:18:31.053 "memory_domains": [ 00:18:31.053 { 00:18:31.053 "dma_device_id": "system", 00:18:31.053 "dma_device_type": 1 00:18:31.053 }, 00:18:31.053 { 00:18:31.053 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:31.053 "dma_device_type": 2 00:18:31.053 } 00:18:31.053 ], 00:18:31.053 "driver_specific": {} 00:18:31.053 }' 00:18:31.053 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:31.053 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:31.053 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:18:31.053 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:31.310 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:31.310 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:18:31.310 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:31.310 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:31.310 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:18:31.310 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:31.310 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:31.310 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:18:31.310 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:18:31.310 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:18:31.310 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:18:31.568 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:18:31.568 "name": "BaseBdev3", 00:18:31.568 "aliases": [ 00:18:31.568 "02aa1a62-8717-41dd-b4c4-6cb58c3c41d6" 00:18:31.568 ], 00:18:31.568 "product_name": "Malloc disk", 00:18:31.568 "block_size": 512, 00:18:31.568 "num_blocks": 65536, 00:18:31.568 "uuid": "02aa1a62-8717-41dd-b4c4-6cb58c3c41d6", 00:18:31.568 "assigned_rate_limits": { 00:18:31.568 "rw_ios_per_sec": 0, 00:18:31.568 "rw_mbytes_per_sec": 0, 00:18:31.568 "r_mbytes_per_sec": 0, 00:18:31.568 "w_mbytes_per_sec": 0 00:18:31.568 }, 00:18:31.568 "claimed": true, 00:18:31.568 "claim_type": "exclusive_write", 00:18:31.568 "zoned": false, 00:18:31.568 "supported_io_types": { 00:18:31.568 "read": true, 00:18:31.568 "write": true, 00:18:31.568 "unmap": true, 00:18:31.568 "flush": true, 00:18:31.568 "reset": true, 00:18:31.568 "nvme_admin": false, 00:18:31.568 "nvme_io": false, 00:18:31.568 "nvme_io_md": false, 00:18:31.568 "write_zeroes": true, 00:18:31.568 "zcopy": true, 00:18:31.568 "get_zone_info": false, 00:18:31.568 "zone_management": false, 00:18:31.568 "zone_append": false, 00:18:31.568 "compare": false, 00:18:31.568 "compare_and_write": false, 00:18:31.568 "abort": true, 00:18:31.568 "seek_hole": false, 00:18:31.568 "seek_data": false, 00:18:31.568 "copy": true, 00:18:31.568 "nvme_iov_md": false 00:18:31.568 }, 00:18:31.568 "memory_domains": [ 00:18:31.568 { 00:18:31.568 "dma_device_id": "system", 00:18:31.568 "dma_device_type": 1 00:18:31.568 }, 00:18:31.568 { 00:18:31.568 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:31.568 "dma_device_type": 2 00:18:31.568 } 00:18:31.568 ], 00:18:31.568 "driver_specific": {} 00:18:31.568 }' 00:18:31.568 11:29:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:31.568 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:31.568 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:18:31.568 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:31.568 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:31.568 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:18:31.568 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:31.826 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:31.826 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:18:31.826 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:31.826 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:31.826 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:18:31.826 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:18:31.826 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 00:18:31.826 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:18:32.392 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:18:32.392 "name": "BaseBdev4", 00:18:32.392 "aliases": [ 00:18:32.392 "15b70ca5-b0c3-4f5c-b8b7-dd0fb0e712c6" 00:18:32.392 ], 00:18:32.392 "product_name": "Malloc disk", 00:18:32.392 "block_size": 512, 00:18:32.392 "num_blocks": 65536, 00:18:32.392 "uuid": "15b70ca5-b0c3-4f5c-b8b7-dd0fb0e712c6", 00:18:32.392 "assigned_rate_limits": { 00:18:32.392 "rw_ios_per_sec": 0, 00:18:32.392 "rw_mbytes_per_sec": 0, 00:18:32.392 "r_mbytes_per_sec": 0, 00:18:32.392 "w_mbytes_per_sec": 0 00:18:32.392 }, 00:18:32.392 "claimed": true, 00:18:32.392 "claim_type": "exclusive_write", 00:18:32.392 "zoned": false, 00:18:32.392 "supported_io_types": { 00:18:32.392 "read": true, 00:18:32.392 "write": true, 00:18:32.392 "unmap": true, 00:18:32.392 "flush": true, 00:18:32.392 "reset": true, 00:18:32.392 "nvme_admin": false, 00:18:32.392 "nvme_io": false, 00:18:32.392 "nvme_io_md": false, 00:18:32.392 "write_zeroes": true, 00:18:32.392 "zcopy": true, 00:18:32.392 "get_zone_info": false, 00:18:32.392 "zone_management": false, 00:18:32.392 "zone_append": false, 00:18:32.392 "compare": false, 00:18:32.392 "compare_and_write": false, 00:18:32.392 "abort": true, 00:18:32.392 "seek_hole": false, 00:18:32.392 "seek_data": false, 00:18:32.392 "copy": true, 00:18:32.392 "nvme_iov_md": false 00:18:32.392 }, 00:18:32.392 "memory_domains": [ 00:18:32.392 { 00:18:32.392 "dma_device_id": "system", 00:18:32.392 "dma_device_type": 1 00:18:32.392 }, 00:18:32.392 { 00:18:32.392 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:32.392 "dma_device_type": 2 00:18:32.392 } 00:18:32.392 ], 00:18:32.392 "driver_specific": {} 00:18:32.392 }' 00:18:32.392 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:32.392 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:32.392 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:18:32.392 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:32.392 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:32.392 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:18:32.392 11:29:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:32.649 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:32.649 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:18:32.649 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:32.649 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:32.649 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:18:32.649 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:18:32.908 [2024-07-15 11:29:16.395306] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:18:32.908 [2024-07-15 11:29:16.395331] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:18:32.908 [2024-07-15 11:29:16.395378] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:18:32.908 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@275 -- # local expected_state 00:18:32.908 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # has_redundancy raid0 00:18:32.908 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # case $1 in 00:18:32.908 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # return 1 00:18:32.908 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@277 -- # expected_state=offline 00:18:32.908 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:18:32.908 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:32.908 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=offline 00:18:32.908 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:32.908 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:32.908 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:18:32.908 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:32.908 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:32.908 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:32.908 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:32.908 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:32.908 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:33.168 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:33.168 "name": "Existed_Raid", 00:18:33.168 "uuid": "88cfb57e-f525-4bc6-abcd-1d53e1b1ed94", 00:18:33.168 "strip_size_kb": 64, 00:18:33.168 "state": "offline", 00:18:33.168 "raid_level": "raid0", 00:18:33.168 "superblock": true, 00:18:33.168 "num_base_bdevs": 4, 00:18:33.168 "num_base_bdevs_discovered": 3, 00:18:33.168 "num_base_bdevs_operational": 3, 00:18:33.168 "base_bdevs_list": [ 00:18:33.168 { 00:18:33.168 "name": null, 00:18:33.168 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:33.168 "is_configured": false, 00:18:33.168 "data_offset": 2048, 00:18:33.168 "data_size": 63488 00:18:33.168 }, 00:18:33.168 { 00:18:33.168 "name": "BaseBdev2", 00:18:33.168 "uuid": "3827b760-0b14-4f31-b7b8-6b47cd1ac326", 00:18:33.168 "is_configured": true, 00:18:33.168 "data_offset": 2048, 00:18:33.168 "data_size": 63488 00:18:33.168 }, 00:18:33.168 { 00:18:33.168 "name": "BaseBdev3", 00:18:33.168 "uuid": "02aa1a62-8717-41dd-b4c4-6cb58c3c41d6", 00:18:33.168 "is_configured": true, 00:18:33.168 "data_offset": 2048, 00:18:33.168 "data_size": 63488 00:18:33.168 }, 00:18:33.168 { 00:18:33.168 "name": "BaseBdev4", 00:18:33.168 "uuid": "15b70ca5-b0c3-4f5c-b8b7-dd0fb0e712c6", 00:18:33.168 "is_configured": true, 00:18:33.168 "data_offset": 2048, 00:18:33.168 "data_size": 63488 00:18:33.168 } 00:18:33.168 ] 00:18:33.168 }' 00:18:33.168 11:29:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:33.168 11:29:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:18:33.735 11:29:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:18:33.735 11:29:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:18:33.735 11:29:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:33.735 11:29:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:18:33.993 11:29:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:18:33.993 11:29:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:18:33.993 11:29:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:18:34.252 [2024-07-15 11:29:17.723907] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:18:34.252 11:29:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:18:34.252 11:29:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:18:34.252 11:29:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:34.252 11:29:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:18:34.512 11:29:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:18:34.512 11:29:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:18:34.512 11:29:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev3 00:18:34.771 [2024-07-15 11:29:18.212127] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:18:34.771 11:29:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:18:34.771 11:29:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:18:34.771 11:29:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:34.771 11:29:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:18:35.029 11:29:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:18:35.029 11:29:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:18:35.029 11:29:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev4 00:18:35.287 [2024-07-15 11:29:18.633673] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:18:35.287 [2024-07-15 11:29:18.633719] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xe83350 name Existed_Raid, state offline 00:18:35.288 11:29:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:18:35.288 11:29:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:18:35.288 11:29:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:35.288 11:29:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:18:35.288 11:29:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:18:35.288 11:29:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:18:35.288 11:29:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # '[' 4 -gt 2 ']' 00:18:35.288 11:29:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i = 1 )) 00:18:35.288 11:29:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:18:35.288 11:29:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:18:35.546 BaseBdev2 00:18:35.546 11:29:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev2 00:18:35.546 11:29:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:18:35.546 11:29:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:18:35.546 11:29:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:18:35.546 11:29:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:18:35.546 11:29:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:18:35.546 11:29:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:18:35.804 11:29:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:18:36.063 [ 00:18:36.063 { 00:18:36.063 "name": "BaseBdev2", 00:18:36.063 "aliases": [ 00:18:36.063 "049f4bec-c5d9-4501-90b7-1e4997ff96cb" 00:18:36.063 ], 00:18:36.063 "product_name": "Malloc disk", 00:18:36.063 "block_size": 512, 00:18:36.063 "num_blocks": 65536, 00:18:36.063 "uuid": "049f4bec-c5d9-4501-90b7-1e4997ff96cb", 00:18:36.063 "assigned_rate_limits": { 00:18:36.063 "rw_ios_per_sec": 0, 00:18:36.063 "rw_mbytes_per_sec": 0, 00:18:36.063 "r_mbytes_per_sec": 0, 00:18:36.063 "w_mbytes_per_sec": 0 00:18:36.063 }, 00:18:36.063 "claimed": false, 00:18:36.063 "zoned": false, 00:18:36.063 "supported_io_types": { 00:18:36.063 "read": true, 00:18:36.063 "write": true, 00:18:36.063 "unmap": true, 00:18:36.063 "flush": true, 00:18:36.063 "reset": true, 00:18:36.063 "nvme_admin": false, 00:18:36.063 "nvme_io": false, 00:18:36.063 "nvme_io_md": false, 00:18:36.063 "write_zeroes": true, 00:18:36.063 "zcopy": true, 00:18:36.063 "get_zone_info": false, 00:18:36.063 "zone_management": false, 00:18:36.063 "zone_append": false, 00:18:36.063 "compare": false, 00:18:36.063 "compare_and_write": false, 00:18:36.063 "abort": true, 00:18:36.063 "seek_hole": false, 00:18:36.063 "seek_data": false, 00:18:36.063 "copy": true, 00:18:36.063 "nvme_iov_md": false 00:18:36.063 }, 00:18:36.063 "memory_domains": [ 00:18:36.063 { 00:18:36.063 "dma_device_id": "system", 00:18:36.063 "dma_device_type": 1 00:18:36.063 }, 00:18:36.063 { 00:18:36.063 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:36.063 "dma_device_type": 2 00:18:36.063 } 00:18:36.064 ], 00:18:36.064 "driver_specific": {} 00:18:36.064 } 00:18:36.064 ] 00:18:36.064 11:29:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:18:36.064 11:29:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:18:36.064 11:29:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:18:36.064 11:29:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:18:36.064 BaseBdev3 00:18:36.064 11:29:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev3 00:18:36.064 11:29:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:18:36.064 11:29:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:18:36.064 11:29:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:18:36.064 11:29:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:18:36.064 11:29:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:18:36.064 11:29:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:18:36.349 11:29:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:18:36.608 [ 00:18:36.608 { 00:18:36.608 "name": "BaseBdev3", 00:18:36.608 "aliases": [ 00:18:36.608 "e1df1ece-c669-479a-84f9-9bacdf7c9131" 00:18:36.608 ], 00:18:36.608 "product_name": "Malloc disk", 00:18:36.608 "block_size": 512, 00:18:36.608 "num_blocks": 65536, 00:18:36.608 "uuid": "e1df1ece-c669-479a-84f9-9bacdf7c9131", 00:18:36.608 "assigned_rate_limits": { 00:18:36.608 "rw_ios_per_sec": 0, 00:18:36.608 "rw_mbytes_per_sec": 0, 00:18:36.608 "r_mbytes_per_sec": 0, 00:18:36.608 "w_mbytes_per_sec": 0 00:18:36.608 }, 00:18:36.608 "claimed": false, 00:18:36.608 "zoned": false, 00:18:36.608 "supported_io_types": { 00:18:36.608 "read": true, 00:18:36.608 "write": true, 00:18:36.608 "unmap": true, 00:18:36.608 "flush": true, 00:18:36.608 "reset": true, 00:18:36.608 "nvme_admin": false, 00:18:36.608 "nvme_io": false, 00:18:36.608 "nvme_io_md": false, 00:18:36.608 "write_zeroes": true, 00:18:36.608 "zcopy": true, 00:18:36.608 "get_zone_info": false, 00:18:36.608 "zone_management": false, 00:18:36.608 "zone_append": false, 00:18:36.608 "compare": false, 00:18:36.608 "compare_and_write": false, 00:18:36.608 "abort": true, 00:18:36.608 "seek_hole": false, 00:18:36.608 "seek_data": false, 00:18:36.608 "copy": true, 00:18:36.608 "nvme_iov_md": false 00:18:36.608 }, 00:18:36.608 "memory_domains": [ 00:18:36.608 { 00:18:36.608 "dma_device_id": "system", 00:18:36.608 "dma_device_type": 1 00:18:36.608 }, 00:18:36.608 { 00:18:36.608 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:36.608 "dma_device_type": 2 00:18:36.608 } 00:18:36.608 ], 00:18:36.608 "driver_specific": {} 00:18:36.608 } 00:18:36.608 ] 00:18:36.608 11:29:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:18:36.608 11:29:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:18:36.608 11:29:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:18:36.608 11:29:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4 00:18:36.867 BaseBdev4 00:18:36.867 11:29:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev4 00:18:36.867 11:29:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev4 00:18:36.867 11:29:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:18:36.867 11:29:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:18:36.867 11:29:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:18:36.867 11:29:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:18:36.867 11:29:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:18:37.125 11:29:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 -t 2000 00:18:37.125 [ 00:18:37.125 { 00:18:37.125 "name": "BaseBdev4", 00:18:37.125 "aliases": [ 00:18:37.125 "29298378-a06e-47b3-bff6-af8f8c6394c5" 00:18:37.125 ], 00:18:37.125 "product_name": "Malloc disk", 00:18:37.125 "block_size": 512, 00:18:37.125 "num_blocks": 65536, 00:18:37.125 "uuid": "29298378-a06e-47b3-bff6-af8f8c6394c5", 00:18:37.125 "assigned_rate_limits": { 00:18:37.125 "rw_ios_per_sec": 0, 00:18:37.125 "rw_mbytes_per_sec": 0, 00:18:37.125 "r_mbytes_per_sec": 0, 00:18:37.125 "w_mbytes_per_sec": 0 00:18:37.125 }, 00:18:37.125 "claimed": false, 00:18:37.125 "zoned": false, 00:18:37.125 "supported_io_types": { 00:18:37.125 "read": true, 00:18:37.125 "write": true, 00:18:37.125 "unmap": true, 00:18:37.125 "flush": true, 00:18:37.125 "reset": true, 00:18:37.125 "nvme_admin": false, 00:18:37.125 "nvme_io": false, 00:18:37.125 "nvme_io_md": false, 00:18:37.125 "write_zeroes": true, 00:18:37.125 "zcopy": true, 00:18:37.125 "get_zone_info": false, 00:18:37.125 "zone_management": false, 00:18:37.125 "zone_append": false, 00:18:37.125 "compare": false, 00:18:37.125 "compare_and_write": false, 00:18:37.125 "abort": true, 00:18:37.125 "seek_hole": false, 00:18:37.125 "seek_data": false, 00:18:37.125 "copy": true, 00:18:37.125 "nvme_iov_md": false 00:18:37.125 }, 00:18:37.125 "memory_domains": [ 00:18:37.125 { 00:18:37.125 "dma_device_id": "system", 00:18:37.125 "dma_device_type": 1 00:18:37.125 }, 00:18:37.125 { 00:18:37.125 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:37.125 "dma_device_type": 2 00:18:37.125 } 00:18:37.125 ], 00:18:37.125 "driver_specific": {} 00:18:37.125 } 00:18:37.125 ] 00:18:37.125 11:29:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:18:37.125 11:29:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:18:37.125 11:29:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:18:37.125 11:29:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@305 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r raid0 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:18:37.384 [2024-07-15 11:29:20.798609] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:18:37.384 [2024-07-15 11:29:20.798658] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:18:37.384 [2024-07-15 11:29:20.798676] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:18:37.384 [2024-07-15 11:29:20.800053] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:18:37.384 [2024-07-15 11:29:20.800096] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:18:37.384 11:29:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:18:37.384 11:29:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:37.384 11:29:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:18:37.384 11:29:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:37.384 11:29:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:37.384 11:29:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:37.384 11:29:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:37.384 11:29:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:37.384 11:29:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:37.384 11:29:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:37.384 11:29:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:37.384 11:29:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:37.950 11:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:37.950 "name": "Existed_Raid", 00:18:37.950 "uuid": "a9a38945-e8c3-479a-8cd8-ddfe527a18a0", 00:18:37.950 "strip_size_kb": 64, 00:18:37.950 "state": "configuring", 00:18:37.950 "raid_level": "raid0", 00:18:37.950 "superblock": true, 00:18:37.950 "num_base_bdevs": 4, 00:18:37.950 "num_base_bdevs_discovered": 3, 00:18:37.950 "num_base_bdevs_operational": 4, 00:18:37.950 "base_bdevs_list": [ 00:18:37.950 { 00:18:37.950 "name": "BaseBdev1", 00:18:37.950 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:37.950 "is_configured": false, 00:18:37.950 "data_offset": 0, 00:18:37.950 "data_size": 0 00:18:37.950 }, 00:18:37.950 { 00:18:37.950 "name": "BaseBdev2", 00:18:37.950 "uuid": "049f4bec-c5d9-4501-90b7-1e4997ff96cb", 00:18:37.950 "is_configured": true, 00:18:37.950 "data_offset": 2048, 00:18:37.950 "data_size": 63488 00:18:37.950 }, 00:18:37.950 { 00:18:37.950 "name": "BaseBdev3", 00:18:37.950 "uuid": "e1df1ece-c669-479a-84f9-9bacdf7c9131", 00:18:37.950 "is_configured": true, 00:18:37.950 "data_offset": 2048, 00:18:37.950 "data_size": 63488 00:18:37.951 }, 00:18:37.951 { 00:18:37.951 "name": "BaseBdev4", 00:18:37.951 "uuid": "29298378-a06e-47b3-bff6-af8f8c6394c5", 00:18:37.951 "is_configured": true, 00:18:37.951 "data_offset": 2048, 00:18:37.951 "data_size": 63488 00:18:37.951 } 00:18:37.951 ] 00:18:37.951 }' 00:18:37.951 11:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:37.951 11:29:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:18:38.520 11:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev2 00:18:38.520 [2024-07-15 11:29:22.037900] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:18:38.520 11:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@309 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:18:38.520 11:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:38.520 11:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:18:38.520 11:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:38.520 11:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:38.520 11:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:38.520 11:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:38.520 11:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:38.520 11:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:38.520 11:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:38.520 11:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:38.520 11:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:38.779 11:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:38.780 "name": "Existed_Raid", 00:18:38.780 "uuid": "a9a38945-e8c3-479a-8cd8-ddfe527a18a0", 00:18:38.780 "strip_size_kb": 64, 00:18:38.780 "state": "configuring", 00:18:38.780 "raid_level": "raid0", 00:18:38.780 "superblock": true, 00:18:38.780 "num_base_bdevs": 4, 00:18:38.780 "num_base_bdevs_discovered": 2, 00:18:38.780 "num_base_bdevs_operational": 4, 00:18:38.780 "base_bdevs_list": [ 00:18:38.780 { 00:18:38.780 "name": "BaseBdev1", 00:18:38.780 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:38.780 "is_configured": false, 00:18:38.780 "data_offset": 0, 00:18:38.780 "data_size": 0 00:18:38.780 }, 00:18:38.780 { 00:18:38.780 "name": null, 00:18:38.780 "uuid": "049f4bec-c5d9-4501-90b7-1e4997ff96cb", 00:18:38.780 "is_configured": false, 00:18:38.780 "data_offset": 2048, 00:18:38.780 "data_size": 63488 00:18:38.780 }, 00:18:38.780 { 00:18:38.780 "name": "BaseBdev3", 00:18:38.780 "uuid": "e1df1ece-c669-479a-84f9-9bacdf7c9131", 00:18:38.780 "is_configured": true, 00:18:38.780 "data_offset": 2048, 00:18:38.780 "data_size": 63488 00:18:38.780 }, 00:18:38.780 { 00:18:38.780 "name": "BaseBdev4", 00:18:38.780 "uuid": "29298378-a06e-47b3-bff6-af8f8c6394c5", 00:18:38.780 "is_configured": true, 00:18:38.780 "data_offset": 2048, 00:18:38.780 "data_size": 63488 00:18:38.780 } 00:18:38.780 ] 00:18:38.780 }' 00:18:38.780 11:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:38.780 11:29:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:18:39.346 11:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:39.346 11:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:18:39.605 11:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # [[ false == \f\a\l\s\e ]] 00:18:39.605 11:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:18:39.864 [2024-07-15 11:29:23.411791] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:18:39.864 BaseBdev1 00:18:39.864 11:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@313 -- # waitforbdev BaseBdev1 00:18:39.864 11:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:18:39.864 11:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:18:39.864 11:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:18:39.864 11:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:18:39.864 11:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:18:39.864 11:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:18:40.123 11:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:18:40.382 [ 00:18:40.382 { 00:18:40.382 "name": "BaseBdev1", 00:18:40.382 "aliases": [ 00:18:40.382 "73f20069-6b69-41d5-985a-5509c2f6ca24" 00:18:40.382 ], 00:18:40.382 "product_name": "Malloc disk", 00:18:40.382 "block_size": 512, 00:18:40.382 "num_blocks": 65536, 00:18:40.382 "uuid": "73f20069-6b69-41d5-985a-5509c2f6ca24", 00:18:40.382 "assigned_rate_limits": { 00:18:40.382 "rw_ios_per_sec": 0, 00:18:40.382 "rw_mbytes_per_sec": 0, 00:18:40.382 "r_mbytes_per_sec": 0, 00:18:40.382 "w_mbytes_per_sec": 0 00:18:40.382 }, 00:18:40.382 "claimed": true, 00:18:40.382 "claim_type": "exclusive_write", 00:18:40.382 "zoned": false, 00:18:40.382 "supported_io_types": { 00:18:40.382 "read": true, 00:18:40.382 "write": true, 00:18:40.382 "unmap": true, 00:18:40.382 "flush": true, 00:18:40.382 "reset": true, 00:18:40.382 "nvme_admin": false, 00:18:40.382 "nvme_io": false, 00:18:40.382 "nvme_io_md": false, 00:18:40.382 "write_zeroes": true, 00:18:40.382 "zcopy": true, 00:18:40.382 "get_zone_info": false, 00:18:40.382 "zone_management": false, 00:18:40.382 "zone_append": false, 00:18:40.382 "compare": false, 00:18:40.382 "compare_and_write": false, 00:18:40.382 "abort": true, 00:18:40.382 "seek_hole": false, 00:18:40.382 "seek_data": false, 00:18:40.382 "copy": true, 00:18:40.382 "nvme_iov_md": false 00:18:40.382 }, 00:18:40.382 "memory_domains": [ 00:18:40.382 { 00:18:40.382 "dma_device_id": "system", 00:18:40.382 "dma_device_type": 1 00:18:40.382 }, 00:18:40.382 { 00:18:40.382 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:40.382 "dma_device_type": 2 00:18:40.382 } 00:18:40.382 ], 00:18:40.382 "driver_specific": {} 00:18:40.382 } 00:18:40.382 ] 00:18:40.382 11:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:18:40.382 11:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:18:40.382 11:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:40.382 11:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:18:40.382 11:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:40.382 11:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:40.382 11:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:40.382 11:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:40.382 11:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:40.382 11:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:40.382 11:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:40.382 11:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:40.382 11:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:40.641 11:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:40.641 "name": "Existed_Raid", 00:18:40.641 "uuid": "a9a38945-e8c3-479a-8cd8-ddfe527a18a0", 00:18:40.641 "strip_size_kb": 64, 00:18:40.641 "state": "configuring", 00:18:40.641 "raid_level": "raid0", 00:18:40.641 "superblock": true, 00:18:40.641 "num_base_bdevs": 4, 00:18:40.641 "num_base_bdevs_discovered": 3, 00:18:40.641 "num_base_bdevs_operational": 4, 00:18:40.641 "base_bdevs_list": [ 00:18:40.641 { 00:18:40.641 "name": "BaseBdev1", 00:18:40.641 "uuid": "73f20069-6b69-41d5-985a-5509c2f6ca24", 00:18:40.641 "is_configured": true, 00:18:40.641 "data_offset": 2048, 00:18:40.641 "data_size": 63488 00:18:40.641 }, 00:18:40.642 { 00:18:40.642 "name": null, 00:18:40.642 "uuid": "049f4bec-c5d9-4501-90b7-1e4997ff96cb", 00:18:40.642 "is_configured": false, 00:18:40.642 "data_offset": 2048, 00:18:40.642 "data_size": 63488 00:18:40.642 }, 00:18:40.642 { 00:18:40.642 "name": "BaseBdev3", 00:18:40.642 "uuid": "e1df1ece-c669-479a-84f9-9bacdf7c9131", 00:18:40.642 "is_configured": true, 00:18:40.642 "data_offset": 2048, 00:18:40.642 "data_size": 63488 00:18:40.642 }, 00:18:40.642 { 00:18:40.642 "name": "BaseBdev4", 00:18:40.642 "uuid": "29298378-a06e-47b3-bff6-af8f8c6394c5", 00:18:40.642 "is_configured": true, 00:18:40.642 "data_offset": 2048, 00:18:40.642 "data_size": 63488 00:18:40.642 } 00:18:40.642 ] 00:18:40.642 }' 00:18:40.642 11:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:40.642 11:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:18:41.211 11:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:18:41.211 11:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:41.472 11:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # [[ true == \t\r\u\e ]] 00:18:41.472 11:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@317 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev3 00:18:41.792 [2024-07-15 11:29:25.212582] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:18:41.792 11:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:18:41.792 11:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:41.792 11:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:18:41.792 11:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:41.792 11:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:41.792 11:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:41.792 11:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:41.792 11:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:41.792 11:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:41.792 11:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:41.792 11:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:41.792 11:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:42.051 11:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:42.051 "name": "Existed_Raid", 00:18:42.051 "uuid": "a9a38945-e8c3-479a-8cd8-ddfe527a18a0", 00:18:42.051 "strip_size_kb": 64, 00:18:42.051 "state": "configuring", 00:18:42.051 "raid_level": "raid0", 00:18:42.051 "superblock": true, 00:18:42.051 "num_base_bdevs": 4, 00:18:42.051 "num_base_bdevs_discovered": 2, 00:18:42.051 "num_base_bdevs_operational": 4, 00:18:42.051 "base_bdevs_list": [ 00:18:42.051 { 00:18:42.051 "name": "BaseBdev1", 00:18:42.051 "uuid": "73f20069-6b69-41d5-985a-5509c2f6ca24", 00:18:42.051 "is_configured": true, 00:18:42.051 "data_offset": 2048, 00:18:42.051 "data_size": 63488 00:18:42.051 }, 00:18:42.051 { 00:18:42.051 "name": null, 00:18:42.051 "uuid": "049f4bec-c5d9-4501-90b7-1e4997ff96cb", 00:18:42.051 "is_configured": false, 00:18:42.051 "data_offset": 2048, 00:18:42.051 "data_size": 63488 00:18:42.051 }, 00:18:42.051 { 00:18:42.051 "name": null, 00:18:42.051 "uuid": "e1df1ece-c669-479a-84f9-9bacdf7c9131", 00:18:42.051 "is_configured": false, 00:18:42.051 "data_offset": 2048, 00:18:42.051 "data_size": 63488 00:18:42.051 }, 00:18:42.051 { 00:18:42.051 "name": "BaseBdev4", 00:18:42.051 "uuid": "29298378-a06e-47b3-bff6-af8f8c6394c5", 00:18:42.051 "is_configured": true, 00:18:42.051 "data_offset": 2048, 00:18:42.051 "data_size": 63488 00:18:42.051 } 00:18:42.051 ] 00:18:42.051 }' 00:18:42.051 11:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:42.051 11:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:18:42.618 11:29:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:42.618 11:29:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:18:42.876 11:29:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # [[ false == \f\a\l\s\e ]] 00:18:42.876 11:29:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:18:43.135 [2024-07-15 11:29:26.604281] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:18:43.135 11:29:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@322 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:18:43.135 11:29:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:43.135 11:29:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:18:43.135 11:29:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:43.135 11:29:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:43.135 11:29:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:43.135 11:29:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:43.135 11:29:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:43.135 11:29:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:43.135 11:29:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:43.135 11:29:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:43.135 11:29:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:43.394 11:29:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:43.395 "name": "Existed_Raid", 00:18:43.395 "uuid": "a9a38945-e8c3-479a-8cd8-ddfe527a18a0", 00:18:43.395 "strip_size_kb": 64, 00:18:43.395 "state": "configuring", 00:18:43.395 "raid_level": "raid0", 00:18:43.395 "superblock": true, 00:18:43.395 "num_base_bdevs": 4, 00:18:43.395 "num_base_bdevs_discovered": 3, 00:18:43.395 "num_base_bdevs_operational": 4, 00:18:43.395 "base_bdevs_list": [ 00:18:43.395 { 00:18:43.395 "name": "BaseBdev1", 00:18:43.395 "uuid": "73f20069-6b69-41d5-985a-5509c2f6ca24", 00:18:43.395 "is_configured": true, 00:18:43.395 "data_offset": 2048, 00:18:43.395 "data_size": 63488 00:18:43.395 }, 00:18:43.395 { 00:18:43.395 "name": null, 00:18:43.395 "uuid": "049f4bec-c5d9-4501-90b7-1e4997ff96cb", 00:18:43.395 "is_configured": false, 00:18:43.395 "data_offset": 2048, 00:18:43.395 "data_size": 63488 00:18:43.395 }, 00:18:43.395 { 00:18:43.395 "name": "BaseBdev3", 00:18:43.395 "uuid": "e1df1ece-c669-479a-84f9-9bacdf7c9131", 00:18:43.395 "is_configured": true, 00:18:43.395 "data_offset": 2048, 00:18:43.395 "data_size": 63488 00:18:43.395 }, 00:18:43.395 { 00:18:43.395 "name": "BaseBdev4", 00:18:43.395 "uuid": "29298378-a06e-47b3-bff6-af8f8c6394c5", 00:18:43.395 "is_configured": true, 00:18:43.395 "data_offset": 2048, 00:18:43.395 "data_size": 63488 00:18:43.395 } 00:18:43.395 ] 00:18:43.395 }' 00:18:43.395 11:29:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:43.395 11:29:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:18:43.964 11:29:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:43.964 11:29:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:18:44.224 11:29:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # [[ true == \t\r\u\e ]] 00:18:44.224 11:29:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@325 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:18:44.483 [2024-07-15 11:29:27.863641] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:18:44.483 11:29:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:18:44.483 11:29:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:44.483 11:29:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:18:44.483 11:29:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:44.483 11:29:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:44.483 11:29:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:44.483 11:29:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:44.483 11:29:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:44.483 11:29:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:44.483 11:29:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:44.483 11:29:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:44.483 11:29:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:44.742 11:29:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:44.742 "name": "Existed_Raid", 00:18:44.742 "uuid": "a9a38945-e8c3-479a-8cd8-ddfe527a18a0", 00:18:44.742 "strip_size_kb": 64, 00:18:44.742 "state": "configuring", 00:18:44.742 "raid_level": "raid0", 00:18:44.742 "superblock": true, 00:18:44.742 "num_base_bdevs": 4, 00:18:44.742 "num_base_bdevs_discovered": 2, 00:18:44.742 "num_base_bdevs_operational": 4, 00:18:44.742 "base_bdevs_list": [ 00:18:44.742 { 00:18:44.742 "name": null, 00:18:44.742 "uuid": "73f20069-6b69-41d5-985a-5509c2f6ca24", 00:18:44.742 "is_configured": false, 00:18:44.742 "data_offset": 2048, 00:18:44.742 "data_size": 63488 00:18:44.742 }, 00:18:44.742 { 00:18:44.742 "name": null, 00:18:44.742 "uuid": "049f4bec-c5d9-4501-90b7-1e4997ff96cb", 00:18:44.742 "is_configured": false, 00:18:44.742 "data_offset": 2048, 00:18:44.742 "data_size": 63488 00:18:44.742 }, 00:18:44.742 { 00:18:44.742 "name": "BaseBdev3", 00:18:44.742 "uuid": "e1df1ece-c669-479a-84f9-9bacdf7c9131", 00:18:44.742 "is_configured": true, 00:18:44.742 "data_offset": 2048, 00:18:44.742 "data_size": 63488 00:18:44.742 }, 00:18:44.742 { 00:18:44.742 "name": "BaseBdev4", 00:18:44.742 "uuid": "29298378-a06e-47b3-bff6-af8f8c6394c5", 00:18:44.742 "is_configured": true, 00:18:44.742 "data_offset": 2048, 00:18:44.742 "data_size": 63488 00:18:44.742 } 00:18:44.742 ] 00:18:44.742 }' 00:18:44.742 11:29:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:44.742 11:29:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:18:45.311 11:29:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@327 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:18:45.311 11:29:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@327 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:45.570 11:29:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@327 -- # [[ false == \f\a\l\s\e ]] 00:18:45.570 11:29:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@329 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:18:45.828 [2024-07-15 11:29:29.288649] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:18:45.828 11:29:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@330 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:18:45.828 11:29:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:45.828 11:29:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:18:45.829 11:29:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:45.829 11:29:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:45.829 11:29:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:45.829 11:29:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:45.829 11:29:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:45.829 11:29:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:45.829 11:29:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:45.829 11:29:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:45.829 11:29:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:46.397 11:29:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:46.397 "name": "Existed_Raid", 00:18:46.397 "uuid": "a9a38945-e8c3-479a-8cd8-ddfe527a18a0", 00:18:46.397 "strip_size_kb": 64, 00:18:46.397 "state": "configuring", 00:18:46.397 "raid_level": "raid0", 00:18:46.397 "superblock": true, 00:18:46.397 "num_base_bdevs": 4, 00:18:46.397 "num_base_bdevs_discovered": 3, 00:18:46.397 "num_base_bdevs_operational": 4, 00:18:46.397 "base_bdevs_list": [ 00:18:46.397 { 00:18:46.397 "name": null, 00:18:46.397 "uuid": "73f20069-6b69-41d5-985a-5509c2f6ca24", 00:18:46.397 "is_configured": false, 00:18:46.397 "data_offset": 2048, 00:18:46.397 "data_size": 63488 00:18:46.397 }, 00:18:46.397 { 00:18:46.397 "name": "BaseBdev2", 00:18:46.397 "uuid": "049f4bec-c5d9-4501-90b7-1e4997ff96cb", 00:18:46.397 "is_configured": true, 00:18:46.397 "data_offset": 2048, 00:18:46.397 "data_size": 63488 00:18:46.397 }, 00:18:46.397 { 00:18:46.397 "name": "BaseBdev3", 00:18:46.397 "uuid": "e1df1ece-c669-479a-84f9-9bacdf7c9131", 00:18:46.397 "is_configured": true, 00:18:46.397 "data_offset": 2048, 00:18:46.397 "data_size": 63488 00:18:46.397 }, 00:18:46.397 { 00:18:46.397 "name": "BaseBdev4", 00:18:46.397 "uuid": "29298378-a06e-47b3-bff6-af8f8c6394c5", 00:18:46.397 "is_configured": true, 00:18:46.397 "data_offset": 2048, 00:18:46.397 "data_size": 63488 00:18:46.397 } 00:18:46.397 ] 00:18:46.397 }' 00:18:46.397 11:29:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:46.397 11:29:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:18:46.965 11:29:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@331 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:46.965 11:29:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@331 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:18:47.225 11:29:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@331 -- # [[ true == \t\r\u\e ]] 00:18:47.225 11:29:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:47.225 11:29:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@333 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:18:47.484 11:29:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b NewBaseBdev -u 73f20069-6b69-41d5-985a-5509c2f6ca24 00:18:47.485 [2024-07-15 11:29:30.998710] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:18:47.485 [2024-07-15 11:29:30.998955] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xe89470 00:18:47.485 [2024-07-15 11:29:30.998975] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:18:47.485 [2024-07-15 11:29:30.999216] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xe79c40 00:18:47.485 [2024-07-15 11:29:30.999381] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xe89470 00:18:47.485 [2024-07-15 11:29:30.999395] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0xe89470 00:18:47.485 [2024-07-15 11:29:30.999527] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:18:47.485 NewBaseBdev 00:18:47.485 11:29:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@334 -- # waitforbdev NewBaseBdev 00:18:47.485 11:29:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=NewBaseBdev 00:18:47.485 11:29:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:18:47.485 11:29:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:18:47.485 11:29:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:18:47.485 11:29:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:18:47.485 11:29:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:18:47.744 11:29:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev -t 2000 00:18:48.003 [ 00:18:48.003 { 00:18:48.003 "name": "NewBaseBdev", 00:18:48.003 "aliases": [ 00:18:48.003 "73f20069-6b69-41d5-985a-5509c2f6ca24" 00:18:48.003 ], 00:18:48.003 "product_name": "Malloc disk", 00:18:48.003 "block_size": 512, 00:18:48.003 "num_blocks": 65536, 00:18:48.004 "uuid": "73f20069-6b69-41d5-985a-5509c2f6ca24", 00:18:48.004 "assigned_rate_limits": { 00:18:48.004 "rw_ios_per_sec": 0, 00:18:48.004 "rw_mbytes_per_sec": 0, 00:18:48.004 "r_mbytes_per_sec": 0, 00:18:48.004 "w_mbytes_per_sec": 0 00:18:48.004 }, 00:18:48.004 "claimed": true, 00:18:48.004 "claim_type": "exclusive_write", 00:18:48.004 "zoned": false, 00:18:48.004 "supported_io_types": { 00:18:48.004 "read": true, 00:18:48.004 "write": true, 00:18:48.004 "unmap": true, 00:18:48.004 "flush": true, 00:18:48.004 "reset": true, 00:18:48.004 "nvme_admin": false, 00:18:48.004 "nvme_io": false, 00:18:48.004 "nvme_io_md": false, 00:18:48.004 "write_zeroes": true, 00:18:48.004 "zcopy": true, 00:18:48.004 "get_zone_info": false, 00:18:48.004 "zone_management": false, 00:18:48.004 "zone_append": false, 00:18:48.004 "compare": false, 00:18:48.004 "compare_and_write": false, 00:18:48.004 "abort": true, 00:18:48.004 "seek_hole": false, 00:18:48.004 "seek_data": false, 00:18:48.004 "copy": true, 00:18:48.004 "nvme_iov_md": false 00:18:48.004 }, 00:18:48.004 "memory_domains": [ 00:18:48.004 { 00:18:48.004 "dma_device_id": "system", 00:18:48.004 "dma_device_type": 1 00:18:48.004 }, 00:18:48.004 { 00:18:48.004 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:48.004 "dma_device_type": 2 00:18:48.004 } 00:18:48.004 ], 00:18:48.004 "driver_specific": {} 00:18:48.004 } 00:18:48.004 ] 00:18:48.004 11:29:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:18:48.004 11:29:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@335 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:18:48.004 11:29:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:18:48.004 11:29:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:18:48.004 11:29:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:48.004 11:29:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:48.004 11:29:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:48.004 11:29:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:48.004 11:29:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:48.004 11:29:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:48.004 11:29:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:48.004 11:29:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:48.004 11:29:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:18:48.004 11:29:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:48.004 "name": "Existed_Raid", 00:18:48.004 "uuid": "a9a38945-e8c3-479a-8cd8-ddfe527a18a0", 00:18:48.004 "strip_size_kb": 64, 00:18:48.004 "state": "online", 00:18:48.004 "raid_level": "raid0", 00:18:48.004 "superblock": true, 00:18:48.004 "num_base_bdevs": 4, 00:18:48.004 "num_base_bdevs_discovered": 4, 00:18:48.004 "num_base_bdevs_operational": 4, 00:18:48.004 "base_bdevs_list": [ 00:18:48.004 { 00:18:48.004 "name": "NewBaseBdev", 00:18:48.004 "uuid": "73f20069-6b69-41d5-985a-5509c2f6ca24", 00:18:48.004 "is_configured": true, 00:18:48.004 "data_offset": 2048, 00:18:48.004 "data_size": 63488 00:18:48.004 }, 00:18:48.004 { 00:18:48.004 "name": "BaseBdev2", 00:18:48.004 "uuid": "049f4bec-c5d9-4501-90b7-1e4997ff96cb", 00:18:48.004 "is_configured": true, 00:18:48.004 "data_offset": 2048, 00:18:48.004 "data_size": 63488 00:18:48.004 }, 00:18:48.004 { 00:18:48.004 "name": "BaseBdev3", 00:18:48.004 "uuid": "e1df1ece-c669-479a-84f9-9bacdf7c9131", 00:18:48.004 "is_configured": true, 00:18:48.004 "data_offset": 2048, 00:18:48.004 "data_size": 63488 00:18:48.004 }, 00:18:48.004 { 00:18:48.004 "name": "BaseBdev4", 00:18:48.004 "uuid": "29298378-a06e-47b3-bff6-af8f8c6394c5", 00:18:48.004 "is_configured": true, 00:18:48.004 "data_offset": 2048, 00:18:48.004 "data_size": 63488 00:18:48.004 } 00:18:48.004 ] 00:18:48.004 }' 00:18:48.004 11:29:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:48.004 11:29:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:18:48.572 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@336 -- # verify_raid_bdev_properties Existed_Raid 00:18:48.572 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:18:48.572 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:18:48.572 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:18:48.572 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:18:48.572 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # local name 00:18:48.572 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:18:48.572 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:18:48.832 [2024-07-15 11:29:32.326587] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:18:48.832 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:18:48.832 "name": "Existed_Raid", 00:18:48.832 "aliases": [ 00:18:48.832 "a9a38945-e8c3-479a-8cd8-ddfe527a18a0" 00:18:48.832 ], 00:18:48.832 "product_name": "Raid Volume", 00:18:48.832 "block_size": 512, 00:18:48.832 "num_blocks": 253952, 00:18:48.832 "uuid": "a9a38945-e8c3-479a-8cd8-ddfe527a18a0", 00:18:48.832 "assigned_rate_limits": { 00:18:48.832 "rw_ios_per_sec": 0, 00:18:48.832 "rw_mbytes_per_sec": 0, 00:18:48.832 "r_mbytes_per_sec": 0, 00:18:48.832 "w_mbytes_per_sec": 0 00:18:48.832 }, 00:18:48.832 "claimed": false, 00:18:48.832 "zoned": false, 00:18:48.832 "supported_io_types": { 00:18:48.832 "read": true, 00:18:48.832 "write": true, 00:18:48.832 "unmap": true, 00:18:48.832 "flush": true, 00:18:48.832 "reset": true, 00:18:48.832 "nvme_admin": false, 00:18:48.832 "nvme_io": false, 00:18:48.832 "nvme_io_md": false, 00:18:48.832 "write_zeroes": true, 00:18:48.832 "zcopy": false, 00:18:48.832 "get_zone_info": false, 00:18:48.832 "zone_management": false, 00:18:48.832 "zone_append": false, 00:18:48.832 "compare": false, 00:18:48.832 "compare_and_write": false, 00:18:48.832 "abort": false, 00:18:48.832 "seek_hole": false, 00:18:48.832 "seek_data": false, 00:18:48.832 "copy": false, 00:18:48.832 "nvme_iov_md": false 00:18:48.832 }, 00:18:48.832 "memory_domains": [ 00:18:48.832 { 00:18:48.832 "dma_device_id": "system", 00:18:48.832 "dma_device_type": 1 00:18:48.832 }, 00:18:48.832 { 00:18:48.832 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:48.832 "dma_device_type": 2 00:18:48.832 }, 00:18:48.832 { 00:18:48.832 "dma_device_id": "system", 00:18:48.832 "dma_device_type": 1 00:18:48.832 }, 00:18:48.832 { 00:18:48.832 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:48.832 "dma_device_type": 2 00:18:48.832 }, 00:18:48.832 { 00:18:48.832 "dma_device_id": "system", 00:18:48.832 "dma_device_type": 1 00:18:48.832 }, 00:18:48.832 { 00:18:48.832 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:48.832 "dma_device_type": 2 00:18:48.832 }, 00:18:48.832 { 00:18:48.832 "dma_device_id": "system", 00:18:48.832 "dma_device_type": 1 00:18:48.832 }, 00:18:48.832 { 00:18:48.832 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:48.832 "dma_device_type": 2 00:18:48.832 } 00:18:48.832 ], 00:18:48.832 "driver_specific": { 00:18:48.832 "raid": { 00:18:48.832 "uuid": "a9a38945-e8c3-479a-8cd8-ddfe527a18a0", 00:18:48.832 "strip_size_kb": 64, 00:18:48.832 "state": "online", 00:18:48.832 "raid_level": "raid0", 00:18:48.832 "superblock": true, 00:18:48.832 "num_base_bdevs": 4, 00:18:48.832 "num_base_bdevs_discovered": 4, 00:18:48.832 "num_base_bdevs_operational": 4, 00:18:48.832 "base_bdevs_list": [ 00:18:48.832 { 00:18:48.832 "name": "NewBaseBdev", 00:18:48.832 "uuid": "73f20069-6b69-41d5-985a-5509c2f6ca24", 00:18:48.832 "is_configured": true, 00:18:48.832 "data_offset": 2048, 00:18:48.832 "data_size": 63488 00:18:48.832 }, 00:18:48.832 { 00:18:48.832 "name": "BaseBdev2", 00:18:48.832 "uuid": "049f4bec-c5d9-4501-90b7-1e4997ff96cb", 00:18:48.832 "is_configured": true, 00:18:48.832 "data_offset": 2048, 00:18:48.832 "data_size": 63488 00:18:48.832 }, 00:18:48.832 { 00:18:48.832 "name": "BaseBdev3", 00:18:48.832 "uuid": "e1df1ece-c669-479a-84f9-9bacdf7c9131", 00:18:48.832 "is_configured": true, 00:18:48.832 "data_offset": 2048, 00:18:48.832 "data_size": 63488 00:18:48.832 }, 00:18:48.832 { 00:18:48.832 "name": "BaseBdev4", 00:18:48.832 "uuid": "29298378-a06e-47b3-bff6-af8f8c6394c5", 00:18:48.832 "is_configured": true, 00:18:48.832 "data_offset": 2048, 00:18:48.832 "data_size": 63488 00:18:48.832 } 00:18:48.832 ] 00:18:48.832 } 00:18:48.832 } 00:18:48.832 }' 00:18:48.832 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:18:48.832 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # base_bdev_names='NewBaseBdev 00:18:48.832 BaseBdev2 00:18:48.832 BaseBdev3 00:18:48.832 BaseBdev4' 00:18:48.832 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:18:48.832 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev 00:18:48.832 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:18:49.091 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:18:49.091 "name": "NewBaseBdev", 00:18:49.091 "aliases": [ 00:18:49.091 "73f20069-6b69-41d5-985a-5509c2f6ca24" 00:18:49.091 ], 00:18:49.091 "product_name": "Malloc disk", 00:18:49.091 "block_size": 512, 00:18:49.091 "num_blocks": 65536, 00:18:49.091 "uuid": "73f20069-6b69-41d5-985a-5509c2f6ca24", 00:18:49.091 "assigned_rate_limits": { 00:18:49.091 "rw_ios_per_sec": 0, 00:18:49.091 "rw_mbytes_per_sec": 0, 00:18:49.091 "r_mbytes_per_sec": 0, 00:18:49.091 "w_mbytes_per_sec": 0 00:18:49.091 }, 00:18:49.091 "claimed": true, 00:18:49.091 "claim_type": "exclusive_write", 00:18:49.091 "zoned": false, 00:18:49.091 "supported_io_types": { 00:18:49.091 "read": true, 00:18:49.091 "write": true, 00:18:49.091 "unmap": true, 00:18:49.091 "flush": true, 00:18:49.091 "reset": true, 00:18:49.091 "nvme_admin": false, 00:18:49.091 "nvme_io": false, 00:18:49.091 "nvme_io_md": false, 00:18:49.091 "write_zeroes": true, 00:18:49.091 "zcopy": true, 00:18:49.091 "get_zone_info": false, 00:18:49.091 "zone_management": false, 00:18:49.091 "zone_append": false, 00:18:49.091 "compare": false, 00:18:49.091 "compare_and_write": false, 00:18:49.091 "abort": true, 00:18:49.091 "seek_hole": false, 00:18:49.091 "seek_data": false, 00:18:49.091 "copy": true, 00:18:49.091 "nvme_iov_md": false 00:18:49.091 }, 00:18:49.091 "memory_domains": [ 00:18:49.091 { 00:18:49.091 "dma_device_id": "system", 00:18:49.091 "dma_device_type": 1 00:18:49.091 }, 00:18:49.092 { 00:18:49.092 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:49.092 "dma_device_type": 2 00:18:49.092 } 00:18:49.092 ], 00:18:49.092 "driver_specific": {} 00:18:49.092 }' 00:18:49.092 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:49.351 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:49.351 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:18:49.351 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:49.351 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:49.351 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:18:49.351 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:49.351 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:49.351 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:18:49.351 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:49.610 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:49.610 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:18:49.610 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:18:49.610 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:18:49.610 11:29:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:18:49.876 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:18:49.876 "name": "BaseBdev2", 00:18:49.876 "aliases": [ 00:18:49.876 "049f4bec-c5d9-4501-90b7-1e4997ff96cb" 00:18:49.876 ], 00:18:49.876 "product_name": "Malloc disk", 00:18:49.876 "block_size": 512, 00:18:49.876 "num_blocks": 65536, 00:18:49.876 "uuid": "049f4bec-c5d9-4501-90b7-1e4997ff96cb", 00:18:49.876 "assigned_rate_limits": { 00:18:49.876 "rw_ios_per_sec": 0, 00:18:49.876 "rw_mbytes_per_sec": 0, 00:18:49.876 "r_mbytes_per_sec": 0, 00:18:49.876 "w_mbytes_per_sec": 0 00:18:49.876 }, 00:18:49.876 "claimed": true, 00:18:49.876 "claim_type": "exclusive_write", 00:18:49.876 "zoned": false, 00:18:49.876 "supported_io_types": { 00:18:49.876 "read": true, 00:18:49.876 "write": true, 00:18:49.876 "unmap": true, 00:18:49.876 "flush": true, 00:18:49.876 "reset": true, 00:18:49.876 "nvme_admin": false, 00:18:49.876 "nvme_io": false, 00:18:49.876 "nvme_io_md": false, 00:18:49.876 "write_zeroes": true, 00:18:49.876 "zcopy": true, 00:18:49.876 "get_zone_info": false, 00:18:49.876 "zone_management": false, 00:18:49.876 "zone_append": false, 00:18:49.876 "compare": false, 00:18:49.876 "compare_and_write": false, 00:18:49.876 "abort": true, 00:18:49.876 "seek_hole": false, 00:18:49.876 "seek_data": false, 00:18:49.876 "copy": true, 00:18:49.876 "nvme_iov_md": false 00:18:49.876 }, 00:18:49.876 "memory_domains": [ 00:18:49.876 { 00:18:49.876 "dma_device_id": "system", 00:18:49.876 "dma_device_type": 1 00:18:49.876 }, 00:18:49.876 { 00:18:49.876 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:49.876 "dma_device_type": 2 00:18:49.876 } 00:18:49.876 ], 00:18:49.876 "driver_specific": {} 00:18:49.876 }' 00:18:49.876 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:49.876 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:49.876 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:18:49.876 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:49.876 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:49.876 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:18:49.876 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:49.876 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:49.876 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:18:49.876 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:50.137 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:50.137 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:18:50.137 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:18:50.137 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:18:50.137 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:18:50.137 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:18:50.137 "name": "BaseBdev3", 00:18:50.137 "aliases": [ 00:18:50.137 "e1df1ece-c669-479a-84f9-9bacdf7c9131" 00:18:50.137 ], 00:18:50.137 "product_name": "Malloc disk", 00:18:50.137 "block_size": 512, 00:18:50.137 "num_blocks": 65536, 00:18:50.137 "uuid": "e1df1ece-c669-479a-84f9-9bacdf7c9131", 00:18:50.137 "assigned_rate_limits": { 00:18:50.137 "rw_ios_per_sec": 0, 00:18:50.137 "rw_mbytes_per_sec": 0, 00:18:50.137 "r_mbytes_per_sec": 0, 00:18:50.137 "w_mbytes_per_sec": 0 00:18:50.137 }, 00:18:50.137 "claimed": true, 00:18:50.137 "claim_type": "exclusive_write", 00:18:50.137 "zoned": false, 00:18:50.137 "supported_io_types": { 00:18:50.137 "read": true, 00:18:50.137 "write": true, 00:18:50.137 "unmap": true, 00:18:50.137 "flush": true, 00:18:50.137 "reset": true, 00:18:50.137 "nvme_admin": false, 00:18:50.137 "nvme_io": false, 00:18:50.137 "nvme_io_md": false, 00:18:50.137 "write_zeroes": true, 00:18:50.137 "zcopy": true, 00:18:50.137 "get_zone_info": false, 00:18:50.137 "zone_management": false, 00:18:50.137 "zone_append": false, 00:18:50.137 "compare": false, 00:18:50.137 "compare_and_write": false, 00:18:50.137 "abort": true, 00:18:50.137 "seek_hole": false, 00:18:50.137 "seek_data": false, 00:18:50.137 "copy": true, 00:18:50.137 "nvme_iov_md": false 00:18:50.137 }, 00:18:50.137 "memory_domains": [ 00:18:50.137 { 00:18:50.137 "dma_device_id": "system", 00:18:50.137 "dma_device_type": 1 00:18:50.137 }, 00:18:50.137 { 00:18:50.137 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:50.137 "dma_device_type": 2 00:18:50.137 } 00:18:50.137 ], 00:18:50.137 "driver_specific": {} 00:18:50.137 }' 00:18:50.137 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:50.396 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:50.396 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:18:50.396 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:50.396 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:50.396 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:18:50.396 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:50.396 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:50.396 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:18:50.396 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:50.396 11:29:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:50.656 11:29:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:18:50.656 11:29:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:18:50.656 11:29:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 00:18:50.656 11:29:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:18:50.656 11:29:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:18:50.656 "name": "BaseBdev4", 00:18:50.656 "aliases": [ 00:18:50.656 "29298378-a06e-47b3-bff6-af8f8c6394c5" 00:18:50.656 ], 00:18:50.656 "product_name": "Malloc disk", 00:18:50.656 "block_size": 512, 00:18:50.656 "num_blocks": 65536, 00:18:50.656 "uuid": "29298378-a06e-47b3-bff6-af8f8c6394c5", 00:18:50.656 "assigned_rate_limits": { 00:18:50.656 "rw_ios_per_sec": 0, 00:18:50.656 "rw_mbytes_per_sec": 0, 00:18:50.656 "r_mbytes_per_sec": 0, 00:18:50.656 "w_mbytes_per_sec": 0 00:18:50.656 }, 00:18:50.656 "claimed": true, 00:18:50.656 "claim_type": "exclusive_write", 00:18:50.656 "zoned": false, 00:18:50.656 "supported_io_types": { 00:18:50.656 "read": true, 00:18:50.656 "write": true, 00:18:50.656 "unmap": true, 00:18:50.656 "flush": true, 00:18:50.656 "reset": true, 00:18:50.656 "nvme_admin": false, 00:18:50.656 "nvme_io": false, 00:18:50.656 "nvme_io_md": false, 00:18:50.656 "write_zeroes": true, 00:18:50.656 "zcopy": true, 00:18:50.656 "get_zone_info": false, 00:18:50.656 "zone_management": false, 00:18:50.656 "zone_append": false, 00:18:50.656 "compare": false, 00:18:50.656 "compare_and_write": false, 00:18:50.656 "abort": true, 00:18:50.656 "seek_hole": false, 00:18:50.656 "seek_data": false, 00:18:50.656 "copy": true, 00:18:50.656 "nvme_iov_md": false 00:18:50.656 }, 00:18:50.656 "memory_domains": [ 00:18:50.656 { 00:18:50.656 "dma_device_id": "system", 00:18:50.656 "dma_device_type": 1 00:18:50.656 }, 00:18:50.656 { 00:18:50.656 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:50.656 "dma_device_type": 2 00:18:50.656 } 00:18:50.656 ], 00:18:50.656 "driver_specific": {} 00:18:50.656 }' 00:18:50.916 11:29:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:50.916 11:29:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:50.916 11:29:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:18:50.916 11:29:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:50.916 11:29:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:50.916 11:29:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:18:50.916 11:29:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:50.916 11:29:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:50.916 11:29:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:18:50.916 11:29:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:51.176 11:29:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:51.176 11:29:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:18:51.176 11:29:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@338 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:18:51.176 [2024-07-15 11:29:34.728637] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:18:51.176 [2024-07-15 11:29:34.728674] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:18:51.176 [2024-07-15 11:29:34.728746] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:18:51.176 [2024-07-15 11:29:34.728818] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:18:51.176 [2024-07-15 11:29:34.728832] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xe89470 name Existed_Raid, state offline 00:18:51.176 11:29:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@341 -- # killprocess 927846 00:18:51.176 11:29:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@948 -- # '[' -z 927846 ']' 00:18:51.176 11:29:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@952 -- # kill -0 927846 00:18:51.176 11:29:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@953 -- # uname 00:18:51.176 11:29:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:18:51.176 11:29:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 927846 00:18:51.435 11:29:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:18:51.435 11:29:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:18:51.435 11:29:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@966 -- # echo 'killing process with pid 927846' 00:18:51.435 killing process with pid 927846 00:18:51.435 11:29:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@967 -- # kill 927846 00:18:51.435 [2024-07-15 11:29:34.791095] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:18:51.435 11:29:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # wait 927846 00:18:51.435 [2024-07-15 11:29:34.880183] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:18:51.694 11:29:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@343 -- # return 0 00:18:51.694 00:18:51.694 real 0m31.902s 00:18:51.694 user 0m58.297s 00:18:51.694 sys 0m5.678s 00:18:51.694 11:29:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1124 -- # xtrace_disable 00:18:51.694 11:29:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:18:51.694 ************************************ 00:18:51.694 END TEST raid_state_function_test_sb 00:18:51.694 ************************************ 00:18:51.954 11:29:35 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:18:51.954 11:29:35 bdev_raid -- bdev/bdev_raid.sh@869 -- # run_test raid_superblock_test raid_superblock_test raid0 4 00:18:51.954 11:29:35 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:18:51.954 11:29:35 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:18:51.954 11:29:35 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:18:51.954 ************************************ 00:18:51.954 START TEST raid_superblock_test 00:18:51.954 ************************************ 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1123 -- # raid_superblock_test raid0 4 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@392 -- # local raid_level=raid0 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local num_base_bdevs=4 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # base_bdevs_malloc=() 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local base_bdevs_malloc 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_pt=() 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_pt 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt_uuid=() 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt_uuid 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local raid_bdev_name=raid_bdev1 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local strip_size 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size_create_arg 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local raid_bdev_uuid 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@403 -- # '[' raid0 '!=' raid1 ']' 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # strip_size=64 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size_create_arg='-z 64' 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # raid_pid=932612 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # waitforlisten 932612 /var/tmp/spdk-raid.sock 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@410 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -L bdev_raid 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@829 -- # '[' -z 932612 ']' 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:18:51.954 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:18:51.954 11:29:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:18:51.954 [2024-07-15 11:29:35.427227] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:18:51.954 [2024-07-15 11:29:35.427299] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid932612 ] 00:18:52.214 [2024-07-15 11:29:35.557458] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:52.214 [2024-07-15 11:29:35.660254] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:18:52.214 [2024-07-15 11:29:35.715207] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:18:52.214 [2024-07-15 11:29:35.715235] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:18:52.783 11:29:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:18:52.783 11:29:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@862 -- # return 0 00:18:52.783 11:29:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i = 1 )) 00:18:52.783 11:29:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:18:52.783 11:29:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc1 00:18:52.783 11:29:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt1 00:18:52.783 11:29:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:18:52.783 11:29:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:18:52.783 11:29:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:18:52.783 11:29:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:18:52.783 11:29:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc1 00:18:53.042 malloc1 00:18:53.042 11:29:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:18:53.300 [2024-07-15 11:29:36.771488] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:18:53.300 [2024-07-15 11:29:36.771539] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:18:53.300 [2024-07-15 11:29:36.771559] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1743570 00:18:53.300 [2024-07-15 11:29:36.771572] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:18:53.300 [2024-07-15 11:29:36.773216] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:18:53.300 [2024-07-15 11:29:36.773247] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:18:53.300 pt1 00:18:53.300 11:29:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:18:53.300 11:29:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:18:53.300 11:29:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc2 00:18:53.300 11:29:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt2 00:18:53.300 11:29:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:18:53.300 11:29:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:18:53.300 11:29:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:18:53.300 11:29:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:18:53.300 11:29:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc2 00:18:53.559 malloc2 00:18:53.559 11:29:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:18:53.818 [2024-07-15 11:29:37.201485] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:18:53.818 [2024-07-15 11:29:37.201538] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:18:53.818 [2024-07-15 11:29:37.201555] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1744970 00:18:53.818 [2024-07-15 11:29:37.201567] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:18:53.818 [2024-07-15 11:29:37.203111] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:18:53.818 [2024-07-15 11:29:37.203141] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:18:53.818 pt2 00:18:53.818 11:29:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:18:53.818 11:29:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:18:53.818 11:29:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc3 00:18:53.818 11:29:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt3 00:18:53.818 11:29:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:18:53.818 11:29:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:18:53.818 11:29:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:18:53.818 11:29:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:18:53.818 11:29:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc3 00:18:53.818 malloc3 00:18:53.818 11:29:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:18:54.078 [2024-07-15 11:29:37.547034] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:18:54.078 [2024-07-15 11:29:37.547085] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:18:54.078 [2024-07-15 11:29:37.547103] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x18db340 00:18:54.078 [2024-07-15 11:29:37.547115] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:18:54.078 [2024-07-15 11:29:37.548553] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:18:54.078 [2024-07-15 11:29:37.548583] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:18:54.078 pt3 00:18:54.078 11:29:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:18:54.078 11:29:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:18:54.078 11:29:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc4 00:18:54.078 11:29:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt4 00:18:54.078 11:29:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:18:54.078 11:29:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:18:54.078 11:29:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:18:54.078 11:29:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:18:54.078 11:29:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc4 00:18:54.337 malloc4 00:18:54.337 11:29:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:18:54.595 [2024-07-15 11:29:38.048956] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:18:54.595 [2024-07-15 11:29:38.049001] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:18:54.595 [2024-07-15 11:29:38.049020] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x18ddc60 00:18:54.595 [2024-07-15 11:29:38.049033] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:18:54.595 [2024-07-15 11:29:38.050449] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:18:54.595 [2024-07-15 11:29:38.050478] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:18:54.595 pt4 00:18:54.595 11:29:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:18:54.595 11:29:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:18:54.596 11:29:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@429 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'pt1 pt2 pt3 pt4' -n raid_bdev1 -s 00:18:54.853 [2024-07-15 11:29:38.301644] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:18:54.853 [2024-07-15 11:29:38.302827] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:18:54.853 [2024-07-15 11:29:38.302881] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:18:54.853 [2024-07-15 11:29:38.302924] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:18:54.853 [2024-07-15 11:29:38.303096] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x173b530 00:18:54.853 [2024-07-15 11:29:38.303107] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:18:54.853 [2024-07-15 11:29:38.303289] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1739770 00:18:54.853 [2024-07-15 11:29:38.303432] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x173b530 00:18:54.853 [2024-07-15 11:29:38.303442] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x173b530 00:18:54.853 [2024-07-15 11:29:38.303533] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:18:54.853 11:29:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:18:54.853 11:29:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:18:54.853 11:29:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:18:54.853 11:29:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:18:54.853 11:29:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:18:54.853 11:29:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:18:54.853 11:29:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:18:54.853 11:29:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:18:54.853 11:29:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:18:54.853 11:29:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:18:54.853 11:29:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:54.853 11:29:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:18:55.112 11:29:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:18:55.112 "name": "raid_bdev1", 00:18:55.112 "uuid": "5425fa73-79f0-4071-9c37-208b9903b7eb", 00:18:55.112 "strip_size_kb": 64, 00:18:55.112 "state": "online", 00:18:55.112 "raid_level": "raid0", 00:18:55.112 "superblock": true, 00:18:55.112 "num_base_bdevs": 4, 00:18:55.112 "num_base_bdevs_discovered": 4, 00:18:55.112 "num_base_bdevs_operational": 4, 00:18:55.112 "base_bdevs_list": [ 00:18:55.112 { 00:18:55.112 "name": "pt1", 00:18:55.112 "uuid": "00000000-0000-0000-0000-000000000001", 00:18:55.112 "is_configured": true, 00:18:55.112 "data_offset": 2048, 00:18:55.112 "data_size": 63488 00:18:55.112 }, 00:18:55.112 { 00:18:55.112 "name": "pt2", 00:18:55.112 "uuid": "00000000-0000-0000-0000-000000000002", 00:18:55.112 "is_configured": true, 00:18:55.112 "data_offset": 2048, 00:18:55.112 "data_size": 63488 00:18:55.112 }, 00:18:55.112 { 00:18:55.112 "name": "pt3", 00:18:55.112 "uuid": "00000000-0000-0000-0000-000000000003", 00:18:55.112 "is_configured": true, 00:18:55.112 "data_offset": 2048, 00:18:55.112 "data_size": 63488 00:18:55.112 }, 00:18:55.112 { 00:18:55.112 "name": "pt4", 00:18:55.112 "uuid": "00000000-0000-0000-0000-000000000004", 00:18:55.112 "is_configured": true, 00:18:55.112 "data_offset": 2048, 00:18:55.112 "data_size": 63488 00:18:55.112 } 00:18:55.112 ] 00:18:55.112 }' 00:18:55.112 11:29:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:18:55.112 11:29:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:18:55.682 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_properties raid_bdev1 00:18:55.682 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:18:55.682 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:18:55.682 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:18:55.682 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:18:55.682 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # local name 00:18:55.682 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:18:55.682 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:18:55.682 [2024-07-15 11:29:39.156195] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:18:55.682 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:18:55.682 "name": "raid_bdev1", 00:18:55.682 "aliases": [ 00:18:55.682 "5425fa73-79f0-4071-9c37-208b9903b7eb" 00:18:55.682 ], 00:18:55.682 "product_name": "Raid Volume", 00:18:55.682 "block_size": 512, 00:18:55.682 "num_blocks": 253952, 00:18:55.682 "uuid": "5425fa73-79f0-4071-9c37-208b9903b7eb", 00:18:55.682 "assigned_rate_limits": { 00:18:55.682 "rw_ios_per_sec": 0, 00:18:55.682 "rw_mbytes_per_sec": 0, 00:18:55.682 "r_mbytes_per_sec": 0, 00:18:55.682 "w_mbytes_per_sec": 0 00:18:55.682 }, 00:18:55.682 "claimed": false, 00:18:55.682 "zoned": false, 00:18:55.682 "supported_io_types": { 00:18:55.682 "read": true, 00:18:55.682 "write": true, 00:18:55.682 "unmap": true, 00:18:55.682 "flush": true, 00:18:55.682 "reset": true, 00:18:55.682 "nvme_admin": false, 00:18:55.682 "nvme_io": false, 00:18:55.682 "nvme_io_md": false, 00:18:55.682 "write_zeroes": true, 00:18:55.682 "zcopy": false, 00:18:55.682 "get_zone_info": false, 00:18:55.682 "zone_management": false, 00:18:55.682 "zone_append": false, 00:18:55.682 "compare": false, 00:18:55.682 "compare_and_write": false, 00:18:55.682 "abort": false, 00:18:55.682 "seek_hole": false, 00:18:55.682 "seek_data": false, 00:18:55.682 "copy": false, 00:18:55.682 "nvme_iov_md": false 00:18:55.682 }, 00:18:55.682 "memory_domains": [ 00:18:55.682 { 00:18:55.682 "dma_device_id": "system", 00:18:55.682 "dma_device_type": 1 00:18:55.682 }, 00:18:55.682 { 00:18:55.682 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:55.682 "dma_device_type": 2 00:18:55.682 }, 00:18:55.682 { 00:18:55.682 "dma_device_id": "system", 00:18:55.682 "dma_device_type": 1 00:18:55.682 }, 00:18:55.682 { 00:18:55.682 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:55.682 "dma_device_type": 2 00:18:55.682 }, 00:18:55.682 { 00:18:55.682 "dma_device_id": "system", 00:18:55.682 "dma_device_type": 1 00:18:55.682 }, 00:18:55.682 { 00:18:55.682 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:55.682 "dma_device_type": 2 00:18:55.682 }, 00:18:55.682 { 00:18:55.682 "dma_device_id": "system", 00:18:55.682 "dma_device_type": 1 00:18:55.682 }, 00:18:55.682 { 00:18:55.682 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:55.682 "dma_device_type": 2 00:18:55.682 } 00:18:55.682 ], 00:18:55.682 "driver_specific": { 00:18:55.682 "raid": { 00:18:55.682 "uuid": "5425fa73-79f0-4071-9c37-208b9903b7eb", 00:18:55.682 "strip_size_kb": 64, 00:18:55.682 "state": "online", 00:18:55.682 "raid_level": "raid0", 00:18:55.682 "superblock": true, 00:18:55.682 "num_base_bdevs": 4, 00:18:55.682 "num_base_bdevs_discovered": 4, 00:18:55.682 "num_base_bdevs_operational": 4, 00:18:55.682 "base_bdevs_list": [ 00:18:55.682 { 00:18:55.682 "name": "pt1", 00:18:55.682 "uuid": "00000000-0000-0000-0000-000000000001", 00:18:55.682 "is_configured": true, 00:18:55.682 "data_offset": 2048, 00:18:55.682 "data_size": 63488 00:18:55.682 }, 00:18:55.682 { 00:18:55.682 "name": "pt2", 00:18:55.682 "uuid": "00000000-0000-0000-0000-000000000002", 00:18:55.682 "is_configured": true, 00:18:55.682 "data_offset": 2048, 00:18:55.682 "data_size": 63488 00:18:55.682 }, 00:18:55.682 { 00:18:55.682 "name": "pt3", 00:18:55.682 "uuid": "00000000-0000-0000-0000-000000000003", 00:18:55.682 "is_configured": true, 00:18:55.682 "data_offset": 2048, 00:18:55.682 "data_size": 63488 00:18:55.682 }, 00:18:55.682 { 00:18:55.682 "name": "pt4", 00:18:55.682 "uuid": "00000000-0000-0000-0000-000000000004", 00:18:55.682 "is_configured": true, 00:18:55.682 "data_offset": 2048, 00:18:55.682 "data_size": 63488 00:18:55.682 } 00:18:55.682 ] 00:18:55.682 } 00:18:55.682 } 00:18:55.682 }' 00:18:55.682 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:18:55.682 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:18:55.682 pt2 00:18:55.682 pt3 00:18:55.682 pt4' 00:18:55.682 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:18:55.682 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:18:55.682 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:18:55.974 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:18:55.974 "name": "pt1", 00:18:55.974 "aliases": [ 00:18:55.974 "00000000-0000-0000-0000-000000000001" 00:18:55.974 ], 00:18:55.974 "product_name": "passthru", 00:18:55.974 "block_size": 512, 00:18:55.974 "num_blocks": 65536, 00:18:55.974 "uuid": "00000000-0000-0000-0000-000000000001", 00:18:55.974 "assigned_rate_limits": { 00:18:55.974 "rw_ios_per_sec": 0, 00:18:55.974 "rw_mbytes_per_sec": 0, 00:18:55.974 "r_mbytes_per_sec": 0, 00:18:55.974 "w_mbytes_per_sec": 0 00:18:55.974 }, 00:18:55.974 "claimed": true, 00:18:55.974 "claim_type": "exclusive_write", 00:18:55.974 "zoned": false, 00:18:55.974 "supported_io_types": { 00:18:55.974 "read": true, 00:18:55.974 "write": true, 00:18:55.974 "unmap": true, 00:18:55.974 "flush": true, 00:18:55.974 "reset": true, 00:18:55.974 "nvme_admin": false, 00:18:55.974 "nvme_io": false, 00:18:55.974 "nvme_io_md": false, 00:18:55.974 "write_zeroes": true, 00:18:55.974 "zcopy": true, 00:18:55.974 "get_zone_info": false, 00:18:55.974 "zone_management": false, 00:18:55.974 "zone_append": false, 00:18:55.974 "compare": false, 00:18:55.974 "compare_and_write": false, 00:18:55.974 "abort": true, 00:18:55.974 "seek_hole": false, 00:18:55.974 "seek_data": false, 00:18:55.974 "copy": true, 00:18:55.974 "nvme_iov_md": false 00:18:55.974 }, 00:18:55.974 "memory_domains": [ 00:18:55.974 { 00:18:55.974 "dma_device_id": "system", 00:18:55.974 "dma_device_type": 1 00:18:55.974 }, 00:18:55.974 { 00:18:55.974 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:55.974 "dma_device_type": 2 00:18:55.974 } 00:18:55.974 ], 00:18:55.974 "driver_specific": { 00:18:55.974 "passthru": { 00:18:55.974 "name": "pt1", 00:18:55.974 "base_bdev_name": "malloc1" 00:18:55.974 } 00:18:55.974 } 00:18:55.974 }' 00:18:55.974 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:55.974 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:55.974 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:18:55.974 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:56.232 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:56.232 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:18:56.232 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:56.232 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:56.232 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:18:56.232 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:56.232 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:56.232 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:18:56.232 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:18:56.232 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:18:56.232 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:18:56.490 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:18:56.490 "name": "pt2", 00:18:56.490 "aliases": [ 00:18:56.490 "00000000-0000-0000-0000-000000000002" 00:18:56.490 ], 00:18:56.490 "product_name": "passthru", 00:18:56.490 "block_size": 512, 00:18:56.490 "num_blocks": 65536, 00:18:56.490 "uuid": "00000000-0000-0000-0000-000000000002", 00:18:56.490 "assigned_rate_limits": { 00:18:56.490 "rw_ios_per_sec": 0, 00:18:56.490 "rw_mbytes_per_sec": 0, 00:18:56.490 "r_mbytes_per_sec": 0, 00:18:56.490 "w_mbytes_per_sec": 0 00:18:56.490 }, 00:18:56.490 "claimed": true, 00:18:56.490 "claim_type": "exclusive_write", 00:18:56.490 "zoned": false, 00:18:56.490 "supported_io_types": { 00:18:56.490 "read": true, 00:18:56.490 "write": true, 00:18:56.490 "unmap": true, 00:18:56.490 "flush": true, 00:18:56.490 "reset": true, 00:18:56.490 "nvme_admin": false, 00:18:56.490 "nvme_io": false, 00:18:56.490 "nvme_io_md": false, 00:18:56.490 "write_zeroes": true, 00:18:56.490 "zcopy": true, 00:18:56.490 "get_zone_info": false, 00:18:56.490 "zone_management": false, 00:18:56.490 "zone_append": false, 00:18:56.490 "compare": false, 00:18:56.490 "compare_and_write": false, 00:18:56.490 "abort": true, 00:18:56.490 "seek_hole": false, 00:18:56.490 "seek_data": false, 00:18:56.490 "copy": true, 00:18:56.490 "nvme_iov_md": false 00:18:56.490 }, 00:18:56.490 "memory_domains": [ 00:18:56.490 { 00:18:56.490 "dma_device_id": "system", 00:18:56.490 "dma_device_type": 1 00:18:56.490 }, 00:18:56.490 { 00:18:56.490 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:56.490 "dma_device_type": 2 00:18:56.490 } 00:18:56.490 ], 00:18:56.490 "driver_specific": { 00:18:56.490 "passthru": { 00:18:56.490 "name": "pt2", 00:18:56.490 "base_bdev_name": "malloc2" 00:18:56.490 } 00:18:56.490 } 00:18:56.491 }' 00:18:56.491 11:29:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:56.491 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:56.491 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:18:56.491 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:56.749 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:56.749 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:18:56.749 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:56.749 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:56.749 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:18:56.749 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:56.749 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:56.749 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:18:56.749 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:18:56.749 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt3 00:18:56.749 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:18:57.008 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:18:57.008 "name": "pt3", 00:18:57.008 "aliases": [ 00:18:57.008 "00000000-0000-0000-0000-000000000003" 00:18:57.008 ], 00:18:57.008 "product_name": "passthru", 00:18:57.008 "block_size": 512, 00:18:57.008 "num_blocks": 65536, 00:18:57.008 "uuid": "00000000-0000-0000-0000-000000000003", 00:18:57.008 "assigned_rate_limits": { 00:18:57.008 "rw_ios_per_sec": 0, 00:18:57.008 "rw_mbytes_per_sec": 0, 00:18:57.008 "r_mbytes_per_sec": 0, 00:18:57.008 "w_mbytes_per_sec": 0 00:18:57.008 }, 00:18:57.008 "claimed": true, 00:18:57.008 "claim_type": "exclusive_write", 00:18:57.008 "zoned": false, 00:18:57.008 "supported_io_types": { 00:18:57.008 "read": true, 00:18:57.008 "write": true, 00:18:57.008 "unmap": true, 00:18:57.008 "flush": true, 00:18:57.008 "reset": true, 00:18:57.008 "nvme_admin": false, 00:18:57.008 "nvme_io": false, 00:18:57.008 "nvme_io_md": false, 00:18:57.008 "write_zeroes": true, 00:18:57.008 "zcopy": true, 00:18:57.008 "get_zone_info": false, 00:18:57.008 "zone_management": false, 00:18:57.008 "zone_append": false, 00:18:57.008 "compare": false, 00:18:57.008 "compare_and_write": false, 00:18:57.008 "abort": true, 00:18:57.008 "seek_hole": false, 00:18:57.008 "seek_data": false, 00:18:57.008 "copy": true, 00:18:57.008 "nvme_iov_md": false 00:18:57.008 }, 00:18:57.008 "memory_domains": [ 00:18:57.008 { 00:18:57.008 "dma_device_id": "system", 00:18:57.008 "dma_device_type": 1 00:18:57.008 }, 00:18:57.008 { 00:18:57.008 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:57.008 "dma_device_type": 2 00:18:57.008 } 00:18:57.008 ], 00:18:57.008 "driver_specific": { 00:18:57.008 "passthru": { 00:18:57.008 "name": "pt3", 00:18:57.008 "base_bdev_name": "malloc3" 00:18:57.008 } 00:18:57.008 } 00:18:57.008 }' 00:18:57.008 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:57.266 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:57.266 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:18:57.266 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:57.266 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:57.266 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:18:57.266 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:57.266 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:57.266 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:18:57.266 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:57.266 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:57.524 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:18:57.524 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:18:57.524 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt4 00:18:57.524 11:29:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:18:57.783 11:29:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:18:57.783 "name": "pt4", 00:18:57.783 "aliases": [ 00:18:57.783 "00000000-0000-0000-0000-000000000004" 00:18:57.783 ], 00:18:57.783 "product_name": "passthru", 00:18:57.783 "block_size": 512, 00:18:57.783 "num_blocks": 65536, 00:18:57.783 "uuid": "00000000-0000-0000-0000-000000000004", 00:18:57.783 "assigned_rate_limits": { 00:18:57.783 "rw_ios_per_sec": 0, 00:18:57.783 "rw_mbytes_per_sec": 0, 00:18:57.783 "r_mbytes_per_sec": 0, 00:18:57.783 "w_mbytes_per_sec": 0 00:18:57.783 }, 00:18:57.783 "claimed": true, 00:18:57.783 "claim_type": "exclusive_write", 00:18:57.783 "zoned": false, 00:18:57.783 "supported_io_types": { 00:18:57.783 "read": true, 00:18:57.783 "write": true, 00:18:57.783 "unmap": true, 00:18:57.783 "flush": true, 00:18:57.783 "reset": true, 00:18:57.783 "nvme_admin": false, 00:18:57.783 "nvme_io": false, 00:18:57.783 "nvme_io_md": false, 00:18:57.783 "write_zeroes": true, 00:18:57.783 "zcopy": true, 00:18:57.783 "get_zone_info": false, 00:18:57.783 "zone_management": false, 00:18:57.783 "zone_append": false, 00:18:57.783 "compare": false, 00:18:57.783 "compare_and_write": false, 00:18:57.783 "abort": true, 00:18:57.783 "seek_hole": false, 00:18:57.783 "seek_data": false, 00:18:57.783 "copy": true, 00:18:57.783 "nvme_iov_md": false 00:18:57.783 }, 00:18:57.783 "memory_domains": [ 00:18:57.783 { 00:18:57.783 "dma_device_id": "system", 00:18:57.783 "dma_device_type": 1 00:18:57.783 }, 00:18:57.783 { 00:18:57.783 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:18:57.783 "dma_device_type": 2 00:18:57.783 } 00:18:57.783 ], 00:18:57.783 "driver_specific": { 00:18:57.783 "passthru": { 00:18:57.783 "name": "pt4", 00:18:57.783 "base_bdev_name": "malloc4" 00:18:57.783 } 00:18:57.783 } 00:18:57.783 }' 00:18:57.783 11:29:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:57.783 11:29:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:18:57.783 11:29:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:18:57.783 11:29:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:57.783 11:29:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:18:57.783 11:29:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:18:57.783 11:29:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:57.783 11:29:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:18:58.042 11:29:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:18:58.042 11:29:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:58.042 11:29:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:18:58.042 11:29:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:18:58.042 11:29:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:18:58.042 11:29:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # jq -r '.[] | .uuid' 00:18:58.301 [2024-07-15 11:29:41.710970] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:18:58.301 11:29:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # raid_bdev_uuid=5425fa73-79f0-4071-9c37-208b9903b7eb 00:18:58.301 11:29:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # '[' -z 5425fa73-79f0-4071-9c37-208b9903b7eb ']' 00:18:58.301 11:29:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@440 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:18:58.608 [2024-07-15 11:29:41.955298] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:18:58.608 [2024-07-15 11:29:41.955318] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:18:58.608 [2024-07-15 11:29:41.955369] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:18:58.608 [2024-07-15 11:29:41.955432] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:18:58.608 [2024-07-15 11:29:41.955444] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x173b530 name raid_bdev1, state offline 00:18:58.608 11:29:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:18:58.608 11:29:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # jq -r '.[]' 00:18:58.867 11:29:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # raid_bdev= 00:18:58.867 11:29:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # '[' -n '' ']' 00:18:58.867 11:29:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:18:58.867 11:29:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt1 00:18:59.125 11:29:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:18:59.125 11:29:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:18:59.125 11:29:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:18:59.125 11:29:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt3 00:18:59.382 11:29:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:18:59.382 11:29:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt4 00:18:59.642 11:29:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs 00:18:59.642 11:29:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:18:59.901 11:29:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # '[' false == true ']' 00:18:59.901 11:29:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@456 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'malloc1 malloc2 malloc3 malloc4' -n raid_bdev1 00:18:59.901 11:29:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@648 -- # local es=0 00:18:59.901 11:29:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'malloc1 malloc2 malloc3 malloc4' -n raid_bdev1 00:18:59.901 11:29:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:18:59.901 11:29:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:18:59.901 11:29:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:18:59.901 11:29:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:18:59.901 11:29:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:18:59.901 11:29:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:18:59.901 11:29:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:18:59.901 11:29:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py ]] 00:18:59.901 11:29:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'malloc1 malloc2 malloc3 malloc4' -n raid_bdev1 00:19:00.159 [2024-07-15 11:29:43.671790] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:19:00.159 [2024-07-15 11:29:43.673211] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:19:00.159 [2024-07-15 11:29:43.673257] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:19:00.160 [2024-07-15 11:29:43.673291] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:19:00.160 [2024-07-15 11:29:43.673335] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:19:00.160 [2024-07-15 11:29:43.673374] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:19:00.160 [2024-07-15 11:29:43.673397] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:19:00.160 [2024-07-15 11:29:43.673418] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:19:00.160 [2024-07-15 11:29:43.673436] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:19:00.160 [2024-07-15 11:29:43.673447] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x18e6ff0 name raid_bdev1, state configuring 00:19:00.160 request: 00:19:00.160 { 00:19:00.160 "name": "raid_bdev1", 00:19:00.160 "raid_level": "raid0", 00:19:00.160 "base_bdevs": [ 00:19:00.160 "malloc1", 00:19:00.160 "malloc2", 00:19:00.160 "malloc3", 00:19:00.160 "malloc4" 00:19:00.160 ], 00:19:00.160 "strip_size_kb": 64, 00:19:00.160 "superblock": false, 00:19:00.160 "method": "bdev_raid_create", 00:19:00.160 "req_id": 1 00:19:00.160 } 00:19:00.160 Got JSON-RPC error response 00:19:00.160 response: 00:19:00.160 { 00:19:00.160 "code": -17, 00:19:00.160 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:19:00.160 } 00:19:00.160 11:29:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@651 -- # es=1 00:19:00.160 11:29:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:19:00.160 11:29:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:19:00.160 11:29:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:19:00.160 11:29:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:00.160 11:29:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # jq -r '.[]' 00:19:00.418 11:29:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # raid_bdev= 00:19:00.418 11:29:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # '[' -n '' ']' 00:19:00.418 11:29:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@464 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:19:00.677 [2024-07-15 11:29:44.165014] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:19:00.677 [2024-07-15 11:29:44.165057] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:19:00.677 [2024-07-15 11:29:44.165078] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x17437a0 00:19:00.677 [2024-07-15 11:29:44.165091] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:19:00.677 [2024-07-15 11:29:44.166753] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:19:00.677 [2024-07-15 11:29:44.166785] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:19:00.677 [2024-07-15 11:29:44.166856] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:19:00.677 [2024-07-15 11:29:44.166886] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:19:00.677 pt1 00:19:00.677 11:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@467 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:19:00.677 11:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:19:00.677 11:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:19:00.677 11:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:19:00.677 11:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:00.677 11:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:00.677 11:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:00.677 11:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:00.677 11:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:00.677 11:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:00.677 11:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:00.677 11:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:19:00.935 11:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:00.935 "name": "raid_bdev1", 00:19:00.935 "uuid": "5425fa73-79f0-4071-9c37-208b9903b7eb", 00:19:00.935 "strip_size_kb": 64, 00:19:00.935 "state": "configuring", 00:19:00.935 "raid_level": "raid0", 00:19:00.935 "superblock": true, 00:19:00.935 "num_base_bdevs": 4, 00:19:00.935 "num_base_bdevs_discovered": 1, 00:19:00.935 "num_base_bdevs_operational": 4, 00:19:00.935 "base_bdevs_list": [ 00:19:00.935 { 00:19:00.935 "name": "pt1", 00:19:00.935 "uuid": "00000000-0000-0000-0000-000000000001", 00:19:00.935 "is_configured": true, 00:19:00.935 "data_offset": 2048, 00:19:00.935 "data_size": 63488 00:19:00.935 }, 00:19:00.935 { 00:19:00.935 "name": null, 00:19:00.935 "uuid": "00000000-0000-0000-0000-000000000002", 00:19:00.935 "is_configured": false, 00:19:00.935 "data_offset": 2048, 00:19:00.935 "data_size": 63488 00:19:00.935 }, 00:19:00.935 { 00:19:00.935 "name": null, 00:19:00.935 "uuid": "00000000-0000-0000-0000-000000000003", 00:19:00.935 "is_configured": false, 00:19:00.935 "data_offset": 2048, 00:19:00.935 "data_size": 63488 00:19:00.935 }, 00:19:00.935 { 00:19:00.935 "name": null, 00:19:00.935 "uuid": "00000000-0000-0000-0000-000000000004", 00:19:00.935 "is_configured": false, 00:19:00.935 "data_offset": 2048, 00:19:00.935 "data_size": 63488 00:19:00.935 } 00:19:00.935 ] 00:19:00.935 }' 00:19:00.935 11:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:00.936 11:29:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:01.503 11:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@469 -- # '[' 4 -gt 2 ']' 00:19:01.503 11:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@471 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:19:01.762 [2024-07-15 11:29:45.223841] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:19:01.762 [2024-07-15 11:29:45.223893] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:19:01.762 [2024-07-15 11:29:45.223914] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x18dc940 00:19:01.762 [2024-07-15 11:29:45.223932] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:19:01.762 [2024-07-15 11:29:45.224282] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:19:01.762 [2024-07-15 11:29:45.224300] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:19:01.762 [2024-07-15 11:29:45.224362] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:19:01.762 [2024-07-15 11:29:45.224381] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:19:01.762 pt2 00:19:01.762 11:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:19:02.021 [2024-07-15 11:29:45.464508] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:19:02.021 11:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:19:02.021 11:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:19:02.021 11:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:19:02.021 11:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:19:02.021 11:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:02.021 11:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:02.021 11:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:02.021 11:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:02.021 11:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:02.021 11:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:02.021 11:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:02.021 11:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:19:02.280 11:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:02.280 "name": "raid_bdev1", 00:19:02.280 "uuid": "5425fa73-79f0-4071-9c37-208b9903b7eb", 00:19:02.280 "strip_size_kb": 64, 00:19:02.280 "state": "configuring", 00:19:02.280 "raid_level": "raid0", 00:19:02.280 "superblock": true, 00:19:02.280 "num_base_bdevs": 4, 00:19:02.280 "num_base_bdevs_discovered": 1, 00:19:02.280 "num_base_bdevs_operational": 4, 00:19:02.280 "base_bdevs_list": [ 00:19:02.280 { 00:19:02.280 "name": "pt1", 00:19:02.280 "uuid": "00000000-0000-0000-0000-000000000001", 00:19:02.280 "is_configured": true, 00:19:02.280 "data_offset": 2048, 00:19:02.280 "data_size": 63488 00:19:02.280 }, 00:19:02.280 { 00:19:02.280 "name": null, 00:19:02.280 "uuid": "00000000-0000-0000-0000-000000000002", 00:19:02.280 "is_configured": false, 00:19:02.280 "data_offset": 2048, 00:19:02.280 "data_size": 63488 00:19:02.280 }, 00:19:02.280 { 00:19:02.280 "name": null, 00:19:02.280 "uuid": "00000000-0000-0000-0000-000000000003", 00:19:02.280 "is_configured": false, 00:19:02.281 "data_offset": 2048, 00:19:02.281 "data_size": 63488 00:19:02.281 }, 00:19:02.281 { 00:19:02.281 "name": null, 00:19:02.281 "uuid": "00000000-0000-0000-0000-000000000004", 00:19:02.281 "is_configured": false, 00:19:02.281 "data_offset": 2048, 00:19:02.281 "data_size": 63488 00:19:02.281 } 00:19:02.281 ] 00:19:02.281 }' 00:19:02.281 11:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:02.281 11:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:02.848 11:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i = 1 )) 00:19:02.848 11:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:19:02.848 11:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:19:03.108 [2024-07-15 11:29:46.551358] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:19:03.108 [2024-07-15 11:29:46.551407] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:19:03.108 [2024-07-15 11:29:46.551426] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x173a060 00:19:03.108 [2024-07-15 11:29:46.551439] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:19:03.108 [2024-07-15 11:29:46.551773] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:19:03.108 [2024-07-15 11:29:46.551790] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:19:03.108 [2024-07-15 11:29:46.551850] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:19:03.108 [2024-07-15 11:29:46.551868] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:19:03.108 pt2 00:19:03.108 11:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:19:03.108 11:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:19:03.108 11:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:19:03.366 [2024-07-15 11:29:46.796012] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:19:03.366 [2024-07-15 11:29:46.796045] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:19:03.366 [2024-07-15 11:29:46.796065] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x173c8d0 00:19:03.366 [2024-07-15 11:29:46.796077] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:19:03.366 [2024-07-15 11:29:46.796371] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:19:03.366 [2024-07-15 11:29:46.796388] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:19:03.366 [2024-07-15 11:29:46.796446] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:19:03.366 [2024-07-15 11:29:46.796463] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:19:03.366 pt3 00:19:03.367 11:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:19:03.367 11:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:19:03.367 11:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:19:03.626 [2024-07-15 11:29:47.040669] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:19:03.626 [2024-07-15 11:29:47.040702] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:19:03.626 [2024-07-15 11:29:47.040718] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x173db80 00:19:03.626 [2024-07-15 11:29:47.040730] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:19:03.626 [2024-07-15 11:29:47.041029] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:19:03.626 [2024-07-15 11:29:47.041047] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:19:03.626 [2024-07-15 11:29:47.041099] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:19:03.626 [2024-07-15 11:29:47.041117] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:19:03.626 [2024-07-15 11:29:47.041233] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x173a780 00:19:03.626 [2024-07-15 11:29:47.041243] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:19:03.626 [2024-07-15 11:29:47.041415] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x173fd70 00:19:03.626 [2024-07-15 11:29:47.041543] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x173a780 00:19:03.626 [2024-07-15 11:29:47.041552] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x173a780 00:19:03.626 [2024-07-15 11:29:47.041647] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:03.626 pt4 00:19:03.626 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:19:03.626 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:19:03.626 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@482 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:19:03.626 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:19:03.626 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:19:03.626 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:19:03.626 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:03.626 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:03.626 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:03.626 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:03.626 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:03.626 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:03.626 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:03.626 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:19:03.885 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:03.885 "name": "raid_bdev1", 00:19:03.885 "uuid": "5425fa73-79f0-4071-9c37-208b9903b7eb", 00:19:03.885 "strip_size_kb": 64, 00:19:03.885 "state": "online", 00:19:03.885 "raid_level": "raid0", 00:19:03.885 "superblock": true, 00:19:03.885 "num_base_bdevs": 4, 00:19:03.885 "num_base_bdevs_discovered": 4, 00:19:03.885 "num_base_bdevs_operational": 4, 00:19:03.885 "base_bdevs_list": [ 00:19:03.885 { 00:19:03.885 "name": "pt1", 00:19:03.885 "uuid": "00000000-0000-0000-0000-000000000001", 00:19:03.885 "is_configured": true, 00:19:03.885 "data_offset": 2048, 00:19:03.885 "data_size": 63488 00:19:03.885 }, 00:19:03.885 { 00:19:03.885 "name": "pt2", 00:19:03.886 "uuid": "00000000-0000-0000-0000-000000000002", 00:19:03.886 "is_configured": true, 00:19:03.886 "data_offset": 2048, 00:19:03.886 "data_size": 63488 00:19:03.886 }, 00:19:03.886 { 00:19:03.886 "name": "pt3", 00:19:03.886 "uuid": "00000000-0000-0000-0000-000000000003", 00:19:03.886 "is_configured": true, 00:19:03.886 "data_offset": 2048, 00:19:03.886 "data_size": 63488 00:19:03.886 }, 00:19:03.886 { 00:19:03.886 "name": "pt4", 00:19:03.886 "uuid": "00000000-0000-0000-0000-000000000004", 00:19:03.886 "is_configured": true, 00:19:03.886 "data_offset": 2048, 00:19:03.886 "data_size": 63488 00:19:03.886 } 00:19:03.886 ] 00:19:03.886 }' 00:19:03.886 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:03.886 11:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:04.451 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_properties raid_bdev1 00:19:04.451 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:19:04.451 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:19:04.451 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:19:04.451 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:19:04.451 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # local name 00:19:04.451 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:19:04.451 11:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:19:04.709 [2024-07-15 11:29:48.135889] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:19:04.709 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:19:04.709 "name": "raid_bdev1", 00:19:04.709 "aliases": [ 00:19:04.709 "5425fa73-79f0-4071-9c37-208b9903b7eb" 00:19:04.709 ], 00:19:04.709 "product_name": "Raid Volume", 00:19:04.709 "block_size": 512, 00:19:04.709 "num_blocks": 253952, 00:19:04.709 "uuid": "5425fa73-79f0-4071-9c37-208b9903b7eb", 00:19:04.709 "assigned_rate_limits": { 00:19:04.709 "rw_ios_per_sec": 0, 00:19:04.709 "rw_mbytes_per_sec": 0, 00:19:04.709 "r_mbytes_per_sec": 0, 00:19:04.709 "w_mbytes_per_sec": 0 00:19:04.709 }, 00:19:04.709 "claimed": false, 00:19:04.709 "zoned": false, 00:19:04.709 "supported_io_types": { 00:19:04.709 "read": true, 00:19:04.709 "write": true, 00:19:04.709 "unmap": true, 00:19:04.709 "flush": true, 00:19:04.709 "reset": true, 00:19:04.709 "nvme_admin": false, 00:19:04.709 "nvme_io": false, 00:19:04.709 "nvme_io_md": false, 00:19:04.709 "write_zeroes": true, 00:19:04.709 "zcopy": false, 00:19:04.709 "get_zone_info": false, 00:19:04.709 "zone_management": false, 00:19:04.709 "zone_append": false, 00:19:04.709 "compare": false, 00:19:04.709 "compare_and_write": false, 00:19:04.709 "abort": false, 00:19:04.709 "seek_hole": false, 00:19:04.709 "seek_data": false, 00:19:04.709 "copy": false, 00:19:04.709 "nvme_iov_md": false 00:19:04.709 }, 00:19:04.709 "memory_domains": [ 00:19:04.709 { 00:19:04.709 "dma_device_id": "system", 00:19:04.709 "dma_device_type": 1 00:19:04.709 }, 00:19:04.709 { 00:19:04.709 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:04.709 "dma_device_type": 2 00:19:04.709 }, 00:19:04.709 { 00:19:04.709 "dma_device_id": "system", 00:19:04.709 "dma_device_type": 1 00:19:04.709 }, 00:19:04.709 { 00:19:04.709 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:04.709 "dma_device_type": 2 00:19:04.709 }, 00:19:04.709 { 00:19:04.709 "dma_device_id": "system", 00:19:04.709 "dma_device_type": 1 00:19:04.709 }, 00:19:04.709 { 00:19:04.709 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:04.709 "dma_device_type": 2 00:19:04.709 }, 00:19:04.709 { 00:19:04.709 "dma_device_id": "system", 00:19:04.709 "dma_device_type": 1 00:19:04.709 }, 00:19:04.709 { 00:19:04.709 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:04.709 "dma_device_type": 2 00:19:04.709 } 00:19:04.709 ], 00:19:04.709 "driver_specific": { 00:19:04.709 "raid": { 00:19:04.709 "uuid": "5425fa73-79f0-4071-9c37-208b9903b7eb", 00:19:04.709 "strip_size_kb": 64, 00:19:04.709 "state": "online", 00:19:04.709 "raid_level": "raid0", 00:19:04.709 "superblock": true, 00:19:04.709 "num_base_bdevs": 4, 00:19:04.709 "num_base_bdevs_discovered": 4, 00:19:04.709 "num_base_bdevs_operational": 4, 00:19:04.709 "base_bdevs_list": [ 00:19:04.709 { 00:19:04.709 "name": "pt1", 00:19:04.709 "uuid": "00000000-0000-0000-0000-000000000001", 00:19:04.709 "is_configured": true, 00:19:04.709 "data_offset": 2048, 00:19:04.709 "data_size": 63488 00:19:04.709 }, 00:19:04.709 { 00:19:04.709 "name": "pt2", 00:19:04.709 "uuid": "00000000-0000-0000-0000-000000000002", 00:19:04.709 "is_configured": true, 00:19:04.709 "data_offset": 2048, 00:19:04.709 "data_size": 63488 00:19:04.709 }, 00:19:04.709 { 00:19:04.709 "name": "pt3", 00:19:04.709 "uuid": "00000000-0000-0000-0000-000000000003", 00:19:04.709 "is_configured": true, 00:19:04.709 "data_offset": 2048, 00:19:04.709 "data_size": 63488 00:19:04.709 }, 00:19:04.709 { 00:19:04.709 "name": "pt4", 00:19:04.709 "uuid": "00000000-0000-0000-0000-000000000004", 00:19:04.709 "is_configured": true, 00:19:04.709 "data_offset": 2048, 00:19:04.709 "data_size": 63488 00:19:04.709 } 00:19:04.709 ] 00:19:04.709 } 00:19:04.709 } 00:19:04.709 }' 00:19:04.709 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:19:04.709 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:19:04.709 pt2 00:19:04.709 pt3 00:19:04.709 pt4' 00:19:04.709 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:19:04.709 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:19:04.709 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:19:04.967 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:19:04.967 "name": "pt1", 00:19:04.967 "aliases": [ 00:19:04.967 "00000000-0000-0000-0000-000000000001" 00:19:04.967 ], 00:19:04.967 "product_name": "passthru", 00:19:04.967 "block_size": 512, 00:19:04.967 "num_blocks": 65536, 00:19:04.967 "uuid": "00000000-0000-0000-0000-000000000001", 00:19:04.967 "assigned_rate_limits": { 00:19:04.967 "rw_ios_per_sec": 0, 00:19:04.967 "rw_mbytes_per_sec": 0, 00:19:04.967 "r_mbytes_per_sec": 0, 00:19:04.967 "w_mbytes_per_sec": 0 00:19:04.967 }, 00:19:04.967 "claimed": true, 00:19:04.967 "claim_type": "exclusive_write", 00:19:04.967 "zoned": false, 00:19:04.967 "supported_io_types": { 00:19:04.967 "read": true, 00:19:04.967 "write": true, 00:19:04.967 "unmap": true, 00:19:04.967 "flush": true, 00:19:04.967 "reset": true, 00:19:04.967 "nvme_admin": false, 00:19:04.967 "nvme_io": false, 00:19:04.967 "nvme_io_md": false, 00:19:04.967 "write_zeroes": true, 00:19:04.967 "zcopy": true, 00:19:04.967 "get_zone_info": false, 00:19:04.967 "zone_management": false, 00:19:04.967 "zone_append": false, 00:19:04.967 "compare": false, 00:19:04.967 "compare_and_write": false, 00:19:04.967 "abort": true, 00:19:04.967 "seek_hole": false, 00:19:04.967 "seek_data": false, 00:19:04.967 "copy": true, 00:19:04.967 "nvme_iov_md": false 00:19:04.967 }, 00:19:04.967 "memory_domains": [ 00:19:04.967 { 00:19:04.967 "dma_device_id": "system", 00:19:04.967 "dma_device_type": 1 00:19:04.967 }, 00:19:04.967 { 00:19:04.967 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:04.967 "dma_device_type": 2 00:19:04.967 } 00:19:04.967 ], 00:19:04.967 "driver_specific": { 00:19:04.967 "passthru": { 00:19:04.967 "name": "pt1", 00:19:04.967 "base_bdev_name": "malloc1" 00:19:04.967 } 00:19:04.967 } 00:19:04.967 }' 00:19:04.967 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:04.967 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:04.967 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:19:04.967 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:05.225 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:05.225 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:19:05.225 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:05.225 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:05.225 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:19:05.225 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:05.225 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:05.225 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:19:05.225 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:19:05.225 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:19:05.225 11:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:19:05.483 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:19:05.483 "name": "pt2", 00:19:05.483 "aliases": [ 00:19:05.483 "00000000-0000-0000-0000-000000000002" 00:19:05.483 ], 00:19:05.483 "product_name": "passthru", 00:19:05.483 "block_size": 512, 00:19:05.483 "num_blocks": 65536, 00:19:05.483 "uuid": "00000000-0000-0000-0000-000000000002", 00:19:05.483 "assigned_rate_limits": { 00:19:05.483 "rw_ios_per_sec": 0, 00:19:05.483 "rw_mbytes_per_sec": 0, 00:19:05.483 "r_mbytes_per_sec": 0, 00:19:05.483 "w_mbytes_per_sec": 0 00:19:05.483 }, 00:19:05.483 "claimed": true, 00:19:05.483 "claim_type": "exclusive_write", 00:19:05.483 "zoned": false, 00:19:05.483 "supported_io_types": { 00:19:05.483 "read": true, 00:19:05.483 "write": true, 00:19:05.483 "unmap": true, 00:19:05.483 "flush": true, 00:19:05.483 "reset": true, 00:19:05.483 "nvme_admin": false, 00:19:05.483 "nvme_io": false, 00:19:05.483 "nvme_io_md": false, 00:19:05.483 "write_zeroes": true, 00:19:05.483 "zcopy": true, 00:19:05.483 "get_zone_info": false, 00:19:05.483 "zone_management": false, 00:19:05.483 "zone_append": false, 00:19:05.483 "compare": false, 00:19:05.483 "compare_and_write": false, 00:19:05.483 "abort": true, 00:19:05.483 "seek_hole": false, 00:19:05.483 "seek_data": false, 00:19:05.483 "copy": true, 00:19:05.483 "nvme_iov_md": false 00:19:05.483 }, 00:19:05.483 "memory_domains": [ 00:19:05.483 { 00:19:05.483 "dma_device_id": "system", 00:19:05.483 "dma_device_type": 1 00:19:05.483 }, 00:19:05.483 { 00:19:05.483 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:05.483 "dma_device_type": 2 00:19:05.483 } 00:19:05.483 ], 00:19:05.483 "driver_specific": { 00:19:05.483 "passthru": { 00:19:05.483 "name": "pt2", 00:19:05.483 "base_bdev_name": "malloc2" 00:19:05.483 } 00:19:05.483 } 00:19:05.483 }' 00:19:05.483 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:05.483 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:05.742 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:19:05.742 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:05.742 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:05.742 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:19:05.742 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:05.742 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:05.742 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:19:05.742 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:05.742 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:06.000 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:19:06.000 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:19:06.000 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt3 00:19:06.000 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:19:06.258 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:19:06.258 "name": "pt3", 00:19:06.258 "aliases": [ 00:19:06.258 "00000000-0000-0000-0000-000000000003" 00:19:06.258 ], 00:19:06.258 "product_name": "passthru", 00:19:06.258 "block_size": 512, 00:19:06.258 "num_blocks": 65536, 00:19:06.258 "uuid": "00000000-0000-0000-0000-000000000003", 00:19:06.258 "assigned_rate_limits": { 00:19:06.258 "rw_ios_per_sec": 0, 00:19:06.258 "rw_mbytes_per_sec": 0, 00:19:06.258 "r_mbytes_per_sec": 0, 00:19:06.258 "w_mbytes_per_sec": 0 00:19:06.258 }, 00:19:06.258 "claimed": true, 00:19:06.258 "claim_type": "exclusive_write", 00:19:06.258 "zoned": false, 00:19:06.258 "supported_io_types": { 00:19:06.258 "read": true, 00:19:06.258 "write": true, 00:19:06.258 "unmap": true, 00:19:06.258 "flush": true, 00:19:06.258 "reset": true, 00:19:06.258 "nvme_admin": false, 00:19:06.258 "nvme_io": false, 00:19:06.258 "nvme_io_md": false, 00:19:06.258 "write_zeroes": true, 00:19:06.258 "zcopy": true, 00:19:06.258 "get_zone_info": false, 00:19:06.258 "zone_management": false, 00:19:06.258 "zone_append": false, 00:19:06.258 "compare": false, 00:19:06.258 "compare_and_write": false, 00:19:06.258 "abort": true, 00:19:06.258 "seek_hole": false, 00:19:06.258 "seek_data": false, 00:19:06.258 "copy": true, 00:19:06.258 "nvme_iov_md": false 00:19:06.258 }, 00:19:06.258 "memory_domains": [ 00:19:06.258 { 00:19:06.258 "dma_device_id": "system", 00:19:06.258 "dma_device_type": 1 00:19:06.258 }, 00:19:06.258 { 00:19:06.258 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:06.258 "dma_device_type": 2 00:19:06.258 } 00:19:06.258 ], 00:19:06.258 "driver_specific": { 00:19:06.258 "passthru": { 00:19:06.258 "name": "pt3", 00:19:06.259 "base_bdev_name": "malloc3" 00:19:06.259 } 00:19:06.259 } 00:19:06.259 }' 00:19:06.259 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:06.259 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:06.259 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:19:06.259 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:06.259 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:06.259 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:19:06.259 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:06.259 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:06.517 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:19:06.517 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:06.517 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:06.517 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:19:06.517 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:19:06.517 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt4 00:19:06.517 11:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:19:06.776 11:29:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:19:06.776 "name": "pt4", 00:19:06.776 "aliases": [ 00:19:06.776 "00000000-0000-0000-0000-000000000004" 00:19:06.776 ], 00:19:06.776 "product_name": "passthru", 00:19:06.776 "block_size": 512, 00:19:06.776 "num_blocks": 65536, 00:19:06.776 "uuid": "00000000-0000-0000-0000-000000000004", 00:19:06.776 "assigned_rate_limits": { 00:19:06.776 "rw_ios_per_sec": 0, 00:19:06.776 "rw_mbytes_per_sec": 0, 00:19:06.776 "r_mbytes_per_sec": 0, 00:19:06.776 "w_mbytes_per_sec": 0 00:19:06.776 }, 00:19:06.776 "claimed": true, 00:19:06.776 "claim_type": "exclusive_write", 00:19:06.776 "zoned": false, 00:19:06.776 "supported_io_types": { 00:19:06.776 "read": true, 00:19:06.776 "write": true, 00:19:06.776 "unmap": true, 00:19:06.776 "flush": true, 00:19:06.776 "reset": true, 00:19:06.776 "nvme_admin": false, 00:19:06.776 "nvme_io": false, 00:19:06.776 "nvme_io_md": false, 00:19:06.776 "write_zeroes": true, 00:19:06.776 "zcopy": true, 00:19:06.776 "get_zone_info": false, 00:19:06.776 "zone_management": false, 00:19:06.776 "zone_append": false, 00:19:06.776 "compare": false, 00:19:06.776 "compare_and_write": false, 00:19:06.776 "abort": true, 00:19:06.776 "seek_hole": false, 00:19:06.776 "seek_data": false, 00:19:06.776 "copy": true, 00:19:06.776 "nvme_iov_md": false 00:19:06.776 }, 00:19:06.776 "memory_domains": [ 00:19:06.776 { 00:19:06.776 "dma_device_id": "system", 00:19:06.776 "dma_device_type": 1 00:19:06.776 }, 00:19:06.776 { 00:19:06.776 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:06.776 "dma_device_type": 2 00:19:06.776 } 00:19:06.776 ], 00:19:06.776 "driver_specific": { 00:19:06.776 "passthru": { 00:19:06.776 "name": "pt4", 00:19:06.776 "base_bdev_name": "malloc4" 00:19:06.776 } 00:19:06.776 } 00:19:06.776 }' 00:19:06.776 11:29:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:06.776 11:29:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:06.776 11:29:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:19:06.776 11:29:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:06.776 11:29:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:07.034 11:29:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:19:07.034 11:29:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:07.034 11:29:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:07.034 11:29:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:19:07.034 11:29:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:07.034 11:29:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:07.034 11:29:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:19:07.034 11:29:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:19:07.034 11:29:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # jq -r '.[] | .uuid' 00:19:07.292 [2024-07-15 11:29:50.795017] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:19:07.292 11:29:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # '[' 5425fa73-79f0-4071-9c37-208b9903b7eb '!=' 5425fa73-79f0-4071-9c37-208b9903b7eb ']' 00:19:07.292 11:29:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@490 -- # has_redundancy raid0 00:19:07.292 11:29:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:19:07.292 11:29:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@215 -- # return 1 00:19:07.292 11:29:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@562 -- # killprocess 932612 00:19:07.292 11:29:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@948 -- # '[' -z 932612 ']' 00:19:07.292 11:29:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@952 -- # kill -0 932612 00:19:07.293 11:29:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@953 -- # uname 00:19:07.293 11:29:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:19:07.293 11:29:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 932612 00:19:07.293 11:29:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:19:07.293 11:29:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:19:07.293 11:29:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 932612' 00:19:07.293 killing process with pid 932612 00:19:07.293 11:29:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@967 -- # kill 932612 00:19:07.293 [2024-07-15 11:29:50.868700] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:19:07.293 [2024-07-15 11:29:50.868761] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:19:07.293 [2024-07-15 11:29:50.868822] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:19:07.293 [2024-07-15 11:29:50.868835] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x173a780 name raid_bdev1, state offline 00:19:07.293 11:29:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # wait 932612 00:19:07.552 [2024-07-15 11:29:50.907781] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:19:07.552 11:29:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@564 -- # return 0 00:19:07.552 00:19:07.552 real 0m15.769s 00:19:07.552 user 0m28.366s 00:19:07.552 sys 0m2.924s 00:19:07.552 11:29:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:19:07.552 11:29:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:07.552 ************************************ 00:19:07.552 END TEST raid_superblock_test 00:19:07.552 ************************************ 00:19:07.811 11:29:51 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:19:07.811 11:29:51 bdev_raid -- bdev/bdev_raid.sh@870 -- # run_test raid_read_error_test raid_io_error_test raid0 4 read 00:19:07.811 11:29:51 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:19:07.811 11:29:51 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:19:07.811 11:29:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:19:07.811 ************************************ 00:19:07.811 START TEST raid_read_error_test 00:19:07.811 ************************************ 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1123 -- # raid_io_error_test raid0 4 read 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@788 -- # local raid_level=raid0 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@789 -- # local num_base_bdevs=4 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local error_io_type=read 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i = 1 )) 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev1 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev2 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev3 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev4 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local base_bdevs 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local raid_bdev_name=raid_bdev1 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local strip_size 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local create_arg 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local bdevperf_log 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local fail_per_s 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # '[' raid0 '!=' raid1 ']' 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@799 -- # strip_size=64 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # create_arg+=' -z 64' 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@805 -- # mktemp -p /raidtest 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@805 -- # bdevperf_log=/raidtest/tmp.yHCP036M5S 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@808 -- # raid_pid=934993 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # waitforlisten 934993 /var/tmp/spdk-raid.sock 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@829 -- # '[' -z 934993 ']' 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:19:07.811 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:19:07.811 11:29:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:19:07.811 [2024-07-15 11:29:51.289324] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:19:07.811 [2024-07-15 11:29:51.289383] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid934993 ] 00:19:07.811 [2024-07-15 11:29:51.403766] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:08.070 [2024-07-15 11:29:51.511117] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:19:08.070 [2024-07-15 11:29:51.572795] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:08.070 [2024-07-15 11:29:51.572822] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:08.637 11:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:19:08.637 11:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@862 -- # return 0 00:19:08.637 11:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:19:08.637 11:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:19:08.896 BaseBdev1_malloc 00:19:08.896 11:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev1_malloc 00:19:09.155 true 00:19:09.155 11:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:19:09.413 [2024-07-15 11:29:52.874379] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:19:09.413 [2024-07-15 11:29:52.874425] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:19:09.413 [2024-07-15 11:29:52.874444] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x27320d0 00:19:09.413 [2024-07-15 11:29:52.874456] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:19:09.413 [2024-07-15 11:29:52.876149] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:19:09.413 [2024-07-15 11:29:52.876177] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:19:09.413 BaseBdev1 00:19:09.413 11:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:19:09.413 11:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:19:09.674 BaseBdev2_malloc 00:19:09.674 11:29:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev2_malloc 00:19:10.004 true 00:19:10.004 11:29:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:19:10.263 [2024-07-15 11:29:53.614128] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:19:10.263 [2024-07-15 11:29:53.614174] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:19:10.263 [2024-07-15 11:29:53.614195] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2736910 00:19:10.263 [2024-07-15 11:29:53.614208] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:19:10.263 [2024-07-15 11:29:53.615769] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:19:10.263 [2024-07-15 11:29:53.615797] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:19:10.263 BaseBdev2 00:19:10.263 11:29:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:19:10.263 11:29:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:19:10.521 BaseBdev3_malloc 00:19:10.521 11:29:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev3_malloc 00:19:10.521 true 00:19:10.779 11:29:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:19:10.779 [2024-07-15 11:29:54.352640] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:19:10.779 [2024-07-15 11:29:54.352693] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:19:10.779 [2024-07-15 11:29:54.352715] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2738bd0 00:19:10.779 [2024-07-15 11:29:54.352730] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:19:10.779 [2024-07-15 11:29:54.354358] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:19:10.779 [2024-07-15 11:29:54.354389] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:19:10.779 BaseBdev3 00:19:10.779 11:29:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:19:10.779 11:29:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:19:11.039 BaseBdev4_malloc 00:19:11.039 11:29:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev4_malloc 00:19:11.298 true 00:19:11.298 11:29:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:19:11.557 [2024-07-15 11:29:55.076389] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:19:11.557 [2024-07-15 11:29:55.076435] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:19:11.557 [2024-07-15 11:29:55.076457] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2739aa0 00:19:11.557 [2024-07-15 11:29:55.076469] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:19:11.557 [2024-07-15 11:29:55.078045] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:19:11.557 [2024-07-15 11:29:55.078074] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:19:11.557 BaseBdev4 00:19:11.557 11:29:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@819 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n raid_bdev1 -s 00:19:11.816 [2024-07-15 11:29:55.317061] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:19:11.816 [2024-07-15 11:29:55.318397] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:19:11.816 [2024-07-15 11:29:55.318466] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:19:11.816 [2024-07-15 11:29:55.318527] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:19:11.816 [2024-07-15 11:29:55.318759] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x2733c20 00:19:11.816 [2024-07-15 11:29:55.318771] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:19:11.816 [2024-07-15 11:29:55.318985] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x2588260 00:19:11.816 [2024-07-15 11:29:55.319136] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x2733c20 00:19:11.816 [2024-07-15 11:29:55.319146] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x2733c20 00:19:11.816 [2024-07-15 11:29:55.319251] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:11.817 11:29:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@820 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:19:11.817 11:29:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:19:11.817 11:29:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:19:11.817 11:29:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:19:11.817 11:29:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:11.817 11:29:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:11.817 11:29:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:11.817 11:29:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:11.817 11:29:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:11.817 11:29:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:11.817 11:29:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:11.817 11:29:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:19:12.076 11:29:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:12.076 "name": "raid_bdev1", 00:19:12.076 "uuid": "9b1f6fff-da1a-4999-bb4c-e221d3dba772", 00:19:12.076 "strip_size_kb": 64, 00:19:12.076 "state": "online", 00:19:12.076 "raid_level": "raid0", 00:19:12.076 "superblock": true, 00:19:12.076 "num_base_bdevs": 4, 00:19:12.076 "num_base_bdevs_discovered": 4, 00:19:12.076 "num_base_bdevs_operational": 4, 00:19:12.076 "base_bdevs_list": [ 00:19:12.076 { 00:19:12.076 "name": "BaseBdev1", 00:19:12.076 "uuid": "525b0328-3115-5d12-9441-c5d454d8516b", 00:19:12.076 "is_configured": true, 00:19:12.076 "data_offset": 2048, 00:19:12.076 "data_size": 63488 00:19:12.076 }, 00:19:12.076 { 00:19:12.076 "name": "BaseBdev2", 00:19:12.076 "uuid": "08b868d2-d9ad-5f73-b81d-9945e01a10d2", 00:19:12.076 "is_configured": true, 00:19:12.076 "data_offset": 2048, 00:19:12.076 "data_size": 63488 00:19:12.076 }, 00:19:12.076 { 00:19:12.076 "name": "BaseBdev3", 00:19:12.076 "uuid": "1d3e74f2-e757-54f2-b871-3be9f472940a", 00:19:12.076 "is_configured": true, 00:19:12.076 "data_offset": 2048, 00:19:12.076 "data_size": 63488 00:19:12.076 }, 00:19:12.076 { 00:19:12.076 "name": "BaseBdev4", 00:19:12.076 "uuid": "a1344a66-5ef7-5eb6-8a95-fa933228b641", 00:19:12.076 "is_configured": true, 00:19:12.076 "data_offset": 2048, 00:19:12.076 "data_size": 63488 00:19:12.076 } 00:19:12.076 ] 00:19:12.076 }' 00:19:12.076 11:29:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:12.076 11:29:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:19:12.643 11:29:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@824 -- # sleep 1 00:19:12.643 11:29:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@823 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:19:12.902 [2024-07-15 11:29:56.299942] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x2725fc0 00:19:13.839 11:29:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@827 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:19:14.098 11:29:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # local expected_num_base_bdevs 00:19:14.098 11:29:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@830 -- # [[ raid0 = \r\a\i\d\1 ]] 00:19:14.098 11:29:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=4 00:19:14.098 11:29:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:19:14.098 11:29:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:19:14.098 11:29:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:19:14.098 11:29:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:19:14.098 11:29:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:14.098 11:29:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:14.098 11:29:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:14.098 11:29:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:14.098 11:29:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:14.098 11:29:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:14.098 11:29:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:14.098 11:29:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:19:14.098 11:29:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:14.098 "name": "raid_bdev1", 00:19:14.098 "uuid": "9b1f6fff-da1a-4999-bb4c-e221d3dba772", 00:19:14.098 "strip_size_kb": 64, 00:19:14.098 "state": "online", 00:19:14.098 "raid_level": "raid0", 00:19:14.098 "superblock": true, 00:19:14.098 "num_base_bdevs": 4, 00:19:14.098 "num_base_bdevs_discovered": 4, 00:19:14.098 "num_base_bdevs_operational": 4, 00:19:14.098 "base_bdevs_list": [ 00:19:14.098 { 00:19:14.098 "name": "BaseBdev1", 00:19:14.098 "uuid": "525b0328-3115-5d12-9441-c5d454d8516b", 00:19:14.098 "is_configured": true, 00:19:14.098 "data_offset": 2048, 00:19:14.098 "data_size": 63488 00:19:14.098 }, 00:19:14.098 { 00:19:14.098 "name": "BaseBdev2", 00:19:14.098 "uuid": "08b868d2-d9ad-5f73-b81d-9945e01a10d2", 00:19:14.098 "is_configured": true, 00:19:14.098 "data_offset": 2048, 00:19:14.098 "data_size": 63488 00:19:14.098 }, 00:19:14.098 { 00:19:14.098 "name": "BaseBdev3", 00:19:14.098 "uuid": "1d3e74f2-e757-54f2-b871-3be9f472940a", 00:19:14.098 "is_configured": true, 00:19:14.098 "data_offset": 2048, 00:19:14.098 "data_size": 63488 00:19:14.098 }, 00:19:14.098 { 00:19:14.098 "name": "BaseBdev4", 00:19:14.098 "uuid": "a1344a66-5ef7-5eb6-8a95-fa933228b641", 00:19:14.098 "is_configured": true, 00:19:14.098 "data_offset": 2048, 00:19:14.098 "data_size": 63488 00:19:14.098 } 00:19:14.098 ] 00:19:14.098 }' 00:19:14.098 11:29:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:14.098 11:29:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:19:14.666 11:29:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:19:14.926 [2024-07-15 11:29:58.473476] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:19:14.926 [2024-07-15 11:29:58.473516] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:19:14.926 [2024-07-15 11:29:58.476696] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:19:14.926 [2024-07-15 11:29:58.476736] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:14.926 [2024-07-15 11:29:58.476777] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:19:14.926 [2024-07-15 11:29:58.476788] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x2733c20 name raid_bdev1, state offline 00:19:14.926 0 00:19:14.926 11:29:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # killprocess 934993 00:19:14.926 11:29:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@948 -- # '[' -z 934993 ']' 00:19:14.926 11:29:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@952 -- # kill -0 934993 00:19:14.926 11:29:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@953 -- # uname 00:19:14.926 11:29:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:19:14.926 11:29:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 934993 00:19:15.185 11:29:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:19:15.185 11:29:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:19:15.185 11:29:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 934993' 00:19:15.185 killing process with pid 934993 00:19:15.185 11:29:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@967 -- # kill 934993 00:19:15.185 [2024-07-15 11:29:58.541114] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:19:15.185 11:29:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # wait 934993 00:19:15.185 [2024-07-15 11:29:58.574157] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:19:15.445 11:29:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # grep -v Job /raidtest/tmp.yHCP036M5S 00:19:15.445 11:29:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # grep raid_bdev1 00:19:15.445 11:29:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # awk '{print $6}' 00:19:15.445 11:29:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # fail_per_s=0.46 00:19:15.445 11:29:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@844 -- # has_redundancy raid0 00:19:15.445 11:29:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:19:15.445 11:29:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@215 -- # return 1 00:19:15.445 11:29:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.46 != \0\.\0\0 ]] 00:19:15.445 00:19:15.445 real 0m7.602s 00:19:15.445 user 0m12.131s 00:19:15.445 sys 0m1.349s 00:19:15.445 11:29:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:19:15.445 11:29:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:19:15.445 ************************************ 00:19:15.445 END TEST raid_read_error_test 00:19:15.445 ************************************ 00:19:15.445 11:29:58 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:19:15.445 11:29:58 bdev_raid -- bdev/bdev_raid.sh@871 -- # run_test raid_write_error_test raid_io_error_test raid0 4 write 00:19:15.445 11:29:58 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:19:15.445 11:29:58 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:19:15.445 11:29:58 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:19:15.445 ************************************ 00:19:15.445 START TEST raid_write_error_test 00:19:15.445 ************************************ 00:19:15.445 11:29:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1123 -- # raid_io_error_test raid0 4 write 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@788 -- # local raid_level=raid0 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@789 -- # local num_base_bdevs=4 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local error_io_type=write 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i = 1 )) 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev1 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev2 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev3 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev4 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local base_bdevs 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local raid_bdev_name=raid_bdev1 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local strip_size 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local create_arg 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local bdevperf_log 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local fail_per_s 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # '[' raid0 '!=' raid1 ']' 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@799 -- # strip_size=64 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # create_arg+=' -z 64' 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@805 -- # mktemp -p /raidtest 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@805 -- # bdevperf_log=/raidtest/tmp.uSRKkSwXL0 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@808 -- # raid_pid=936146 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # waitforlisten 936146 /var/tmp/spdk-raid.sock 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@829 -- # '[' -z 936146 ']' 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:19:15.446 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:19:15.446 11:29:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:19:15.446 [2024-07-15 11:29:58.981472] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:19:15.446 [2024-07-15 11:29:58.981540] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid936146 ] 00:19:15.705 [2024-07-15 11:29:59.110668] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:15.705 [2024-07-15 11:29:59.213917] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:19:15.705 [2024-07-15 11:29:59.273929] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:15.705 [2024-07-15 11:29:59.273960] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:16.640 11:29:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:19:16.640 11:29:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@862 -- # return 0 00:19:16.640 11:29:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:19:16.640 11:29:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:19:16.640 BaseBdev1_malloc 00:19:16.640 11:30:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev1_malloc 00:19:16.898 true 00:19:16.898 11:30:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:19:17.157 [2024-07-15 11:30:00.635619] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:19:17.157 [2024-07-15 11:30:00.635667] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:19:17.157 [2024-07-15 11:30:00.635690] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x13d00d0 00:19:17.157 [2024-07-15 11:30:00.635704] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:19:17.157 [2024-07-15 11:30:00.637538] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:19:17.157 [2024-07-15 11:30:00.637568] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:19:17.157 BaseBdev1 00:19:17.157 11:30:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:19:17.157 11:30:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:19:17.415 BaseBdev2_malloc 00:19:17.415 11:30:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev2_malloc 00:19:17.673 true 00:19:17.673 11:30:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:19:17.931 [2024-07-15 11:30:01.374343] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:19:17.931 [2024-07-15 11:30:01.374392] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:19:17.931 [2024-07-15 11:30:01.374414] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x13d4910 00:19:17.931 [2024-07-15 11:30:01.374428] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:19:17.931 [2024-07-15 11:30:01.375990] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:19:17.931 [2024-07-15 11:30:01.376018] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:19:17.931 BaseBdev2 00:19:17.931 11:30:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:19:17.931 11:30:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:19:18.188 BaseBdev3_malloc 00:19:18.188 11:30:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev3_malloc 00:19:18.446 true 00:19:18.447 11:30:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:19:18.705 [2024-07-15 11:30:02.121030] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:19:18.705 [2024-07-15 11:30:02.121079] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:19:18.705 [2024-07-15 11:30:02.121101] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x13d6bd0 00:19:18.705 [2024-07-15 11:30:02.121114] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:19:18.705 [2024-07-15 11:30:02.122716] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:19:18.705 [2024-07-15 11:30:02.122743] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:19:18.705 BaseBdev3 00:19:18.705 11:30:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:19:18.705 11:30:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:19:18.964 BaseBdev4_malloc 00:19:18.964 11:30:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev4_malloc 00:19:19.224 true 00:19:19.224 11:30:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:19:19.483 [2024-07-15 11:30:02.851636] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:19:19.483 [2024-07-15 11:30:02.851681] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:19:19.483 [2024-07-15 11:30:02.851703] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x13d7aa0 00:19:19.483 [2024-07-15 11:30:02.851716] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:19:19.483 [2024-07-15 11:30:02.853176] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:19:19.483 [2024-07-15 11:30:02.853204] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:19:19.483 BaseBdev4 00:19:19.483 11:30:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@819 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r raid0 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n raid_bdev1 -s 00:19:19.741 [2024-07-15 11:30:03.096319] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:19:19.741 [2024-07-15 11:30:03.097512] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:19:19.741 [2024-07-15 11:30:03.097578] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:19:19.741 [2024-07-15 11:30:03.097640] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:19:19.741 [2024-07-15 11:30:03.097864] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x13d1c20 00:19:19.741 [2024-07-15 11:30:03.097876] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:19:19.741 [2024-07-15 11:30:03.098071] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1226260 00:19:19.741 [2024-07-15 11:30:03.098216] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x13d1c20 00:19:19.741 [2024-07-15 11:30:03.098225] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x13d1c20 00:19:19.741 [2024-07-15 11:30:03.098322] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:19.741 11:30:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@820 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:19:19.741 11:30:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:19:19.741 11:30:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:19:19.741 11:30:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:19:19.741 11:30:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:19.741 11:30:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:19.741 11:30:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:19.741 11:30:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:19.741 11:30:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:19.741 11:30:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:19.741 11:30:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:19.741 11:30:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:19:20.000 11:30:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:20.000 "name": "raid_bdev1", 00:19:20.000 "uuid": "4f32bee8-54a3-40f8-ad98-0431cc133dce", 00:19:20.000 "strip_size_kb": 64, 00:19:20.000 "state": "online", 00:19:20.000 "raid_level": "raid0", 00:19:20.000 "superblock": true, 00:19:20.000 "num_base_bdevs": 4, 00:19:20.000 "num_base_bdevs_discovered": 4, 00:19:20.000 "num_base_bdevs_operational": 4, 00:19:20.000 "base_bdevs_list": [ 00:19:20.000 { 00:19:20.000 "name": "BaseBdev1", 00:19:20.000 "uuid": "49232a17-dce3-5ad7-8b69-26e53740368f", 00:19:20.000 "is_configured": true, 00:19:20.000 "data_offset": 2048, 00:19:20.000 "data_size": 63488 00:19:20.000 }, 00:19:20.000 { 00:19:20.000 "name": "BaseBdev2", 00:19:20.000 "uuid": "52b1b9f4-3d1e-5f6a-8f35-c376e4bf3363", 00:19:20.000 "is_configured": true, 00:19:20.000 "data_offset": 2048, 00:19:20.000 "data_size": 63488 00:19:20.000 }, 00:19:20.000 { 00:19:20.000 "name": "BaseBdev3", 00:19:20.000 "uuid": "a59d48bd-0242-5efc-aac6-e7d4cc7a4865", 00:19:20.000 "is_configured": true, 00:19:20.000 "data_offset": 2048, 00:19:20.000 "data_size": 63488 00:19:20.000 }, 00:19:20.000 { 00:19:20.000 "name": "BaseBdev4", 00:19:20.000 "uuid": "b9a4323c-1c62-5eb6-ae48-dcc6e9b56c1e", 00:19:20.000 "is_configured": true, 00:19:20.000 "data_offset": 2048, 00:19:20.000 "data_size": 63488 00:19:20.000 } 00:19:20.000 ] 00:19:20.000 }' 00:19:20.000 11:30:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:20.000 11:30:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:19:20.569 11:30:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@824 -- # sleep 1 00:19:20.569 11:30:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@823 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:19:20.569 [2024-07-15 11:30:04.055144] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x13c3fc0 00:19:21.509 11:30:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@827 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:19:21.768 11:30:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # local expected_num_base_bdevs 00:19:21.768 11:30:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@830 -- # [[ raid0 = \r\a\i\d\1 ]] 00:19:21.768 11:30:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=4 00:19:21.768 11:30:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:19:21.768 11:30:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:19:21.768 11:30:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:19:21.768 11:30:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid0 00:19:21.768 11:30:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:21.768 11:30:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:21.768 11:30:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:21.769 11:30:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:21.769 11:30:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:21.769 11:30:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:21.769 11:30:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:19:21.769 11:30:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:22.027 11:30:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:22.027 "name": "raid_bdev1", 00:19:22.027 "uuid": "4f32bee8-54a3-40f8-ad98-0431cc133dce", 00:19:22.027 "strip_size_kb": 64, 00:19:22.027 "state": "online", 00:19:22.027 "raid_level": "raid0", 00:19:22.027 "superblock": true, 00:19:22.028 "num_base_bdevs": 4, 00:19:22.028 "num_base_bdevs_discovered": 4, 00:19:22.028 "num_base_bdevs_operational": 4, 00:19:22.028 "base_bdevs_list": [ 00:19:22.028 { 00:19:22.028 "name": "BaseBdev1", 00:19:22.028 "uuid": "49232a17-dce3-5ad7-8b69-26e53740368f", 00:19:22.028 "is_configured": true, 00:19:22.028 "data_offset": 2048, 00:19:22.028 "data_size": 63488 00:19:22.028 }, 00:19:22.028 { 00:19:22.028 "name": "BaseBdev2", 00:19:22.028 "uuid": "52b1b9f4-3d1e-5f6a-8f35-c376e4bf3363", 00:19:22.028 "is_configured": true, 00:19:22.028 "data_offset": 2048, 00:19:22.028 "data_size": 63488 00:19:22.028 }, 00:19:22.028 { 00:19:22.028 "name": "BaseBdev3", 00:19:22.028 "uuid": "a59d48bd-0242-5efc-aac6-e7d4cc7a4865", 00:19:22.028 "is_configured": true, 00:19:22.028 "data_offset": 2048, 00:19:22.028 "data_size": 63488 00:19:22.028 }, 00:19:22.028 { 00:19:22.028 "name": "BaseBdev4", 00:19:22.028 "uuid": "b9a4323c-1c62-5eb6-ae48-dcc6e9b56c1e", 00:19:22.028 "is_configured": true, 00:19:22.028 "data_offset": 2048, 00:19:22.028 "data_size": 63488 00:19:22.028 } 00:19:22.028 ] 00:19:22.028 }' 00:19:22.028 11:30:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:22.028 11:30:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:19:22.596 11:30:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:19:22.856 [2024-07-15 11:30:06.308967] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:19:22.856 [2024-07-15 11:30:06.308999] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:19:22.856 [2024-07-15 11:30:06.312170] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:19:22.856 [2024-07-15 11:30:06.312211] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:22.857 [2024-07-15 11:30:06.312251] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:19:22.857 [2024-07-15 11:30:06.312263] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x13d1c20 name raid_bdev1, state offline 00:19:22.857 0 00:19:22.857 11:30:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # killprocess 936146 00:19:22.857 11:30:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@948 -- # '[' -z 936146 ']' 00:19:22.857 11:30:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@952 -- # kill -0 936146 00:19:22.857 11:30:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@953 -- # uname 00:19:22.857 11:30:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:19:22.857 11:30:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 936146 00:19:22.857 11:30:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:19:22.857 11:30:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:19:22.857 11:30:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 936146' 00:19:22.857 killing process with pid 936146 00:19:22.857 11:30:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@967 -- # kill 936146 00:19:22.857 [2024-07-15 11:30:06.377996] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:19:22.857 11:30:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # wait 936146 00:19:22.857 [2024-07-15 11:30:06.408433] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:19:23.119 11:30:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # grep -v Job /raidtest/tmp.uSRKkSwXL0 00:19:23.119 11:30:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # grep raid_bdev1 00:19:23.119 11:30:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # awk '{print $6}' 00:19:23.119 11:30:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # fail_per_s=0.45 00:19:23.119 11:30:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@844 -- # has_redundancy raid0 00:19:23.119 11:30:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:19:23.119 11:30:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@215 -- # return 1 00:19:23.119 11:30:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.45 != \0\.\0\0 ]] 00:19:23.119 00:19:23.119 real 0m7.731s 00:19:23.119 user 0m12.355s 00:19:23.119 sys 0m1.360s 00:19:23.119 11:30:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:19:23.119 11:30:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:19:23.119 ************************************ 00:19:23.119 END TEST raid_write_error_test 00:19:23.119 ************************************ 00:19:23.119 11:30:06 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:19:23.119 11:30:06 bdev_raid -- bdev/bdev_raid.sh@866 -- # for level in raid0 concat raid1 00:19:23.119 11:30:06 bdev_raid -- bdev/bdev_raid.sh@867 -- # run_test raid_state_function_test raid_state_function_test concat 4 false 00:19:23.119 11:30:06 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:19:23.119 11:30:06 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:19:23.119 11:30:06 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:19:23.379 ************************************ 00:19:23.379 START TEST raid_state_function_test 00:19:23.379 ************************************ 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1123 -- # raid_state_function_test concat 4 false 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@220 -- # local raid_level=concat 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=4 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # local superblock=false 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev3 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev4 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # local strip_size 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # '[' concat '!=' raid1 ']' 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # strip_size=64 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@232 -- # strip_size_create_arg='-z 64' 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # '[' false = true ']' 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@240 -- # superblock_create_arg= 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # raid_pid=937799 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 937799' 00:19:23.379 Process raid pid: 937799 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@246 -- # waitforlisten 937799 /var/tmp/spdk-raid.sock 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@829 -- # '[' -z 937799 ']' 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:19:23.379 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:19:23.379 11:30:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:23.379 [2024-07-15 11:30:06.791131] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:19:23.379 [2024-07-15 11:30:06.791184] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:23.379 [2024-07-15 11:30:06.906638] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:23.639 [2024-07-15 11:30:07.009044] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:19:23.639 [2024-07-15 11:30:07.072759] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:23.639 [2024-07-15 11:30:07.072794] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:24.206 11:30:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:19:24.206 11:30:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@862 -- # return 0 00:19:24.206 11:30:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:19:24.543 [2024-07-15 11:30:07.959220] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:19:24.543 [2024-07-15 11:30:07.959263] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:19:24.543 [2024-07-15 11:30:07.959274] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:19:24.543 [2024-07-15 11:30:07.959286] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:19:24.543 [2024-07-15 11:30:07.959295] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:19:24.543 [2024-07-15 11:30:07.959306] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:19:24.543 [2024-07-15 11:30:07.959314] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:19:24.543 [2024-07-15 11:30:07.959326] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:19:24.543 11:30:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:19:24.543 11:30:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:19:24.543 11:30:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:19:24.543 11:30:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:19:24.543 11:30:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:24.543 11:30:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:24.543 11:30:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:24.543 11:30:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:24.543 11:30:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:24.543 11:30:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:24.543 11:30:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:24.543 11:30:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:24.803 11:30:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:24.803 "name": "Existed_Raid", 00:19:24.803 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:24.803 "strip_size_kb": 64, 00:19:24.803 "state": "configuring", 00:19:24.803 "raid_level": "concat", 00:19:24.803 "superblock": false, 00:19:24.803 "num_base_bdevs": 4, 00:19:24.803 "num_base_bdevs_discovered": 0, 00:19:24.803 "num_base_bdevs_operational": 4, 00:19:24.803 "base_bdevs_list": [ 00:19:24.803 { 00:19:24.803 "name": "BaseBdev1", 00:19:24.803 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:24.803 "is_configured": false, 00:19:24.803 "data_offset": 0, 00:19:24.803 "data_size": 0 00:19:24.803 }, 00:19:24.803 { 00:19:24.803 "name": "BaseBdev2", 00:19:24.803 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:24.803 "is_configured": false, 00:19:24.803 "data_offset": 0, 00:19:24.803 "data_size": 0 00:19:24.803 }, 00:19:24.803 { 00:19:24.803 "name": "BaseBdev3", 00:19:24.803 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:24.803 "is_configured": false, 00:19:24.803 "data_offset": 0, 00:19:24.803 "data_size": 0 00:19:24.803 }, 00:19:24.803 { 00:19:24.803 "name": "BaseBdev4", 00:19:24.803 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:24.803 "is_configured": false, 00:19:24.803 "data_offset": 0, 00:19:24.803 "data_size": 0 00:19:24.803 } 00:19:24.803 ] 00:19:24.803 }' 00:19:24.803 11:30:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:24.803 11:30:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:25.371 11:30:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:19:25.629 [2024-07-15 11:30:09.041948] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:19:25.629 [2024-07-15 11:30:09.041977] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1329aa0 name Existed_Raid, state configuring 00:19:25.629 11:30:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:19:25.888 [2024-07-15 11:30:09.290627] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:19:25.888 [2024-07-15 11:30:09.290656] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:19:25.888 [2024-07-15 11:30:09.290666] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:19:25.888 [2024-07-15 11:30:09.290677] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:19:25.888 [2024-07-15 11:30:09.290686] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:19:25.888 [2024-07-15 11:30:09.290697] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:19:25.888 [2024-07-15 11:30:09.290706] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:19:25.888 [2024-07-15 11:30:09.290716] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:19:25.888 11:30:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:19:25.888 [2024-07-15 11:30:09.477023] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:19:25.888 BaseBdev1 00:19:26.147 11:30:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:19:26.147 11:30:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:19:26.147 11:30:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:19:26.147 11:30:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:19:26.147 11:30:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:19:26.147 11:30:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:19:26.147 11:30:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:19:26.147 11:30:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:19:26.406 [ 00:19:26.406 { 00:19:26.406 "name": "BaseBdev1", 00:19:26.406 "aliases": [ 00:19:26.406 "f2ddc0e6-9dc3-410c-8c78-19594159a130" 00:19:26.406 ], 00:19:26.406 "product_name": "Malloc disk", 00:19:26.406 "block_size": 512, 00:19:26.406 "num_blocks": 65536, 00:19:26.406 "uuid": "f2ddc0e6-9dc3-410c-8c78-19594159a130", 00:19:26.406 "assigned_rate_limits": { 00:19:26.406 "rw_ios_per_sec": 0, 00:19:26.406 "rw_mbytes_per_sec": 0, 00:19:26.406 "r_mbytes_per_sec": 0, 00:19:26.406 "w_mbytes_per_sec": 0 00:19:26.406 }, 00:19:26.406 "claimed": true, 00:19:26.406 "claim_type": "exclusive_write", 00:19:26.406 "zoned": false, 00:19:26.406 "supported_io_types": { 00:19:26.406 "read": true, 00:19:26.406 "write": true, 00:19:26.406 "unmap": true, 00:19:26.406 "flush": true, 00:19:26.406 "reset": true, 00:19:26.406 "nvme_admin": false, 00:19:26.406 "nvme_io": false, 00:19:26.406 "nvme_io_md": false, 00:19:26.406 "write_zeroes": true, 00:19:26.406 "zcopy": true, 00:19:26.406 "get_zone_info": false, 00:19:26.406 "zone_management": false, 00:19:26.406 "zone_append": false, 00:19:26.406 "compare": false, 00:19:26.406 "compare_and_write": false, 00:19:26.406 "abort": true, 00:19:26.406 "seek_hole": false, 00:19:26.406 "seek_data": false, 00:19:26.406 "copy": true, 00:19:26.406 "nvme_iov_md": false 00:19:26.406 }, 00:19:26.406 "memory_domains": [ 00:19:26.406 { 00:19:26.406 "dma_device_id": "system", 00:19:26.406 "dma_device_type": 1 00:19:26.406 }, 00:19:26.406 { 00:19:26.406 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:26.406 "dma_device_type": 2 00:19:26.406 } 00:19:26.406 ], 00:19:26.406 "driver_specific": {} 00:19:26.406 } 00:19:26.406 ] 00:19:26.406 11:30:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:19:26.406 11:30:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:19:26.406 11:30:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:19:26.406 11:30:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:19:26.406 11:30:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:19:26.406 11:30:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:26.406 11:30:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:26.406 11:30:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:26.406 11:30:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:26.406 11:30:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:26.406 11:30:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:26.406 11:30:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:26.406 11:30:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:26.665 11:30:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:26.665 "name": "Existed_Raid", 00:19:26.665 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:26.665 "strip_size_kb": 64, 00:19:26.665 "state": "configuring", 00:19:26.665 "raid_level": "concat", 00:19:26.665 "superblock": false, 00:19:26.665 "num_base_bdevs": 4, 00:19:26.665 "num_base_bdevs_discovered": 1, 00:19:26.665 "num_base_bdevs_operational": 4, 00:19:26.665 "base_bdevs_list": [ 00:19:26.665 { 00:19:26.665 "name": "BaseBdev1", 00:19:26.665 "uuid": "f2ddc0e6-9dc3-410c-8c78-19594159a130", 00:19:26.665 "is_configured": true, 00:19:26.665 "data_offset": 0, 00:19:26.665 "data_size": 65536 00:19:26.665 }, 00:19:26.665 { 00:19:26.665 "name": "BaseBdev2", 00:19:26.665 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:26.665 "is_configured": false, 00:19:26.665 "data_offset": 0, 00:19:26.665 "data_size": 0 00:19:26.665 }, 00:19:26.665 { 00:19:26.665 "name": "BaseBdev3", 00:19:26.665 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:26.665 "is_configured": false, 00:19:26.665 "data_offset": 0, 00:19:26.665 "data_size": 0 00:19:26.665 }, 00:19:26.665 { 00:19:26.665 "name": "BaseBdev4", 00:19:26.665 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:26.665 "is_configured": false, 00:19:26.665 "data_offset": 0, 00:19:26.665 "data_size": 0 00:19:26.665 } 00:19:26.665 ] 00:19:26.665 }' 00:19:26.665 11:30:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:26.665 11:30:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:27.233 11:30:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:19:27.492 [2024-07-15 11:30:10.904825] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:19:27.492 [2024-07-15 11:30:10.904866] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1329310 name Existed_Raid, state configuring 00:19:27.492 11:30:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:19:27.492 [2024-07-15 11:30:11.069302] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:19:27.492 [2024-07-15 11:30:11.070733] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:19:27.492 [2024-07-15 11:30:11.070767] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:19:27.492 [2024-07-15 11:30:11.070777] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:19:27.492 [2024-07-15 11:30:11.070789] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:19:27.492 [2024-07-15 11:30:11.070798] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:19:27.492 [2024-07-15 11:30:11.070809] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:19:27.751 11:30:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:19:27.751 11:30:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:19:27.751 11:30:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:19:27.751 11:30:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:19:27.751 11:30:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:19:27.751 11:30:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:19:27.751 11:30:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:27.751 11:30:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:27.751 11:30:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:27.751 11:30:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:27.751 11:30:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:27.751 11:30:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:27.751 11:30:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:27.751 11:30:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:27.751 11:30:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:27.751 "name": "Existed_Raid", 00:19:27.751 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:27.752 "strip_size_kb": 64, 00:19:27.752 "state": "configuring", 00:19:27.752 "raid_level": "concat", 00:19:27.752 "superblock": false, 00:19:27.752 "num_base_bdevs": 4, 00:19:27.752 "num_base_bdevs_discovered": 1, 00:19:27.752 "num_base_bdevs_operational": 4, 00:19:27.752 "base_bdevs_list": [ 00:19:27.752 { 00:19:27.752 "name": "BaseBdev1", 00:19:27.752 "uuid": "f2ddc0e6-9dc3-410c-8c78-19594159a130", 00:19:27.752 "is_configured": true, 00:19:27.752 "data_offset": 0, 00:19:27.752 "data_size": 65536 00:19:27.752 }, 00:19:27.752 { 00:19:27.752 "name": "BaseBdev2", 00:19:27.752 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:27.752 "is_configured": false, 00:19:27.752 "data_offset": 0, 00:19:27.752 "data_size": 0 00:19:27.752 }, 00:19:27.752 { 00:19:27.752 "name": "BaseBdev3", 00:19:27.752 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:27.752 "is_configured": false, 00:19:27.752 "data_offset": 0, 00:19:27.752 "data_size": 0 00:19:27.752 }, 00:19:27.752 { 00:19:27.752 "name": "BaseBdev4", 00:19:27.752 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:27.752 "is_configured": false, 00:19:27.752 "data_offset": 0, 00:19:27.752 "data_size": 0 00:19:27.752 } 00:19:27.752 ] 00:19:27.752 }' 00:19:27.752 11:30:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:27.752 11:30:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:28.687 11:30:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:19:28.687 [2024-07-15 11:30:12.188887] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:19:28.687 BaseBdev2 00:19:28.687 11:30:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:19:28.687 11:30:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:19:28.687 11:30:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:19:28.687 11:30:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:19:28.687 11:30:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:19:28.687 11:30:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:19:28.687 11:30:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:19:28.945 11:30:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:19:29.204 [ 00:19:29.204 { 00:19:29.204 "name": "BaseBdev2", 00:19:29.204 "aliases": [ 00:19:29.204 "d23e25d6-16ba-4938-9ad4-3afddb2df86d" 00:19:29.204 ], 00:19:29.204 "product_name": "Malloc disk", 00:19:29.204 "block_size": 512, 00:19:29.204 "num_blocks": 65536, 00:19:29.204 "uuid": "d23e25d6-16ba-4938-9ad4-3afddb2df86d", 00:19:29.204 "assigned_rate_limits": { 00:19:29.204 "rw_ios_per_sec": 0, 00:19:29.204 "rw_mbytes_per_sec": 0, 00:19:29.204 "r_mbytes_per_sec": 0, 00:19:29.204 "w_mbytes_per_sec": 0 00:19:29.204 }, 00:19:29.204 "claimed": true, 00:19:29.204 "claim_type": "exclusive_write", 00:19:29.204 "zoned": false, 00:19:29.204 "supported_io_types": { 00:19:29.204 "read": true, 00:19:29.204 "write": true, 00:19:29.204 "unmap": true, 00:19:29.204 "flush": true, 00:19:29.204 "reset": true, 00:19:29.204 "nvme_admin": false, 00:19:29.204 "nvme_io": false, 00:19:29.204 "nvme_io_md": false, 00:19:29.204 "write_zeroes": true, 00:19:29.204 "zcopy": true, 00:19:29.204 "get_zone_info": false, 00:19:29.204 "zone_management": false, 00:19:29.204 "zone_append": false, 00:19:29.204 "compare": false, 00:19:29.204 "compare_and_write": false, 00:19:29.204 "abort": true, 00:19:29.204 "seek_hole": false, 00:19:29.204 "seek_data": false, 00:19:29.204 "copy": true, 00:19:29.204 "nvme_iov_md": false 00:19:29.204 }, 00:19:29.204 "memory_domains": [ 00:19:29.204 { 00:19:29.204 "dma_device_id": "system", 00:19:29.204 "dma_device_type": 1 00:19:29.204 }, 00:19:29.204 { 00:19:29.204 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:29.204 "dma_device_type": 2 00:19:29.204 } 00:19:29.204 ], 00:19:29.204 "driver_specific": {} 00:19:29.204 } 00:19:29.204 ] 00:19:29.204 11:30:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:19:29.204 11:30:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:19:29.204 11:30:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:19:29.204 11:30:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:19:29.204 11:30:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:19:29.204 11:30:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:19:29.204 11:30:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:19:29.204 11:30:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:29.204 11:30:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:29.204 11:30:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:29.204 11:30:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:29.204 11:30:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:29.204 11:30:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:29.204 11:30:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:29.204 11:30:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:29.462 11:30:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:29.462 "name": "Existed_Raid", 00:19:29.462 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:29.462 "strip_size_kb": 64, 00:19:29.462 "state": "configuring", 00:19:29.462 "raid_level": "concat", 00:19:29.462 "superblock": false, 00:19:29.462 "num_base_bdevs": 4, 00:19:29.462 "num_base_bdevs_discovered": 2, 00:19:29.462 "num_base_bdevs_operational": 4, 00:19:29.462 "base_bdevs_list": [ 00:19:29.462 { 00:19:29.462 "name": "BaseBdev1", 00:19:29.462 "uuid": "f2ddc0e6-9dc3-410c-8c78-19594159a130", 00:19:29.462 "is_configured": true, 00:19:29.462 "data_offset": 0, 00:19:29.462 "data_size": 65536 00:19:29.462 }, 00:19:29.462 { 00:19:29.462 "name": "BaseBdev2", 00:19:29.462 "uuid": "d23e25d6-16ba-4938-9ad4-3afddb2df86d", 00:19:29.462 "is_configured": true, 00:19:29.462 "data_offset": 0, 00:19:29.462 "data_size": 65536 00:19:29.462 }, 00:19:29.462 { 00:19:29.462 "name": "BaseBdev3", 00:19:29.462 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:29.462 "is_configured": false, 00:19:29.462 "data_offset": 0, 00:19:29.462 "data_size": 0 00:19:29.462 }, 00:19:29.462 { 00:19:29.462 "name": "BaseBdev4", 00:19:29.462 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:29.462 "is_configured": false, 00:19:29.462 "data_offset": 0, 00:19:29.462 "data_size": 0 00:19:29.462 } 00:19:29.462 ] 00:19:29.462 }' 00:19:29.462 11:30:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:29.462 11:30:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:30.027 11:30:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:19:30.285 [2024-07-15 11:30:13.792545] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:19:30.285 BaseBdev3 00:19:30.285 11:30:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev3 00:19:30.285 11:30:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:19:30.285 11:30:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:19:30.285 11:30:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:19:30.285 11:30:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:19:30.285 11:30:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:19:30.285 11:30:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:19:30.544 11:30:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:19:30.803 [ 00:19:30.803 { 00:19:30.803 "name": "BaseBdev3", 00:19:30.803 "aliases": [ 00:19:30.803 "60a05ca7-4e34-4264-9c0c-7f8f82d40447" 00:19:30.803 ], 00:19:30.803 "product_name": "Malloc disk", 00:19:30.803 "block_size": 512, 00:19:30.803 "num_blocks": 65536, 00:19:30.803 "uuid": "60a05ca7-4e34-4264-9c0c-7f8f82d40447", 00:19:30.803 "assigned_rate_limits": { 00:19:30.803 "rw_ios_per_sec": 0, 00:19:30.803 "rw_mbytes_per_sec": 0, 00:19:30.803 "r_mbytes_per_sec": 0, 00:19:30.803 "w_mbytes_per_sec": 0 00:19:30.803 }, 00:19:30.803 "claimed": true, 00:19:30.803 "claim_type": "exclusive_write", 00:19:30.803 "zoned": false, 00:19:30.803 "supported_io_types": { 00:19:30.803 "read": true, 00:19:30.803 "write": true, 00:19:30.803 "unmap": true, 00:19:30.803 "flush": true, 00:19:30.803 "reset": true, 00:19:30.803 "nvme_admin": false, 00:19:30.803 "nvme_io": false, 00:19:30.803 "nvme_io_md": false, 00:19:30.803 "write_zeroes": true, 00:19:30.803 "zcopy": true, 00:19:30.803 "get_zone_info": false, 00:19:30.803 "zone_management": false, 00:19:30.803 "zone_append": false, 00:19:30.803 "compare": false, 00:19:30.803 "compare_and_write": false, 00:19:30.803 "abort": true, 00:19:30.803 "seek_hole": false, 00:19:30.803 "seek_data": false, 00:19:30.803 "copy": true, 00:19:30.803 "nvme_iov_md": false 00:19:30.803 }, 00:19:30.803 "memory_domains": [ 00:19:30.803 { 00:19:30.803 "dma_device_id": "system", 00:19:30.803 "dma_device_type": 1 00:19:30.803 }, 00:19:30.803 { 00:19:30.803 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:30.803 "dma_device_type": 2 00:19:30.803 } 00:19:30.803 ], 00:19:30.803 "driver_specific": {} 00:19:30.803 } 00:19:30.803 ] 00:19:30.803 11:30:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:19:30.803 11:30:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:19:30.803 11:30:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:19:30.803 11:30:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:19:30.803 11:30:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:19:30.803 11:30:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:19:30.803 11:30:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:19:30.803 11:30:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:30.803 11:30:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:30.803 11:30:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:30.803 11:30:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:30.803 11:30:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:30.803 11:30:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:30.803 11:30:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:30.803 11:30:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:30.803 11:30:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:30.803 "name": "Existed_Raid", 00:19:30.803 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:30.803 "strip_size_kb": 64, 00:19:30.803 "state": "configuring", 00:19:30.803 "raid_level": "concat", 00:19:30.803 "superblock": false, 00:19:30.803 "num_base_bdevs": 4, 00:19:30.803 "num_base_bdevs_discovered": 3, 00:19:30.803 "num_base_bdevs_operational": 4, 00:19:30.803 "base_bdevs_list": [ 00:19:30.803 { 00:19:30.803 "name": "BaseBdev1", 00:19:30.803 "uuid": "f2ddc0e6-9dc3-410c-8c78-19594159a130", 00:19:30.803 "is_configured": true, 00:19:30.803 "data_offset": 0, 00:19:30.803 "data_size": 65536 00:19:30.803 }, 00:19:30.803 { 00:19:30.803 "name": "BaseBdev2", 00:19:30.803 "uuid": "d23e25d6-16ba-4938-9ad4-3afddb2df86d", 00:19:30.803 "is_configured": true, 00:19:30.803 "data_offset": 0, 00:19:30.803 "data_size": 65536 00:19:30.803 }, 00:19:30.803 { 00:19:30.803 "name": "BaseBdev3", 00:19:30.803 "uuid": "60a05ca7-4e34-4264-9c0c-7f8f82d40447", 00:19:30.803 "is_configured": true, 00:19:30.803 "data_offset": 0, 00:19:30.803 "data_size": 65536 00:19:30.803 }, 00:19:30.803 { 00:19:30.803 "name": "BaseBdev4", 00:19:30.803 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:30.803 "is_configured": false, 00:19:30.803 "data_offset": 0, 00:19:30.803 "data_size": 0 00:19:30.803 } 00:19:30.803 ] 00:19:30.803 }' 00:19:30.803 11:30:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:30.803 11:30:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:31.371 11:30:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4 00:19:31.939 [2024-07-15 11:30:15.440313] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:19:31.939 [2024-07-15 11:30:15.440352] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x132a350 00:19:31.939 [2024-07-15 11:30:15.440360] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:19:31.939 [2024-07-15 11:30:15.440620] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x132a020 00:19:31.939 [2024-07-15 11:30:15.440743] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x132a350 00:19:31.939 [2024-07-15 11:30:15.440753] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x132a350 00:19:31.939 [2024-07-15 11:30:15.440918] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:31.939 BaseBdev4 00:19:31.939 11:30:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev4 00:19:31.939 11:30:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev4 00:19:31.939 11:30:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:19:31.939 11:30:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:19:31.939 11:30:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:19:31.939 11:30:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:19:31.939 11:30:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:19:32.198 11:30:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 -t 2000 00:19:32.457 [ 00:19:32.457 { 00:19:32.457 "name": "BaseBdev4", 00:19:32.457 "aliases": [ 00:19:32.457 "da56d31d-0932-442b-b096-f76808b14eac" 00:19:32.457 ], 00:19:32.457 "product_name": "Malloc disk", 00:19:32.457 "block_size": 512, 00:19:32.457 "num_blocks": 65536, 00:19:32.457 "uuid": "da56d31d-0932-442b-b096-f76808b14eac", 00:19:32.457 "assigned_rate_limits": { 00:19:32.457 "rw_ios_per_sec": 0, 00:19:32.457 "rw_mbytes_per_sec": 0, 00:19:32.457 "r_mbytes_per_sec": 0, 00:19:32.457 "w_mbytes_per_sec": 0 00:19:32.457 }, 00:19:32.457 "claimed": true, 00:19:32.457 "claim_type": "exclusive_write", 00:19:32.457 "zoned": false, 00:19:32.457 "supported_io_types": { 00:19:32.457 "read": true, 00:19:32.457 "write": true, 00:19:32.457 "unmap": true, 00:19:32.457 "flush": true, 00:19:32.457 "reset": true, 00:19:32.457 "nvme_admin": false, 00:19:32.457 "nvme_io": false, 00:19:32.457 "nvme_io_md": false, 00:19:32.457 "write_zeroes": true, 00:19:32.457 "zcopy": true, 00:19:32.457 "get_zone_info": false, 00:19:32.457 "zone_management": false, 00:19:32.457 "zone_append": false, 00:19:32.457 "compare": false, 00:19:32.457 "compare_and_write": false, 00:19:32.457 "abort": true, 00:19:32.457 "seek_hole": false, 00:19:32.458 "seek_data": false, 00:19:32.458 "copy": true, 00:19:32.458 "nvme_iov_md": false 00:19:32.458 }, 00:19:32.458 "memory_domains": [ 00:19:32.458 { 00:19:32.458 "dma_device_id": "system", 00:19:32.458 "dma_device_type": 1 00:19:32.458 }, 00:19:32.458 { 00:19:32.458 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:32.458 "dma_device_type": 2 00:19:32.458 } 00:19:32.458 ], 00:19:32.458 "driver_specific": {} 00:19:32.458 } 00:19:32.458 ] 00:19:32.458 11:30:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:19:32.458 11:30:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:19:32.458 11:30:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:19:32.458 11:30:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:19:32.458 11:30:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:19:32.458 11:30:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:19:32.458 11:30:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:19:32.458 11:30:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:32.458 11:30:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:32.458 11:30:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:32.458 11:30:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:32.458 11:30:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:32.458 11:30:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:32.458 11:30:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:32.458 11:30:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:32.717 11:30:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:32.717 "name": "Existed_Raid", 00:19:32.717 "uuid": "257d5dea-827d-40ab-b632-6b2403c1f2b9", 00:19:32.717 "strip_size_kb": 64, 00:19:32.717 "state": "online", 00:19:32.717 "raid_level": "concat", 00:19:32.717 "superblock": false, 00:19:32.717 "num_base_bdevs": 4, 00:19:32.717 "num_base_bdevs_discovered": 4, 00:19:32.717 "num_base_bdevs_operational": 4, 00:19:32.717 "base_bdevs_list": [ 00:19:32.717 { 00:19:32.717 "name": "BaseBdev1", 00:19:32.717 "uuid": "f2ddc0e6-9dc3-410c-8c78-19594159a130", 00:19:32.717 "is_configured": true, 00:19:32.717 "data_offset": 0, 00:19:32.717 "data_size": 65536 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "name": "BaseBdev2", 00:19:32.717 "uuid": "d23e25d6-16ba-4938-9ad4-3afddb2df86d", 00:19:32.717 "is_configured": true, 00:19:32.717 "data_offset": 0, 00:19:32.717 "data_size": 65536 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "name": "BaseBdev3", 00:19:32.717 "uuid": "60a05ca7-4e34-4264-9c0c-7f8f82d40447", 00:19:32.717 "is_configured": true, 00:19:32.717 "data_offset": 0, 00:19:32.717 "data_size": 65536 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "name": "BaseBdev4", 00:19:32.717 "uuid": "da56d31d-0932-442b-b096-f76808b14eac", 00:19:32.717 "is_configured": true, 00:19:32.717 "data_offset": 0, 00:19:32.717 "data_size": 65536 00:19:32.717 } 00:19:32.717 ] 00:19:32.717 }' 00:19:32.717 11:30:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:32.717 11:30:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:33.284 11:30:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:19:33.284 11:30:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:19:33.284 11:30:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:19:33.284 11:30:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:19:33.284 11:30:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:19:33.284 11:30:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # local name 00:19:33.284 11:30:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:19:33.284 11:30:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:19:33.543 [2024-07-15 11:30:16.916569] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:19:33.543 11:30:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:19:33.543 "name": "Existed_Raid", 00:19:33.543 "aliases": [ 00:19:33.543 "257d5dea-827d-40ab-b632-6b2403c1f2b9" 00:19:33.543 ], 00:19:33.543 "product_name": "Raid Volume", 00:19:33.543 "block_size": 512, 00:19:33.543 "num_blocks": 262144, 00:19:33.543 "uuid": "257d5dea-827d-40ab-b632-6b2403c1f2b9", 00:19:33.543 "assigned_rate_limits": { 00:19:33.543 "rw_ios_per_sec": 0, 00:19:33.543 "rw_mbytes_per_sec": 0, 00:19:33.543 "r_mbytes_per_sec": 0, 00:19:33.543 "w_mbytes_per_sec": 0 00:19:33.543 }, 00:19:33.543 "claimed": false, 00:19:33.543 "zoned": false, 00:19:33.543 "supported_io_types": { 00:19:33.543 "read": true, 00:19:33.543 "write": true, 00:19:33.543 "unmap": true, 00:19:33.543 "flush": true, 00:19:33.543 "reset": true, 00:19:33.543 "nvme_admin": false, 00:19:33.543 "nvme_io": false, 00:19:33.543 "nvme_io_md": false, 00:19:33.543 "write_zeroes": true, 00:19:33.543 "zcopy": false, 00:19:33.543 "get_zone_info": false, 00:19:33.543 "zone_management": false, 00:19:33.543 "zone_append": false, 00:19:33.543 "compare": false, 00:19:33.543 "compare_and_write": false, 00:19:33.543 "abort": false, 00:19:33.543 "seek_hole": false, 00:19:33.543 "seek_data": false, 00:19:33.543 "copy": false, 00:19:33.543 "nvme_iov_md": false 00:19:33.543 }, 00:19:33.543 "memory_domains": [ 00:19:33.543 { 00:19:33.543 "dma_device_id": "system", 00:19:33.543 "dma_device_type": 1 00:19:33.543 }, 00:19:33.543 { 00:19:33.543 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:33.543 "dma_device_type": 2 00:19:33.543 }, 00:19:33.543 { 00:19:33.543 "dma_device_id": "system", 00:19:33.543 "dma_device_type": 1 00:19:33.543 }, 00:19:33.543 { 00:19:33.543 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:33.543 "dma_device_type": 2 00:19:33.543 }, 00:19:33.543 { 00:19:33.543 "dma_device_id": "system", 00:19:33.543 "dma_device_type": 1 00:19:33.543 }, 00:19:33.543 { 00:19:33.543 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:33.543 "dma_device_type": 2 00:19:33.543 }, 00:19:33.543 { 00:19:33.543 "dma_device_id": "system", 00:19:33.543 "dma_device_type": 1 00:19:33.543 }, 00:19:33.543 { 00:19:33.543 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:33.543 "dma_device_type": 2 00:19:33.543 } 00:19:33.543 ], 00:19:33.543 "driver_specific": { 00:19:33.543 "raid": { 00:19:33.543 "uuid": "257d5dea-827d-40ab-b632-6b2403c1f2b9", 00:19:33.543 "strip_size_kb": 64, 00:19:33.543 "state": "online", 00:19:33.543 "raid_level": "concat", 00:19:33.543 "superblock": false, 00:19:33.543 "num_base_bdevs": 4, 00:19:33.543 "num_base_bdevs_discovered": 4, 00:19:33.543 "num_base_bdevs_operational": 4, 00:19:33.543 "base_bdevs_list": [ 00:19:33.543 { 00:19:33.543 "name": "BaseBdev1", 00:19:33.543 "uuid": "f2ddc0e6-9dc3-410c-8c78-19594159a130", 00:19:33.543 "is_configured": true, 00:19:33.543 "data_offset": 0, 00:19:33.543 "data_size": 65536 00:19:33.543 }, 00:19:33.543 { 00:19:33.543 "name": "BaseBdev2", 00:19:33.543 "uuid": "d23e25d6-16ba-4938-9ad4-3afddb2df86d", 00:19:33.543 "is_configured": true, 00:19:33.543 "data_offset": 0, 00:19:33.543 "data_size": 65536 00:19:33.543 }, 00:19:33.543 { 00:19:33.543 "name": "BaseBdev3", 00:19:33.543 "uuid": "60a05ca7-4e34-4264-9c0c-7f8f82d40447", 00:19:33.543 "is_configured": true, 00:19:33.543 "data_offset": 0, 00:19:33.543 "data_size": 65536 00:19:33.543 }, 00:19:33.543 { 00:19:33.543 "name": "BaseBdev4", 00:19:33.543 "uuid": "da56d31d-0932-442b-b096-f76808b14eac", 00:19:33.543 "is_configured": true, 00:19:33.543 "data_offset": 0, 00:19:33.543 "data_size": 65536 00:19:33.543 } 00:19:33.543 ] 00:19:33.543 } 00:19:33.543 } 00:19:33.543 }' 00:19:33.543 11:30:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:19:33.543 11:30:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:19:33.543 BaseBdev2 00:19:33.543 BaseBdev3 00:19:33.543 BaseBdev4' 00:19:33.543 11:30:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:19:33.543 11:30:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:19:33.543 11:30:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:19:33.804 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:19:33.804 "name": "BaseBdev1", 00:19:33.804 "aliases": [ 00:19:33.804 "f2ddc0e6-9dc3-410c-8c78-19594159a130" 00:19:33.804 ], 00:19:33.804 "product_name": "Malloc disk", 00:19:33.804 "block_size": 512, 00:19:33.805 "num_blocks": 65536, 00:19:33.805 "uuid": "f2ddc0e6-9dc3-410c-8c78-19594159a130", 00:19:33.805 "assigned_rate_limits": { 00:19:33.805 "rw_ios_per_sec": 0, 00:19:33.805 "rw_mbytes_per_sec": 0, 00:19:33.805 "r_mbytes_per_sec": 0, 00:19:33.805 "w_mbytes_per_sec": 0 00:19:33.805 }, 00:19:33.805 "claimed": true, 00:19:33.805 "claim_type": "exclusive_write", 00:19:33.805 "zoned": false, 00:19:33.805 "supported_io_types": { 00:19:33.805 "read": true, 00:19:33.805 "write": true, 00:19:33.805 "unmap": true, 00:19:33.805 "flush": true, 00:19:33.805 "reset": true, 00:19:33.805 "nvme_admin": false, 00:19:33.805 "nvme_io": false, 00:19:33.805 "nvme_io_md": false, 00:19:33.805 "write_zeroes": true, 00:19:33.805 "zcopy": true, 00:19:33.805 "get_zone_info": false, 00:19:33.805 "zone_management": false, 00:19:33.805 "zone_append": false, 00:19:33.805 "compare": false, 00:19:33.805 "compare_and_write": false, 00:19:33.805 "abort": true, 00:19:33.805 "seek_hole": false, 00:19:33.805 "seek_data": false, 00:19:33.805 "copy": true, 00:19:33.805 "nvme_iov_md": false 00:19:33.805 }, 00:19:33.805 "memory_domains": [ 00:19:33.805 { 00:19:33.805 "dma_device_id": "system", 00:19:33.805 "dma_device_type": 1 00:19:33.805 }, 00:19:33.805 { 00:19:33.805 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:33.805 "dma_device_type": 2 00:19:33.805 } 00:19:33.805 ], 00:19:33.805 "driver_specific": {} 00:19:33.805 }' 00:19:33.805 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:33.805 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:33.805 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:19:33.805 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:33.805 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:34.063 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:19:34.063 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:34.063 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:34.063 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:19:34.063 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:34.063 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:34.063 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:19:34.063 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:19:34.063 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:19:34.063 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:19:34.321 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:19:34.321 "name": "BaseBdev2", 00:19:34.321 "aliases": [ 00:19:34.321 "d23e25d6-16ba-4938-9ad4-3afddb2df86d" 00:19:34.321 ], 00:19:34.321 "product_name": "Malloc disk", 00:19:34.321 "block_size": 512, 00:19:34.321 "num_blocks": 65536, 00:19:34.321 "uuid": "d23e25d6-16ba-4938-9ad4-3afddb2df86d", 00:19:34.321 "assigned_rate_limits": { 00:19:34.321 "rw_ios_per_sec": 0, 00:19:34.321 "rw_mbytes_per_sec": 0, 00:19:34.321 "r_mbytes_per_sec": 0, 00:19:34.321 "w_mbytes_per_sec": 0 00:19:34.321 }, 00:19:34.321 "claimed": true, 00:19:34.321 "claim_type": "exclusive_write", 00:19:34.321 "zoned": false, 00:19:34.321 "supported_io_types": { 00:19:34.321 "read": true, 00:19:34.321 "write": true, 00:19:34.321 "unmap": true, 00:19:34.321 "flush": true, 00:19:34.321 "reset": true, 00:19:34.321 "nvme_admin": false, 00:19:34.321 "nvme_io": false, 00:19:34.321 "nvme_io_md": false, 00:19:34.321 "write_zeroes": true, 00:19:34.321 "zcopy": true, 00:19:34.321 "get_zone_info": false, 00:19:34.321 "zone_management": false, 00:19:34.321 "zone_append": false, 00:19:34.321 "compare": false, 00:19:34.321 "compare_and_write": false, 00:19:34.321 "abort": true, 00:19:34.321 "seek_hole": false, 00:19:34.321 "seek_data": false, 00:19:34.321 "copy": true, 00:19:34.321 "nvme_iov_md": false 00:19:34.321 }, 00:19:34.321 "memory_domains": [ 00:19:34.321 { 00:19:34.321 "dma_device_id": "system", 00:19:34.321 "dma_device_type": 1 00:19:34.321 }, 00:19:34.321 { 00:19:34.321 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:34.321 "dma_device_type": 2 00:19:34.321 } 00:19:34.321 ], 00:19:34.321 "driver_specific": {} 00:19:34.321 }' 00:19:34.321 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:34.321 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:34.580 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:19:34.580 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:34.580 11:30:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:34.580 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:19:34.580 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:34.580 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:34.580 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:19:34.580 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:34.580 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:34.839 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:19:34.839 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:19:34.839 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:19:34.839 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:19:35.099 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:19:35.099 "name": "BaseBdev3", 00:19:35.099 "aliases": [ 00:19:35.099 "60a05ca7-4e34-4264-9c0c-7f8f82d40447" 00:19:35.099 ], 00:19:35.099 "product_name": "Malloc disk", 00:19:35.099 "block_size": 512, 00:19:35.099 "num_blocks": 65536, 00:19:35.099 "uuid": "60a05ca7-4e34-4264-9c0c-7f8f82d40447", 00:19:35.099 "assigned_rate_limits": { 00:19:35.099 "rw_ios_per_sec": 0, 00:19:35.099 "rw_mbytes_per_sec": 0, 00:19:35.099 "r_mbytes_per_sec": 0, 00:19:35.099 "w_mbytes_per_sec": 0 00:19:35.099 }, 00:19:35.099 "claimed": true, 00:19:35.099 "claim_type": "exclusive_write", 00:19:35.099 "zoned": false, 00:19:35.099 "supported_io_types": { 00:19:35.099 "read": true, 00:19:35.099 "write": true, 00:19:35.099 "unmap": true, 00:19:35.099 "flush": true, 00:19:35.099 "reset": true, 00:19:35.099 "nvme_admin": false, 00:19:35.099 "nvme_io": false, 00:19:35.099 "nvme_io_md": false, 00:19:35.099 "write_zeroes": true, 00:19:35.099 "zcopy": true, 00:19:35.099 "get_zone_info": false, 00:19:35.099 "zone_management": false, 00:19:35.099 "zone_append": false, 00:19:35.099 "compare": false, 00:19:35.099 "compare_and_write": false, 00:19:35.099 "abort": true, 00:19:35.099 "seek_hole": false, 00:19:35.099 "seek_data": false, 00:19:35.099 "copy": true, 00:19:35.099 "nvme_iov_md": false 00:19:35.099 }, 00:19:35.099 "memory_domains": [ 00:19:35.099 { 00:19:35.099 "dma_device_id": "system", 00:19:35.099 "dma_device_type": 1 00:19:35.099 }, 00:19:35.099 { 00:19:35.099 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:35.099 "dma_device_type": 2 00:19:35.099 } 00:19:35.099 ], 00:19:35.099 "driver_specific": {} 00:19:35.099 }' 00:19:35.099 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:35.099 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:35.099 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:19:35.099 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:35.099 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:35.099 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:19:35.099 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:35.099 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:35.358 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:19:35.358 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:35.358 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:35.358 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:19:35.358 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:19:35.358 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 00:19:35.358 11:30:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:19:35.617 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:19:35.617 "name": "BaseBdev4", 00:19:35.617 "aliases": [ 00:19:35.617 "da56d31d-0932-442b-b096-f76808b14eac" 00:19:35.617 ], 00:19:35.617 "product_name": "Malloc disk", 00:19:35.617 "block_size": 512, 00:19:35.617 "num_blocks": 65536, 00:19:35.617 "uuid": "da56d31d-0932-442b-b096-f76808b14eac", 00:19:35.617 "assigned_rate_limits": { 00:19:35.617 "rw_ios_per_sec": 0, 00:19:35.617 "rw_mbytes_per_sec": 0, 00:19:35.617 "r_mbytes_per_sec": 0, 00:19:35.617 "w_mbytes_per_sec": 0 00:19:35.617 }, 00:19:35.617 "claimed": true, 00:19:35.617 "claim_type": "exclusive_write", 00:19:35.617 "zoned": false, 00:19:35.617 "supported_io_types": { 00:19:35.617 "read": true, 00:19:35.617 "write": true, 00:19:35.617 "unmap": true, 00:19:35.617 "flush": true, 00:19:35.617 "reset": true, 00:19:35.617 "nvme_admin": false, 00:19:35.617 "nvme_io": false, 00:19:35.617 "nvme_io_md": false, 00:19:35.617 "write_zeroes": true, 00:19:35.617 "zcopy": true, 00:19:35.617 "get_zone_info": false, 00:19:35.617 "zone_management": false, 00:19:35.617 "zone_append": false, 00:19:35.617 "compare": false, 00:19:35.617 "compare_and_write": false, 00:19:35.617 "abort": true, 00:19:35.617 "seek_hole": false, 00:19:35.617 "seek_data": false, 00:19:35.617 "copy": true, 00:19:35.617 "nvme_iov_md": false 00:19:35.617 }, 00:19:35.617 "memory_domains": [ 00:19:35.617 { 00:19:35.617 "dma_device_id": "system", 00:19:35.617 "dma_device_type": 1 00:19:35.617 }, 00:19:35.617 { 00:19:35.617 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:35.617 "dma_device_type": 2 00:19:35.617 } 00:19:35.617 ], 00:19:35.617 "driver_specific": {} 00:19:35.617 }' 00:19:35.617 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:35.617 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:35.617 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:19:35.617 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:35.617 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:35.617 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:19:35.617 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:35.876 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:35.876 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:19:35.876 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:35.876 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:35.876 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:19:35.876 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:19:36.135 [2024-07-15 11:30:19.603420] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:19:36.135 [2024-07-15 11:30:19.603449] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:19:36.135 [2024-07-15 11:30:19.603498] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:19:36.135 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@275 -- # local expected_state 00:19:36.135 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # has_redundancy concat 00:19:36.135 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:19:36.135 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # return 1 00:19:36.135 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@277 -- # expected_state=offline 00:19:36.135 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:19:36.135 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:19:36.135 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=offline 00:19:36.135 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:19:36.135 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:36.135 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:19:36.135 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:36.135 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:36.135 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:36.135 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:36.135 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:36.135 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:36.394 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:36.394 "name": "Existed_Raid", 00:19:36.395 "uuid": "257d5dea-827d-40ab-b632-6b2403c1f2b9", 00:19:36.395 "strip_size_kb": 64, 00:19:36.395 "state": "offline", 00:19:36.395 "raid_level": "concat", 00:19:36.395 "superblock": false, 00:19:36.395 "num_base_bdevs": 4, 00:19:36.395 "num_base_bdevs_discovered": 3, 00:19:36.395 "num_base_bdevs_operational": 3, 00:19:36.395 "base_bdevs_list": [ 00:19:36.395 { 00:19:36.395 "name": null, 00:19:36.395 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:36.395 "is_configured": false, 00:19:36.395 "data_offset": 0, 00:19:36.395 "data_size": 65536 00:19:36.395 }, 00:19:36.395 { 00:19:36.395 "name": "BaseBdev2", 00:19:36.395 "uuid": "d23e25d6-16ba-4938-9ad4-3afddb2df86d", 00:19:36.395 "is_configured": true, 00:19:36.395 "data_offset": 0, 00:19:36.395 "data_size": 65536 00:19:36.395 }, 00:19:36.395 { 00:19:36.395 "name": "BaseBdev3", 00:19:36.395 "uuid": "60a05ca7-4e34-4264-9c0c-7f8f82d40447", 00:19:36.395 "is_configured": true, 00:19:36.395 "data_offset": 0, 00:19:36.395 "data_size": 65536 00:19:36.395 }, 00:19:36.395 { 00:19:36.395 "name": "BaseBdev4", 00:19:36.395 "uuid": "da56d31d-0932-442b-b096-f76808b14eac", 00:19:36.395 "is_configured": true, 00:19:36.395 "data_offset": 0, 00:19:36.395 "data_size": 65536 00:19:36.395 } 00:19:36.395 ] 00:19:36.395 }' 00:19:36.395 11:30:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:36.395 11:30:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:36.961 11:30:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:19:36.961 11:30:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:19:36.961 11:30:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:36.961 11:30:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:19:37.220 11:30:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:19:37.220 11:30:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:19:37.220 11:30:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:19:37.478 [2024-07-15 11:30:20.940892] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:19:37.478 11:30:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:19:37.478 11:30:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:19:37.478 11:30:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:37.478 11:30:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:19:37.737 11:30:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:19:37.737 11:30:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:19:37.737 11:30:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev3 00:19:37.996 [2024-07-15 11:30:21.452905] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:19:37.996 11:30:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:19:37.996 11:30:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:19:37.996 11:30:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:37.996 11:30:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:19:38.285 11:30:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:19:38.285 11:30:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:19:38.285 11:30:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev4 00:19:38.552 [2024-07-15 11:30:21.962602] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:19:38.552 [2024-07-15 11:30:21.962650] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x132a350 name Existed_Raid, state offline 00:19:38.552 11:30:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:19:38.552 11:30:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:19:38.552 11:30:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:38.552 11:30:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:19:38.809 11:30:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:19:38.809 11:30:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:19:38.809 11:30:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # '[' 4 -gt 2 ']' 00:19:38.809 11:30:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i = 1 )) 00:19:38.809 11:30:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:19:38.809 11:30:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:19:39.068 BaseBdev2 00:19:39.068 11:30:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev2 00:19:39.068 11:30:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:19:39.068 11:30:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:19:39.068 11:30:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:19:39.068 11:30:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:19:39.068 11:30:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:19:39.068 11:30:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:19:39.326 11:30:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:19:39.585 [ 00:19:39.585 { 00:19:39.585 "name": "BaseBdev2", 00:19:39.585 "aliases": [ 00:19:39.585 "a1ecaccc-b270-4f0f-8099-a5a4ffd9d05e" 00:19:39.585 ], 00:19:39.585 "product_name": "Malloc disk", 00:19:39.585 "block_size": 512, 00:19:39.585 "num_blocks": 65536, 00:19:39.585 "uuid": "a1ecaccc-b270-4f0f-8099-a5a4ffd9d05e", 00:19:39.585 "assigned_rate_limits": { 00:19:39.585 "rw_ios_per_sec": 0, 00:19:39.585 "rw_mbytes_per_sec": 0, 00:19:39.585 "r_mbytes_per_sec": 0, 00:19:39.585 "w_mbytes_per_sec": 0 00:19:39.585 }, 00:19:39.585 "claimed": false, 00:19:39.585 "zoned": false, 00:19:39.585 "supported_io_types": { 00:19:39.585 "read": true, 00:19:39.585 "write": true, 00:19:39.585 "unmap": true, 00:19:39.585 "flush": true, 00:19:39.585 "reset": true, 00:19:39.585 "nvme_admin": false, 00:19:39.585 "nvme_io": false, 00:19:39.585 "nvme_io_md": false, 00:19:39.585 "write_zeroes": true, 00:19:39.585 "zcopy": true, 00:19:39.585 "get_zone_info": false, 00:19:39.585 "zone_management": false, 00:19:39.585 "zone_append": false, 00:19:39.585 "compare": false, 00:19:39.585 "compare_and_write": false, 00:19:39.585 "abort": true, 00:19:39.585 "seek_hole": false, 00:19:39.585 "seek_data": false, 00:19:39.585 "copy": true, 00:19:39.585 "nvme_iov_md": false 00:19:39.585 }, 00:19:39.585 "memory_domains": [ 00:19:39.585 { 00:19:39.585 "dma_device_id": "system", 00:19:39.585 "dma_device_type": 1 00:19:39.585 }, 00:19:39.585 { 00:19:39.585 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:39.585 "dma_device_type": 2 00:19:39.585 } 00:19:39.585 ], 00:19:39.585 "driver_specific": {} 00:19:39.585 } 00:19:39.585 ] 00:19:39.585 11:30:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:19:39.585 11:30:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:19:39.585 11:30:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:19:39.585 11:30:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:19:39.843 BaseBdev3 00:19:39.843 11:30:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev3 00:19:39.843 11:30:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:19:39.843 11:30:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:19:39.843 11:30:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:19:39.843 11:30:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:19:39.843 11:30:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:19:39.843 11:30:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:19:40.101 11:30:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:19:40.358 [ 00:19:40.358 { 00:19:40.358 "name": "BaseBdev3", 00:19:40.358 "aliases": [ 00:19:40.358 "3c3a272d-8184-4866-810c-8c512613a0d5" 00:19:40.358 ], 00:19:40.358 "product_name": "Malloc disk", 00:19:40.358 "block_size": 512, 00:19:40.358 "num_blocks": 65536, 00:19:40.358 "uuid": "3c3a272d-8184-4866-810c-8c512613a0d5", 00:19:40.358 "assigned_rate_limits": { 00:19:40.358 "rw_ios_per_sec": 0, 00:19:40.358 "rw_mbytes_per_sec": 0, 00:19:40.358 "r_mbytes_per_sec": 0, 00:19:40.358 "w_mbytes_per_sec": 0 00:19:40.358 }, 00:19:40.358 "claimed": false, 00:19:40.358 "zoned": false, 00:19:40.358 "supported_io_types": { 00:19:40.358 "read": true, 00:19:40.358 "write": true, 00:19:40.358 "unmap": true, 00:19:40.358 "flush": true, 00:19:40.358 "reset": true, 00:19:40.358 "nvme_admin": false, 00:19:40.358 "nvme_io": false, 00:19:40.358 "nvme_io_md": false, 00:19:40.358 "write_zeroes": true, 00:19:40.358 "zcopy": true, 00:19:40.358 "get_zone_info": false, 00:19:40.358 "zone_management": false, 00:19:40.358 "zone_append": false, 00:19:40.358 "compare": false, 00:19:40.358 "compare_and_write": false, 00:19:40.358 "abort": true, 00:19:40.358 "seek_hole": false, 00:19:40.358 "seek_data": false, 00:19:40.358 "copy": true, 00:19:40.358 "nvme_iov_md": false 00:19:40.358 }, 00:19:40.358 "memory_domains": [ 00:19:40.358 { 00:19:40.358 "dma_device_id": "system", 00:19:40.358 "dma_device_type": 1 00:19:40.358 }, 00:19:40.358 { 00:19:40.358 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:40.358 "dma_device_type": 2 00:19:40.358 } 00:19:40.358 ], 00:19:40.358 "driver_specific": {} 00:19:40.358 } 00:19:40.358 ] 00:19:40.358 11:30:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:19:40.358 11:30:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:19:40.358 11:30:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:19:40.358 11:30:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4 00:19:40.615 BaseBdev4 00:19:40.615 11:30:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev4 00:19:40.615 11:30:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev4 00:19:40.615 11:30:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:19:40.615 11:30:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:19:40.615 11:30:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:19:40.615 11:30:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:19:40.615 11:30:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:19:40.873 11:30:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 -t 2000 00:19:40.873 [ 00:19:40.873 { 00:19:40.873 "name": "BaseBdev4", 00:19:40.873 "aliases": [ 00:19:40.873 "28a269d6-c2ec-444b-ac2c-259984fcb510" 00:19:40.873 ], 00:19:40.873 "product_name": "Malloc disk", 00:19:40.873 "block_size": 512, 00:19:40.873 "num_blocks": 65536, 00:19:40.873 "uuid": "28a269d6-c2ec-444b-ac2c-259984fcb510", 00:19:40.873 "assigned_rate_limits": { 00:19:40.873 "rw_ios_per_sec": 0, 00:19:40.873 "rw_mbytes_per_sec": 0, 00:19:40.873 "r_mbytes_per_sec": 0, 00:19:40.873 "w_mbytes_per_sec": 0 00:19:40.873 }, 00:19:40.873 "claimed": false, 00:19:40.873 "zoned": false, 00:19:40.873 "supported_io_types": { 00:19:40.873 "read": true, 00:19:40.873 "write": true, 00:19:40.873 "unmap": true, 00:19:40.873 "flush": true, 00:19:40.873 "reset": true, 00:19:40.873 "nvme_admin": false, 00:19:40.873 "nvme_io": false, 00:19:40.873 "nvme_io_md": false, 00:19:40.873 "write_zeroes": true, 00:19:40.873 "zcopy": true, 00:19:40.873 "get_zone_info": false, 00:19:40.873 "zone_management": false, 00:19:40.873 "zone_append": false, 00:19:40.873 "compare": false, 00:19:40.873 "compare_and_write": false, 00:19:40.873 "abort": true, 00:19:40.873 "seek_hole": false, 00:19:40.873 "seek_data": false, 00:19:40.873 "copy": true, 00:19:40.873 "nvme_iov_md": false 00:19:40.873 }, 00:19:40.873 "memory_domains": [ 00:19:40.873 { 00:19:40.873 "dma_device_id": "system", 00:19:40.873 "dma_device_type": 1 00:19:40.873 }, 00:19:40.873 { 00:19:40.873 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:40.873 "dma_device_type": 2 00:19:40.873 } 00:19:40.873 ], 00:19:40.873 "driver_specific": {} 00:19:40.873 } 00:19:40.873 ] 00:19:40.873 11:30:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:19:40.873 11:30:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:19:40.873 11:30:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:19:40.873 11:30:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@305 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:19:41.131 [2024-07-15 11:30:24.609276] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:19:41.131 [2024-07-15 11:30:24.609318] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:19:41.131 [2024-07-15 11:30:24.609339] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:19:41.131 [2024-07-15 11:30:24.610653] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:19:41.131 [2024-07-15 11:30:24.610702] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:19:41.131 11:30:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:19:41.131 11:30:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:19:41.131 11:30:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:19:41.131 11:30:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:19:41.131 11:30:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:41.131 11:30:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:41.131 11:30:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:41.131 11:30:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:41.131 11:30:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:41.131 11:30:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:41.131 11:30:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:41.131 11:30:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:41.389 11:30:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:41.389 "name": "Existed_Raid", 00:19:41.389 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:41.389 "strip_size_kb": 64, 00:19:41.389 "state": "configuring", 00:19:41.389 "raid_level": "concat", 00:19:41.389 "superblock": false, 00:19:41.389 "num_base_bdevs": 4, 00:19:41.389 "num_base_bdevs_discovered": 3, 00:19:41.389 "num_base_bdevs_operational": 4, 00:19:41.389 "base_bdevs_list": [ 00:19:41.389 { 00:19:41.389 "name": "BaseBdev1", 00:19:41.389 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:41.389 "is_configured": false, 00:19:41.389 "data_offset": 0, 00:19:41.389 "data_size": 0 00:19:41.389 }, 00:19:41.389 { 00:19:41.389 "name": "BaseBdev2", 00:19:41.389 "uuid": "a1ecaccc-b270-4f0f-8099-a5a4ffd9d05e", 00:19:41.389 "is_configured": true, 00:19:41.389 "data_offset": 0, 00:19:41.389 "data_size": 65536 00:19:41.389 }, 00:19:41.389 { 00:19:41.389 "name": "BaseBdev3", 00:19:41.389 "uuid": "3c3a272d-8184-4866-810c-8c512613a0d5", 00:19:41.389 "is_configured": true, 00:19:41.389 "data_offset": 0, 00:19:41.389 "data_size": 65536 00:19:41.389 }, 00:19:41.389 { 00:19:41.389 "name": "BaseBdev4", 00:19:41.389 "uuid": "28a269d6-c2ec-444b-ac2c-259984fcb510", 00:19:41.389 "is_configured": true, 00:19:41.389 "data_offset": 0, 00:19:41.389 "data_size": 65536 00:19:41.389 } 00:19:41.389 ] 00:19:41.389 }' 00:19:41.389 11:30:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:41.389 11:30:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:41.955 11:30:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev2 00:19:42.214 [2024-07-15 11:30:25.712178] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:19:42.214 11:30:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@309 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:19:42.214 11:30:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:19:42.214 11:30:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:19:42.214 11:30:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:19:42.214 11:30:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:42.214 11:30:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:42.214 11:30:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:42.214 11:30:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:42.214 11:30:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:42.214 11:30:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:42.214 11:30:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:42.214 11:30:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:42.473 11:30:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:42.473 "name": "Existed_Raid", 00:19:42.473 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:42.473 "strip_size_kb": 64, 00:19:42.473 "state": "configuring", 00:19:42.473 "raid_level": "concat", 00:19:42.473 "superblock": false, 00:19:42.473 "num_base_bdevs": 4, 00:19:42.473 "num_base_bdevs_discovered": 2, 00:19:42.473 "num_base_bdevs_operational": 4, 00:19:42.473 "base_bdevs_list": [ 00:19:42.473 { 00:19:42.473 "name": "BaseBdev1", 00:19:42.473 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:42.473 "is_configured": false, 00:19:42.473 "data_offset": 0, 00:19:42.473 "data_size": 0 00:19:42.473 }, 00:19:42.473 { 00:19:42.473 "name": null, 00:19:42.473 "uuid": "a1ecaccc-b270-4f0f-8099-a5a4ffd9d05e", 00:19:42.473 "is_configured": false, 00:19:42.473 "data_offset": 0, 00:19:42.473 "data_size": 65536 00:19:42.473 }, 00:19:42.473 { 00:19:42.473 "name": "BaseBdev3", 00:19:42.473 "uuid": "3c3a272d-8184-4866-810c-8c512613a0d5", 00:19:42.473 "is_configured": true, 00:19:42.473 "data_offset": 0, 00:19:42.473 "data_size": 65536 00:19:42.473 }, 00:19:42.473 { 00:19:42.473 "name": "BaseBdev4", 00:19:42.473 "uuid": "28a269d6-c2ec-444b-ac2c-259984fcb510", 00:19:42.473 "is_configured": true, 00:19:42.473 "data_offset": 0, 00:19:42.473 "data_size": 65536 00:19:42.473 } 00:19:42.473 ] 00:19:42.473 }' 00:19:42.473 11:30:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:42.473 11:30:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:43.041 11:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:43.041 11:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:19:43.300 11:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # [[ false == \f\a\l\s\e ]] 00:19:43.300 11:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:19:43.567 [2024-07-15 11:30:27.003873] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:19:43.567 BaseBdev1 00:19:43.567 11:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@313 -- # waitforbdev BaseBdev1 00:19:43.567 11:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:19:43.567 11:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:19:43.567 11:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:19:43.567 11:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:19:43.567 11:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:19:43.567 11:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:19:43.828 11:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:19:44.087 [ 00:19:44.087 { 00:19:44.087 "name": "BaseBdev1", 00:19:44.087 "aliases": [ 00:19:44.087 "15b3a9f0-c4a7-4e94-9a29-c2ca91f82f3d" 00:19:44.087 ], 00:19:44.087 "product_name": "Malloc disk", 00:19:44.087 "block_size": 512, 00:19:44.087 "num_blocks": 65536, 00:19:44.087 "uuid": "15b3a9f0-c4a7-4e94-9a29-c2ca91f82f3d", 00:19:44.087 "assigned_rate_limits": { 00:19:44.087 "rw_ios_per_sec": 0, 00:19:44.087 "rw_mbytes_per_sec": 0, 00:19:44.087 "r_mbytes_per_sec": 0, 00:19:44.087 "w_mbytes_per_sec": 0 00:19:44.087 }, 00:19:44.087 "claimed": true, 00:19:44.087 "claim_type": "exclusive_write", 00:19:44.087 "zoned": false, 00:19:44.087 "supported_io_types": { 00:19:44.087 "read": true, 00:19:44.087 "write": true, 00:19:44.087 "unmap": true, 00:19:44.087 "flush": true, 00:19:44.087 "reset": true, 00:19:44.087 "nvme_admin": false, 00:19:44.087 "nvme_io": false, 00:19:44.087 "nvme_io_md": false, 00:19:44.087 "write_zeroes": true, 00:19:44.087 "zcopy": true, 00:19:44.087 "get_zone_info": false, 00:19:44.087 "zone_management": false, 00:19:44.087 "zone_append": false, 00:19:44.087 "compare": false, 00:19:44.087 "compare_and_write": false, 00:19:44.087 "abort": true, 00:19:44.087 "seek_hole": false, 00:19:44.087 "seek_data": false, 00:19:44.087 "copy": true, 00:19:44.087 "nvme_iov_md": false 00:19:44.087 }, 00:19:44.087 "memory_domains": [ 00:19:44.087 { 00:19:44.087 "dma_device_id": "system", 00:19:44.087 "dma_device_type": 1 00:19:44.087 }, 00:19:44.087 { 00:19:44.087 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:44.087 "dma_device_type": 2 00:19:44.087 } 00:19:44.087 ], 00:19:44.087 "driver_specific": {} 00:19:44.087 } 00:19:44.087 ] 00:19:44.087 11:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:19:44.087 11:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:19:44.087 11:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:19:44.087 11:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:19:44.087 11:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:19:44.087 11:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:44.087 11:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:44.087 11:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:44.087 11:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:44.087 11:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:44.087 11:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:44.087 11:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:44.087 11:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:44.346 11:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:44.346 "name": "Existed_Raid", 00:19:44.346 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:44.346 "strip_size_kb": 64, 00:19:44.346 "state": "configuring", 00:19:44.346 "raid_level": "concat", 00:19:44.346 "superblock": false, 00:19:44.346 "num_base_bdevs": 4, 00:19:44.346 "num_base_bdevs_discovered": 3, 00:19:44.346 "num_base_bdevs_operational": 4, 00:19:44.346 "base_bdevs_list": [ 00:19:44.346 { 00:19:44.346 "name": "BaseBdev1", 00:19:44.346 "uuid": "15b3a9f0-c4a7-4e94-9a29-c2ca91f82f3d", 00:19:44.346 "is_configured": true, 00:19:44.346 "data_offset": 0, 00:19:44.346 "data_size": 65536 00:19:44.346 }, 00:19:44.346 { 00:19:44.346 "name": null, 00:19:44.346 "uuid": "a1ecaccc-b270-4f0f-8099-a5a4ffd9d05e", 00:19:44.346 "is_configured": false, 00:19:44.347 "data_offset": 0, 00:19:44.347 "data_size": 65536 00:19:44.347 }, 00:19:44.347 { 00:19:44.347 "name": "BaseBdev3", 00:19:44.347 "uuid": "3c3a272d-8184-4866-810c-8c512613a0d5", 00:19:44.347 "is_configured": true, 00:19:44.347 "data_offset": 0, 00:19:44.347 "data_size": 65536 00:19:44.347 }, 00:19:44.347 { 00:19:44.347 "name": "BaseBdev4", 00:19:44.347 "uuid": "28a269d6-c2ec-444b-ac2c-259984fcb510", 00:19:44.347 "is_configured": true, 00:19:44.347 "data_offset": 0, 00:19:44.347 "data_size": 65536 00:19:44.347 } 00:19:44.347 ] 00:19:44.347 }' 00:19:44.347 11:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:44.347 11:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:44.914 11:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:44.914 11:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:19:45.173 11:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # [[ true == \t\r\u\e ]] 00:19:45.173 11:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@317 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev3 00:19:45.433 [2024-07-15 11:30:28.832749] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:19:45.433 11:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:19:45.433 11:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:19:45.433 11:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:19:45.433 11:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:19:45.433 11:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:45.433 11:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:45.433 11:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:45.433 11:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:45.433 11:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:45.433 11:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:45.433 11:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:45.433 11:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:45.691 11:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:45.691 "name": "Existed_Raid", 00:19:45.691 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:45.691 "strip_size_kb": 64, 00:19:45.691 "state": "configuring", 00:19:45.691 "raid_level": "concat", 00:19:45.691 "superblock": false, 00:19:45.691 "num_base_bdevs": 4, 00:19:45.691 "num_base_bdevs_discovered": 2, 00:19:45.691 "num_base_bdevs_operational": 4, 00:19:45.692 "base_bdevs_list": [ 00:19:45.692 { 00:19:45.692 "name": "BaseBdev1", 00:19:45.692 "uuid": "15b3a9f0-c4a7-4e94-9a29-c2ca91f82f3d", 00:19:45.692 "is_configured": true, 00:19:45.692 "data_offset": 0, 00:19:45.692 "data_size": 65536 00:19:45.692 }, 00:19:45.692 { 00:19:45.692 "name": null, 00:19:45.692 "uuid": "a1ecaccc-b270-4f0f-8099-a5a4ffd9d05e", 00:19:45.692 "is_configured": false, 00:19:45.692 "data_offset": 0, 00:19:45.692 "data_size": 65536 00:19:45.692 }, 00:19:45.692 { 00:19:45.692 "name": null, 00:19:45.692 "uuid": "3c3a272d-8184-4866-810c-8c512613a0d5", 00:19:45.692 "is_configured": false, 00:19:45.692 "data_offset": 0, 00:19:45.692 "data_size": 65536 00:19:45.692 }, 00:19:45.692 { 00:19:45.692 "name": "BaseBdev4", 00:19:45.692 "uuid": "28a269d6-c2ec-444b-ac2c-259984fcb510", 00:19:45.692 "is_configured": true, 00:19:45.692 "data_offset": 0, 00:19:45.692 "data_size": 65536 00:19:45.692 } 00:19:45.692 ] 00:19:45.692 }' 00:19:45.692 11:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:45.692 11:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:46.259 11:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:46.259 11:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:19:46.518 11:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # [[ false == \f\a\l\s\e ]] 00:19:46.518 11:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:19:46.776 [2024-07-15 11:30:30.128226] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:19:46.776 11:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@322 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:19:46.776 11:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:19:46.776 11:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:19:46.776 11:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:19:46.776 11:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:46.776 11:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:46.776 11:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:46.776 11:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:46.776 11:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:46.776 11:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:46.776 11:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:46.776 11:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:47.034 11:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:47.034 "name": "Existed_Raid", 00:19:47.034 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:47.034 "strip_size_kb": 64, 00:19:47.034 "state": "configuring", 00:19:47.034 "raid_level": "concat", 00:19:47.034 "superblock": false, 00:19:47.034 "num_base_bdevs": 4, 00:19:47.034 "num_base_bdevs_discovered": 3, 00:19:47.034 "num_base_bdevs_operational": 4, 00:19:47.034 "base_bdevs_list": [ 00:19:47.034 { 00:19:47.034 "name": "BaseBdev1", 00:19:47.034 "uuid": "15b3a9f0-c4a7-4e94-9a29-c2ca91f82f3d", 00:19:47.034 "is_configured": true, 00:19:47.034 "data_offset": 0, 00:19:47.034 "data_size": 65536 00:19:47.034 }, 00:19:47.034 { 00:19:47.034 "name": null, 00:19:47.034 "uuid": "a1ecaccc-b270-4f0f-8099-a5a4ffd9d05e", 00:19:47.034 "is_configured": false, 00:19:47.034 "data_offset": 0, 00:19:47.034 "data_size": 65536 00:19:47.034 }, 00:19:47.034 { 00:19:47.034 "name": "BaseBdev3", 00:19:47.034 "uuid": "3c3a272d-8184-4866-810c-8c512613a0d5", 00:19:47.034 "is_configured": true, 00:19:47.034 "data_offset": 0, 00:19:47.034 "data_size": 65536 00:19:47.034 }, 00:19:47.034 { 00:19:47.034 "name": "BaseBdev4", 00:19:47.034 "uuid": "28a269d6-c2ec-444b-ac2c-259984fcb510", 00:19:47.034 "is_configured": true, 00:19:47.034 "data_offset": 0, 00:19:47.034 "data_size": 65536 00:19:47.034 } 00:19:47.034 ] 00:19:47.034 }' 00:19:47.034 11:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:47.034 11:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:47.598 11:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:47.598 11:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:19:47.598 11:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # [[ true == \t\r\u\e ]] 00:19:47.598 11:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@325 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:19:47.855 [2024-07-15 11:30:31.371509] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:19:47.855 11:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:19:47.855 11:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:19:47.855 11:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:19:47.855 11:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:19:47.855 11:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:47.855 11:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:47.855 11:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:47.855 11:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:47.855 11:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:47.855 11:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:47.855 11:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:47.855 11:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:48.113 11:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:48.113 "name": "Existed_Raid", 00:19:48.113 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:48.113 "strip_size_kb": 64, 00:19:48.113 "state": "configuring", 00:19:48.113 "raid_level": "concat", 00:19:48.113 "superblock": false, 00:19:48.113 "num_base_bdevs": 4, 00:19:48.113 "num_base_bdevs_discovered": 2, 00:19:48.113 "num_base_bdevs_operational": 4, 00:19:48.113 "base_bdevs_list": [ 00:19:48.113 { 00:19:48.113 "name": null, 00:19:48.113 "uuid": "15b3a9f0-c4a7-4e94-9a29-c2ca91f82f3d", 00:19:48.113 "is_configured": false, 00:19:48.113 "data_offset": 0, 00:19:48.113 "data_size": 65536 00:19:48.113 }, 00:19:48.113 { 00:19:48.113 "name": null, 00:19:48.113 "uuid": "a1ecaccc-b270-4f0f-8099-a5a4ffd9d05e", 00:19:48.113 "is_configured": false, 00:19:48.113 "data_offset": 0, 00:19:48.113 "data_size": 65536 00:19:48.113 }, 00:19:48.113 { 00:19:48.113 "name": "BaseBdev3", 00:19:48.113 "uuid": "3c3a272d-8184-4866-810c-8c512613a0d5", 00:19:48.113 "is_configured": true, 00:19:48.113 "data_offset": 0, 00:19:48.113 "data_size": 65536 00:19:48.113 }, 00:19:48.113 { 00:19:48.113 "name": "BaseBdev4", 00:19:48.113 "uuid": "28a269d6-c2ec-444b-ac2c-259984fcb510", 00:19:48.113 "is_configured": true, 00:19:48.113 "data_offset": 0, 00:19:48.113 "data_size": 65536 00:19:48.113 } 00:19:48.113 ] 00:19:48.113 }' 00:19:48.113 11:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:48.113 11:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:48.678 11:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@327 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:48.678 11:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@327 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:19:48.935 11:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@327 -- # [[ false == \f\a\l\s\e ]] 00:19:48.935 11:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@329 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:19:49.193 [2024-07-15 11:30:32.715530] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:19:49.193 11:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@330 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:19:49.193 11:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:19:49.193 11:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:19:49.193 11:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:19:49.193 11:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:49.193 11:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:49.193 11:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:49.193 11:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:49.193 11:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:49.193 11:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:49.193 11:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:49.193 11:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:49.451 11:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:49.451 "name": "Existed_Raid", 00:19:49.451 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:49.451 "strip_size_kb": 64, 00:19:49.451 "state": "configuring", 00:19:49.451 "raid_level": "concat", 00:19:49.451 "superblock": false, 00:19:49.451 "num_base_bdevs": 4, 00:19:49.451 "num_base_bdevs_discovered": 3, 00:19:49.451 "num_base_bdevs_operational": 4, 00:19:49.451 "base_bdevs_list": [ 00:19:49.451 { 00:19:49.451 "name": null, 00:19:49.451 "uuid": "15b3a9f0-c4a7-4e94-9a29-c2ca91f82f3d", 00:19:49.451 "is_configured": false, 00:19:49.451 "data_offset": 0, 00:19:49.451 "data_size": 65536 00:19:49.451 }, 00:19:49.451 { 00:19:49.451 "name": "BaseBdev2", 00:19:49.451 "uuid": "a1ecaccc-b270-4f0f-8099-a5a4ffd9d05e", 00:19:49.451 "is_configured": true, 00:19:49.451 "data_offset": 0, 00:19:49.451 "data_size": 65536 00:19:49.451 }, 00:19:49.451 { 00:19:49.451 "name": "BaseBdev3", 00:19:49.451 "uuid": "3c3a272d-8184-4866-810c-8c512613a0d5", 00:19:49.451 "is_configured": true, 00:19:49.451 "data_offset": 0, 00:19:49.451 "data_size": 65536 00:19:49.451 }, 00:19:49.451 { 00:19:49.451 "name": "BaseBdev4", 00:19:49.451 "uuid": "28a269d6-c2ec-444b-ac2c-259984fcb510", 00:19:49.451 "is_configured": true, 00:19:49.451 "data_offset": 0, 00:19:49.451 "data_size": 65536 00:19:49.451 } 00:19:49.451 ] 00:19:49.451 }' 00:19:49.451 11:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:49.451 11:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:50.015 11:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@331 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:50.015 11:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@331 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:19:50.272 11:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@331 -- # [[ true == \t\r\u\e ]] 00:19:50.272 11:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:50.272 11:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@333 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:19:50.529 11:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b NewBaseBdev -u 15b3a9f0-c4a7-4e94-9a29-c2ca91f82f3d 00:19:50.787 [2024-07-15 11:30:34.299345] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:19:50.787 [2024-07-15 11:30:34.299383] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x132e040 00:19:50.787 [2024-07-15 11:30:34.299392] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:19:50.787 [2024-07-15 11:30:34.299587] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1329a70 00:19:50.787 [2024-07-15 11:30:34.299706] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x132e040 00:19:50.787 [2024-07-15 11:30:34.299716] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x132e040 00:19:50.787 [2024-07-15 11:30:34.299873] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:50.787 NewBaseBdev 00:19:50.787 11:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@334 -- # waitforbdev NewBaseBdev 00:19:50.787 11:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=NewBaseBdev 00:19:50.787 11:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:19:50.787 11:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:19:50.787 11:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:19:50.787 11:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:19:50.787 11:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:19:51.045 11:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev -t 2000 00:19:51.302 [ 00:19:51.302 { 00:19:51.302 "name": "NewBaseBdev", 00:19:51.302 "aliases": [ 00:19:51.302 "15b3a9f0-c4a7-4e94-9a29-c2ca91f82f3d" 00:19:51.302 ], 00:19:51.302 "product_name": "Malloc disk", 00:19:51.302 "block_size": 512, 00:19:51.302 "num_blocks": 65536, 00:19:51.302 "uuid": "15b3a9f0-c4a7-4e94-9a29-c2ca91f82f3d", 00:19:51.302 "assigned_rate_limits": { 00:19:51.302 "rw_ios_per_sec": 0, 00:19:51.302 "rw_mbytes_per_sec": 0, 00:19:51.302 "r_mbytes_per_sec": 0, 00:19:51.302 "w_mbytes_per_sec": 0 00:19:51.302 }, 00:19:51.302 "claimed": true, 00:19:51.302 "claim_type": "exclusive_write", 00:19:51.302 "zoned": false, 00:19:51.302 "supported_io_types": { 00:19:51.302 "read": true, 00:19:51.302 "write": true, 00:19:51.302 "unmap": true, 00:19:51.302 "flush": true, 00:19:51.302 "reset": true, 00:19:51.302 "nvme_admin": false, 00:19:51.302 "nvme_io": false, 00:19:51.302 "nvme_io_md": false, 00:19:51.302 "write_zeroes": true, 00:19:51.302 "zcopy": true, 00:19:51.302 "get_zone_info": false, 00:19:51.302 "zone_management": false, 00:19:51.302 "zone_append": false, 00:19:51.302 "compare": false, 00:19:51.302 "compare_and_write": false, 00:19:51.302 "abort": true, 00:19:51.302 "seek_hole": false, 00:19:51.302 "seek_data": false, 00:19:51.302 "copy": true, 00:19:51.302 "nvme_iov_md": false 00:19:51.302 }, 00:19:51.302 "memory_domains": [ 00:19:51.302 { 00:19:51.302 "dma_device_id": "system", 00:19:51.302 "dma_device_type": 1 00:19:51.302 }, 00:19:51.302 { 00:19:51.302 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:51.302 "dma_device_type": 2 00:19:51.302 } 00:19:51.302 ], 00:19:51.302 "driver_specific": {} 00:19:51.302 } 00:19:51.302 ] 00:19:51.302 11:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:19:51.302 11:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@335 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:19:51.302 11:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:19:51.302 11:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:19:51.302 11:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:19:51.302 11:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:51.302 11:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:51.302 11:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:51.302 11:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:51.302 11:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:51.302 11:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:51.302 11:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:51.303 11:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:51.561 11:30:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:51.561 "name": "Existed_Raid", 00:19:51.561 "uuid": "d13087eb-a9fd-4d45-998c-a7b7b455deba", 00:19:51.561 "strip_size_kb": 64, 00:19:51.561 "state": "online", 00:19:51.561 "raid_level": "concat", 00:19:51.561 "superblock": false, 00:19:51.561 "num_base_bdevs": 4, 00:19:51.561 "num_base_bdevs_discovered": 4, 00:19:51.561 "num_base_bdevs_operational": 4, 00:19:51.561 "base_bdevs_list": [ 00:19:51.561 { 00:19:51.561 "name": "NewBaseBdev", 00:19:51.561 "uuid": "15b3a9f0-c4a7-4e94-9a29-c2ca91f82f3d", 00:19:51.561 "is_configured": true, 00:19:51.561 "data_offset": 0, 00:19:51.561 "data_size": 65536 00:19:51.561 }, 00:19:51.561 { 00:19:51.561 "name": "BaseBdev2", 00:19:51.561 "uuid": "a1ecaccc-b270-4f0f-8099-a5a4ffd9d05e", 00:19:51.561 "is_configured": true, 00:19:51.561 "data_offset": 0, 00:19:51.561 "data_size": 65536 00:19:51.561 }, 00:19:51.561 { 00:19:51.561 "name": "BaseBdev3", 00:19:51.561 "uuid": "3c3a272d-8184-4866-810c-8c512613a0d5", 00:19:51.561 "is_configured": true, 00:19:51.561 "data_offset": 0, 00:19:51.561 "data_size": 65536 00:19:51.561 }, 00:19:51.561 { 00:19:51.561 "name": "BaseBdev4", 00:19:51.561 "uuid": "28a269d6-c2ec-444b-ac2c-259984fcb510", 00:19:51.561 "is_configured": true, 00:19:51.561 "data_offset": 0, 00:19:51.561 "data_size": 65536 00:19:51.561 } 00:19:51.561 ] 00:19:51.561 }' 00:19:51.561 11:30:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:51.561 11:30:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:52.125 11:30:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@336 -- # verify_raid_bdev_properties Existed_Raid 00:19:52.125 11:30:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:19:52.125 11:30:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:19:52.125 11:30:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:19:52.125 11:30:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:19:52.125 11:30:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # local name 00:19:52.125 11:30:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:19:52.125 11:30:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:19:52.385 [2024-07-15 11:30:35.891894] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:19:52.385 11:30:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:19:52.385 "name": "Existed_Raid", 00:19:52.385 "aliases": [ 00:19:52.385 "d13087eb-a9fd-4d45-998c-a7b7b455deba" 00:19:52.385 ], 00:19:52.385 "product_name": "Raid Volume", 00:19:52.385 "block_size": 512, 00:19:52.385 "num_blocks": 262144, 00:19:52.385 "uuid": "d13087eb-a9fd-4d45-998c-a7b7b455deba", 00:19:52.385 "assigned_rate_limits": { 00:19:52.385 "rw_ios_per_sec": 0, 00:19:52.385 "rw_mbytes_per_sec": 0, 00:19:52.385 "r_mbytes_per_sec": 0, 00:19:52.385 "w_mbytes_per_sec": 0 00:19:52.385 }, 00:19:52.385 "claimed": false, 00:19:52.385 "zoned": false, 00:19:52.385 "supported_io_types": { 00:19:52.385 "read": true, 00:19:52.385 "write": true, 00:19:52.385 "unmap": true, 00:19:52.385 "flush": true, 00:19:52.385 "reset": true, 00:19:52.385 "nvme_admin": false, 00:19:52.385 "nvme_io": false, 00:19:52.385 "nvme_io_md": false, 00:19:52.385 "write_zeroes": true, 00:19:52.385 "zcopy": false, 00:19:52.385 "get_zone_info": false, 00:19:52.385 "zone_management": false, 00:19:52.385 "zone_append": false, 00:19:52.385 "compare": false, 00:19:52.385 "compare_and_write": false, 00:19:52.385 "abort": false, 00:19:52.385 "seek_hole": false, 00:19:52.385 "seek_data": false, 00:19:52.385 "copy": false, 00:19:52.385 "nvme_iov_md": false 00:19:52.385 }, 00:19:52.385 "memory_domains": [ 00:19:52.385 { 00:19:52.385 "dma_device_id": "system", 00:19:52.385 "dma_device_type": 1 00:19:52.385 }, 00:19:52.385 { 00:19:52.385 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:52.385 "dma_device_type": 2 00:19:52.385 }, 00:19:52.385 { 00:19:52.385 "dma_device_id": "system", 00:19:52.385 "dma_device_type": 1 00:19:52.385 }, 00:19:52.385 { 00:19:52.385 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:52.385 "dma_device_type": 2 00:19:52.385 }, 00:19:52.385 { 00:19:52.385 "dma_device_id": "system", 00:19:52.385 "dma_device_type": 1 00:19:52.385 }, 00:19:52.385 { 00:19:52.385 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:52.385 "dma_device_type": 2 00:19:52.385 }, 00:19:52.385 { 00:19:52.385 "dma_device_id": "system", 00:19:52.385 "dma_device_type": 1 00:19:52.385 }, 00:19:52.385 { 00:19:52.385 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:52.385 "dma_device_type": 2 00:19:52.385 } 00:19:52.385 ], 00:19:52.385 "driver_specific": { 00:19:52.385 "raid": { 00:19:52.385 "uuid": "d13087eb-a9fd-4d45-998c-a7b7b455deba", 00:19:52.385 "strip_size_kb": 64, 00:19:52.385 "state": "online", 00:19:52.385 "raid_level": "concat", 00:19:52.385 "superblock": false, 00:19:52.385 "num_base_bdevs": 4, 00:19:52.385 "num_base_bdevs_discovered": 4, 00:19:52.385 "num_base_bdevs_operational": 4, 00:19:52.385 "base_bdevs_list": [ 00:19:52.385 { 00:19:52.385 "name": "NewBaseBdev", 00:19:52.385 "uuid": "15b3a9f0-c4a7-4e94-9a29-c2ca91f82f3d", 00:19:52.385 "is_configured": true, 00:19:52.385 "data_offset": 0, 00:19:52.385 "data_size": 65536 00:19:52.385 }, 00:19:52.385 { 00:19:52.385 "name": "BaseBdev2", 00:19:52.385 "uuid": "a1ecaccc-b270-4f0f-8099-a5a4ffd9d05e", 00:19:52.385 "is_configured": true, 00:19:52.385 "data_offset": 0, 00:19:52.385 "data_size": 65536 00:19:52.385 }, 00:19:52.385 { 00:19:52.385 "name": "BaseBdev3", 00:19:52.385 "uuid": "3c3a272d-8184-4866-810c-8c512613a0d5", 00:19:52.385 "is_configured": true, 00:19:52.385 "data_offset": 0, 00:19:52.385 "data_size": 65536 00:19:52.385 }, 00:19:52.385 { 00:19:52.385 "name": "BaseBdev4", 00:19:52.385 "uuid": "28a269d6-c2ec-444b-ac2c-259984fcb510", 00:19:52.385 "is_configured": true, 00:19:52.385 "data_offset": 0, 00:19:52.385 "data_size": 65536 00:19:52.385 } 00:19:52.385 ] 00:19:52.385 } 00:19:52.385 } 00:19:52.385 }' 00:19:52.385 11:30:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:19:52.385 11:30:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='NewBaseBdev 00:19:52.385 BaseBdev2 00:19:52.385 BaseBdev3 00:19:52.385 BaseBdev4' 00:19:52.385 11:30:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:19:52.385 11:30:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev 00:19:52.385 11:30:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:19:52.699 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:19:52.699 "name": "NewBaseBdev", 00:19:52.699 "aliases": [ 00:19:52.699 "15b3a9f0-c4a7-4e94-9a29-c2ca91f82f3d" 00:19:52.699 ], 00:19:52.699 "product_name": "Malloc disk", 00:19:52.699 "block_size": 512, 00:19:52.699 "num_blocks": 65536, 00:19:52.699 "uuid": "15b3a9f0-c4a7-4e94-9a29-c2ca91f82f3d", 00:19:52.699 "assigned_rate_limits": { 00:19:52.699 "rw_ios_per_sec": 0, 00:19:52.699 "rw_mbytes_per_sec": 0, 00:19:52.699 "r_mbytes_per_sec": 0, 00:19:52.699 "w_mbytes_per_sec": 0 00:19:52.699 }, 00:19:52.699 "claimed": true, 00:19:52.699 "claim_type": "exclusive_write", 00:19:52.699 "zoned": false, 00:19:52.699 "supported_io_types": { 00:19:52.699 "read": true, 00:19:52.699 "write": true, 00:19:52.699 "unmap": true, 00:19:52.699 "flush": true, 00:19:52.699 "reset": true, 00:19:52.699 "nvme_admin": false, 00:19:52.699 "nvme_io": false, 00:19:52.699 "nvme_io_md": false, 00:19:52.699 "write_zeroes": true, 00:19:52.699 "zcopy": true, 00:19:52.699 "get_zone_info": false, 00:19:52.699 "zone_management": false, 00:19:52.699 "zone_append": false, 00:19:52.699 "compare": false, 00:19:52.699 "compare_and_write": false, 00:19:52.699 "abort": true, 00:19:52.699 "seek_hole": false, 00:19:52.699 "seek_data": false, 00:19:52.699 "copy": true, 00:19:52.699 "nvme_iov_md": false 00:19:52.699 }, 00:19:52.699 "memory_domains": [ 00:19:52.699 { 00:19:52.699 "dma_device_id": "system", 00:19:52.699 "dma_device_type": 1 00:19:52.699 }, 00:19:52.699 { 00:19:52.699 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:52.699 "dma_device_type": 2 00:19:52.699 } 00:19:52.699 ], 00:19:52.699 "driver_specific": {} 00:19:52.699 }' 00:19:52.699 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:52.699 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:52.699 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:19:52.699 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:52.956 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:52.956 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:19:52.957 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:52.957 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:52.957 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:19:52.957 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:52.957 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:52.957 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:19:52.957 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:19:52.957 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:19:52.957 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:19:53.214 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:19:53.214 "name": "BaseBdev2", 00:19:53.214 "aliases": [ 00:19:53.214 "a1ecaccc-b270-4f0f-8099-a5a4ffd9d05e" 00:19:53.214 ], 00:19:53.214 "product_name": "Malloc disk", 00:19:53.214 "block_size": 512, 00:19:53.214 "num_blocks": 65536, 00:19:53.214 "uuid": "a1ecaccc-b270-4f0f-8099-a5a4ffd9d05e", 00:19:53.214 "assigned_rate_limits": { 00:19:53.214 "rw_ios_per_sec": 0, 00:19:53.214 "rw_mbytes_per_sec": 0, 00:19:53.214 "r_mbytes_per_sec": 0, 00:19:53.214 "w_mbytes_per_sec": 0 00:19:53.214 }, 00:19:53.214 "claimed": true, 00:19:53.214 "claim_type": "exclusive_write", 00:19:53.214 "zoned": false, 00:19:53.214 "supported_io_types": { 00:19:53.215 "read": true, 00:19:53.215 "write": true, 00:19:53.215 "unmap": true, 00:19:53.215 "flush": true, 00:19:53.215 "reset": true, 00:19:53.215 "nvme_admin": false, 00:19:53.215 "nvme_io": false, 00:19:53.215 "nvme_io_md": false, 00:19:53.215 "write_zeroes": true, 00:19:53.215 "zcopy": true, 00:19:53.215 "get_zone_info": false, 00:19:53.215 "zone_management": false, 00:19:53.215 "zone_append": false, 00:19:53.215 "compare": false, 00:19:53.215 "compare_and_write": false, 00:19:53.215 "abort": true, 00:19:53.215 "seek_hole": false, 00:19:53.215 "seek_data": false, 00:19:53.215 "copy": true, 00:19:53.215 "nvme_iov_md": false 00:19:53.215 }, 00:19:53.215 "memory_domains": [ 00:19:53.215 { 00:19:53.215 "dma_device_id": "system", 00:19:53.215 "dma_device_type": 1 00:19:53.215 }, 00:19:53.215 { 00:19:53.215 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:53.215 "dma_device_type": 2 00:19:53.215 } 00:19:53.215 ], 00:19:53.215 "driver_specific": {} 00:19:53.215 }' 00:19:53.215 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:53.473 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:53.473 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:19:53.473 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:53.473 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:53.473 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:19:53.473 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:53.473 11:30:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:53.473 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:19:53.473 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:53.473 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:53.731 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:19:53.731 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:19:53.731 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:19:53.731 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:19:53.731 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:19:53.731 "name": "BaseBdev3", 00:19:53.731 "aliases": [ 00:19:53.731 "3c3a272d-8184-4866-810c-8c512613a0d5" 00:19:53.731 ], 00:19:53.731 "product_name": "Malloc disk", 00:19:53.731 "block_size": 512, 00:19:53.731 "num_blocks": 65536, 00:19:53.731 "uuid": "3c3a272d-8184-4866-810c-8c512613a0d5", 00:19:53.731 "assigned_rate_limits": { 00:19:53.731 "rw_ios_per_sec": 0, 00:19:53.731 "rw_mbytes_per_sec": 0, 00:19:53.731 "r_mbytes_per_sec": 0, 00:19:53.731 "w_mbytes_per_sec": 0 00:19:53.731 }, 00:19:53.731 "claimed": true, 00:19:53.731 "claim_type": "exclusive_write", 00:19:53.731 "zoned": false, 00:19:53.731 "supported_io_types": { 00:19:53.731 "read": true, 00:19:53.731 "write": true, 00:19:53.731 "unmap": true, 00:19:53.731 "flush": true, 00:19:53.731 "reset": true, 00:19:53.731 "nvme_admin": false, 00:19:53.731 "nvme_io": false, 00:19:53.731 "nvme_io_md": false, 00:19:53.731 "write_zeroes": true, 00:19:53.731 "zcopy": true, 00:19:53.731 "get_zone_info": false, 00:19:53.731 "zone_management": false, 00:19:53.731 "zone_append": false, 00:19:53.731 "compare": false, 00:19:53.731 "compare_and_write": false, 00:19:53.731 "abort": true, 00:19:53.731 "seek_hole": false, 00:19:53.731 "seek_data": false, 00:19:53.731 "copy": true, 00:19:53.731 "nvme_iov_md": false 00:19:53.731 }, 00:19:53.731 "memory_domains": [ 00:19:53.731 { 00:19:53.731 "dma_device_id": "system", 00:19:53.731 "dma_device_type": 1 00:19:53.731 }, 00:19:53.731 { 00:19:53.731 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:53.731 "dma_device_type": 2 00:19:53.731 } 00:19:53.731 ], 00:19:53.731 "driver_specific": {} 00:19:53.731 }' 00:19:53.731 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:53.731 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:53.989 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:19:53.989 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:53.989 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:53.989 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:19:53.989 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:53.989 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:53.989 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:19:53.989 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:53.989 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:54.248 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:19:54.248 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:19:54.248 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 00:19:54.248 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:19:54.507 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:19:54.507 "name": "BaseBdev4", 00:19:54.507 "aliases": [ 00:19:54.507 "28a269d6-c2ec-444b-ac2c-259984fcb510" 00:19:54.507 ], 00:19:54.507 "product_name": "Malloc disk", 00:19:54.507 "block_size": 512, 00:19:54.507 "num_blocks": 65536, 00:19:54.507 "uuid": "28a269d6-c2ec-444b-ac2c-259984fcb510", 00:19:54.507 "assigned_rate_limits": { 00:19:54.507 "rw_ios_per_sec": 0, 00:19:54.507 "rw_mbytes_per_sec": 0, 00:19:54.507 "r_mbytes_per_sec": 0, 00:19:54.507 "w_mbytes_per_sec": 0 00:19:54.507 }, 00:19:54.507 "claimed": true, 00:19:54.507 "claim_type": "exclusive_write", 00:19:54.507 "zoned": false, 00:19:54.507 "supported_io_types": { 00:19:54.507 "read": true, 00:19:54.507 "write": true, 00:19:54.507 "unmap": true, 00:19:54.507 "flush": true, 00:19:54.507 "reset": true, 00:19:54.507 "nvme_admin": false, 00:19:54.507 "nvme_io": false, 00:19:54.507 "nvme_io_md": false, 00:19:54.507 "write_zeroes": true, 00:19:54.507 "zcopy": true, 00:19:54.507 "get_zone_info": false, 00:19:54.507 "zone_management": false, 00:19:54.507 "zone_append": false, 00:19:54.507 "compare": false, 00:19:54.507 "compare_and_write": false, 00:19:54.507 "abort": true, 00:19:54.507 "seek_hole": false, 00:19:54.507 "seek_data": false, 00:19:54.507 "copy": true, 00:19:54.507 "nvme_iov_md": false 00:19:54.507 }, 00:19:54.507 "memory_domains": [ 00:19:54.507 { 00:19:54.507 "dma_device_id": "system", 00:19:54.507 "dma_device_type": 1 00:19:54.507 }, 00:19:54.507 { 00:19:54.507 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:54.507 "dma_device_type": 2 00:19:54.507 } 00:19:54.507 ], 00:19:54.507 "driver_specific": {} 00:19:54.507 }' 00:19:54.507 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:54.507 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:19:54.507 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:19:54.507 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:54.507 11:30:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:19:54.507 11:30:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:19:54.507 11:30:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:54.507 11:30:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:19:54.507 11:30:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:19:54.765 11:30:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:54.765 11:30:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:19:54.765 11:30:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:19:54.765 11:30:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@338 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:19:55.024 [2024-07-15 11:30:38.394250] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:19:55.024 [2024-07-15 11:30:38.394277] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:19:55.024 [2024-07-15 11:30:38.394331] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:19:55.024 [2024-07-15 11:30:38.394389] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:19:55.024 [2024-07-15 11:30:38.394401] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x132e040 name Existed_Raid, state offline 00:19:55.024 11:30:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@341 -- # killprocess 937799 00:19:55.024 11:30:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@948 -- # '[' -z 937799 ']' 00:19:55.024 11:30:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@952 -- # kill -0 937799 00:19:55.024 11:30:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@953 -- # uname 00:19:55.024 11:30:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:19:55.024 11:30:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 937799 00:19:55.024 11:30:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:19:55.024 11:30:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:19:55.024 11:30:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 937799' 00:19:55.024 killing process with pid 937799 00:19:55.024 11:30:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@967 -- # kill 937799 00:19:55.024 [2024-07-15 11:30:38.466302] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:19:55.024 11:30:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # wait 937799 00:19:55.024 [2024-07-15 11:30:38.503974] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@343 -- # return 0 00:19:55.283 00:19:55.283 real 0m32.004s 00:19:55.283 user 0m58.716s 00:19:55.283 sys 0m5.714s 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:55.283 ************************************ 00:19:55.283 END TEST raid_state_function_test 00:19:55.283 ************************************ 00:19:55.283 11:30:38 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:19:55.283 11:30:38 bdev_raid -- bdev/bdev_raid.sh@868 -- # run_test raid_state_function_test_sb raid_state_function_test concat 4 true 00:19:55.283 11:30:38 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:19:55.283 11:30:38 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:19:55.283 11:30:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:19:55.283 ************************************ 00:19:55.283 START TEST raid_state_function_test_sb 00:19:55.283 ************************************ 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1123 -- # raid_state_function_test concat 4 true 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@220 -- # local raid_level=concat 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=4 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # local superblock=true 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev3 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev4 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # local strip_size 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # '[' concat '!=' raid1 ']' 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # strip_size=64 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@232 -- # strip_size_create_arg='-z 64' 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # '[' true = true ']' 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@238 -- # superblock_create_arg=-s 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # raid_pid=942599 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 942599' 00:19:55.283 Process raid pid: 942599 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:19:55.283 11:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@246 -- # waitforlisten 942599 /var/tmp/spdk-raid.sock 00:19:55.284 11:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@829 -- # '[' -z 942599 ']' 00:19:55.284 11:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:19:55.284 11:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@834 -- # local max_retries=100 00:19:55.284 11:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:19:55.284 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:19:55.284 11:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # xtrace_disable 00:19:55.284 11:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:55.542 [2024-07-15 11:30:38.891122] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:19:55.542 [2024-07-15 11:30:38.891188] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:55.542 [2024-07-15 11:30:39.021083] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:55.542 [2024-07-15 11:30:39.128535] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:19:55.799 [2024-07-15 11:30:39.197017] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:55.799 [2024-07-15 11:30:39.197052] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:56.363 11:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:19:56.363 11:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@862 -- # return 0 00:19:56.363 11:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r concat -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:19:56.621 [2024-07-15 11:30:40.036520] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:19:56.621 [2024-07-15 11:30:40.036563] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:19:56.621 [2024-07-15 11:30:40.036576] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:19:56.621 [2024-07-15 11:30:40.036588] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:19:56.621 [2024-07-15 11:30:40.036597] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:19:56.621 [2024-07-15 11:30:40.036610] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:19:56.621 [2024-07-15 11:30:40.036619] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:19:56.621 [2024-07-15 11:30:40.036630] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:19:56.621 11:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:19:56.621 11:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:19:56.621 11:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:19:56.621 11:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:19:56.621 11:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:56.621 11:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:56.621 11:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:56.621 11:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:56.621 11:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:56.621 11:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:56.621 11:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:56.621 11:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:56.878 11:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:56.878 "name": "Existed_Raid", 00:19:56.878 "uuid": "5b4d1b03-fdaf-4fa1-b8c9-8f600de80cdb", 00:19:56.878 "strip_size_kb": 64, 00:19:56.878 "state": "configuring", 00:19:56.878 "raid_level": "concat", 00:19:56.878 "superblock": true, 00:19:56.878 "num_base_bdevs": 4, 00:19:56.878 "num_base_bdevs_discovered": 0, 00:19:56.878 "num_base_bdevs_operational": 4, 00:19:56.878 "base_bdevs_list": [ 00:19:56.878 { 00:19:56.878 "name": "BaseBdev1", 00:19:56.878 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:56.878 "is_configured": false, 00:19:56.878 "data_offset": 0, 00:19:56.878 "data_size": 0 00:19:56.878 }, 00:19:56.878 { 00:19:56.878 "name": "BaseBdev2", 00:19:56.878 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:56.878 "is_configured": false, 00:19:56.878 "data_offset": 0, 00:19:56.878 "data_size": 0 00:19:56.878 }, 00:19:56.878 { 00:19:56.878 "name": "BaseBdev3", 00:19:56.878 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:56.878 "is_configured": false, 00:19:56.878 "data_offset": 0, 00:19:56.878 "data_size": 0 00:19:56.878 }, 00:19:56.878 { 00:19:56.878 "name": "BaseBdev4", 00:19:56.878 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:56.878 "is_configured": false, 00:19:56.878 "data_offset": 0, 00:19:56.878 "data_size": 0 00:19:56.878 } 00:19:56.878 ] 00:19:56.878 }' 00:19:56.878 11:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:56.878 11:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:57.440 11:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:19:57.699 [2024-07-15 11:30:41.039026] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:19:57.699 [2024-07-15 11:30:41.039054] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xcf1aa0 name Existed_Raid, state configuring 00:19:57.699 11:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r concat -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:19:57.699 [2024-07-15 11:30:41.283696] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:19:57.699 [2024-07-15 11:30:41.283721] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:19:57.699 [2024-07-15 11:30:41.283731] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:19:57.699 [2024-07-15 11:30:41.283742] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:19:57.699 [2024-07-15 11:30:41.283751] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:19:57.699 [2024-07-15 11:30:41.283762] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:19:57.699 [2024-07-15 11:30:41.283770] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:19:57.699 [2024-07-15 11:30:41.283781] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:19:57.956 11:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:19:57.957 [2024-07-15 11:30:41.538107] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:19:57.957 BaseBdev1 00:19:58.214 11:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:19:58.214 11:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:19:58.214 11:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:19:58.214 11:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:19:58.214 11:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:19:58.214 11:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:19:58.214 11:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:19:58.214 11:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:19:58.472 [ 00:19:58.472 { 00:19:58.472 "name": "BaseBdev1", 00:19:58.472 "aliases": [ 00:19:58.472 "4fc773d7-0611-4b07-bfe5-bd437438a12d" 00:19:58.472 ], 00:19:58.472 "product_name": "Malloc disk", 00:19:58.472 "block_size": 512, 00:19:58.472 "num_blocks": 65536, 00:19:58.472 "uuid": "4fc773d7-0611-4b07-bfe5-bd437438a12d", 00:19:58.472 "assigned_rate_limits": { 00:19:58.472 "rw_ios_per_sec": 0, 00:19:58.472 "rw_mbytes_per_sec": 0, 00:19:58.472 "r_mbytes_per_sec": 0, 00:19:58.472 "w_mbytes_per_sec": 0 00:19:58.472 }, 00:19:58.472 "claimed": true, 00:19:58.472 "claim_type": "exclusive_write", 00:19:58.472 "zoned": false, 00:19:58.472 "supported_io_types": { 00:19:58.472 "read": true, 00:19:58.472 "write": true, 00:19:58.472 "unmap": true, 00:19:58.472 "flush": true, 00:19:58.472 "reset": true, 00:19:58.472 "nvme_admin": false, 00:19:58.472 "nvme_io": false, 00:19:58.472 "nvme_io_md": false, 00:19:58.472 "write_zeroes": true, 00:19:58.472 "zcopy": true, 00:19:58.472 "get_zone_info": false, 00:19:58.472 "zone_management": false, 00:19:58.472 "zone_append": false, 00:19:58.472 "compare": false, 00:19:58.472 "compare_and_write": false, 00:19:58.472 "abort": true, 00:19:58.472 "seek_hole": false, 00:19:58.472 "seek_data": false, 00:19:58.472 "copy": true, 00:19:58.472 "nvme_iov_md": false 00:19:58.472 }, 00:19:58.472 "memory_domains": [ 00:19:58.472 { 00:19:58.472 "dma_device_id": "system", 00:19:58.472 "dma_device_type": 1 00:19:58.472 }, 00:19:58.472 { 00:19:58.472 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:58.472 "dma_device_type": 2 00:19:58.472 } 00:19:58.472 ], 00:19:58.472 "driver_specific": {} 00:19:58.472 } 00:19:58.472 ] 00:19:58.472 11:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:19:58.472 11:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:19:58.472 11:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:19:58.472 11:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:19:58.472 11:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:19:58.472 11:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:58.472 11:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:58.472 11:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:58.472 11:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:58.472 11:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:58.472 11:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:58.472 11:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:58.472 11:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:58.730 11:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:19:58.730 "name": "Existed_Raid", 00:19:58.730 "uuid": "810392ce-e034-4c64-84d9-c704f5e4ea0f", 00:19:58.730 "strip_size_kb": 64, 00:19:58.730 "state": "configuring", 00:19:58.730 "raid_level": "concat", 00:19:58.730 "superblock": true, 00:19:58.730 "num_base_bdevs": 4, 00:19:58.730 "num_base_bdevs_discovered": 1, 00:19:58.730 "num_base_bdevs_operational": 4, 00:19:58.730 "base_bdevs_list": [ 00:19:58.730 { 00:19:58.730 "name": "BaseBdev1", 00:19:58.730 "uuid": "4fc773d7-0611-4b07-bfe5-bd437438a12d", 00:19:58.730 "is_configured": true, 00:19:58.730 "data_offset": 2048, 00:19:58.730 "data_size": 63488 00:19:58.730 }, 00:19:58.730 { 00:19:58.730 "name": "BaseBdev2", 00:19:58.730 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:58.730 "is_configured": false, 00:19:58.730 "data_offset": 0, 00:19:58.730 "data_size": 0 00:19:58.730 }, 00:19:58.730 { 00:19:58.730 "name": "BaseBdev3", 00:19:58.730 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:58.730 "is_configured": false, 00:19:58.730 "data_offset": 0, 00:19:58.730 "data_size": 0 00:19:58.730 }, 00:19:58.730 { 00:19:58.730 "name": "BaseBdev4", 00:19:58.730 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:58.730 "is_configured": false, 00:19:58.730 "data_offset": 0, 00:19:58.730 "data_size": 0 00:19:58.730 } 00:19:58.730 ] 00:19:58.730 }' 00:19:58.730 11:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:19:58.730 11:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:59.296 11:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:19:59.555 [2024-07-15 11:30:43.074183] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:19:59.555 [2024-07-15 11:30:43.074219] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xcf1310 name Existed_Raid, state configuring 00:19:59.555 11:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r concat -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:19:59.813 [2024-07-15 11:30:43.318876] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:19:59.813 [2024-07-15 11:30:43.320311] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:19:59.813 [2024-07-15 11:30:43.320342] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:19:59.813 [2024-07-15 11:30:43.320353] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:19:59.813 [2024-07-15 11:30:43.320365] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:19:59.813 [2024-07-15 11:30:43.320374] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:19:59.813 [2024-07-15 11:30:43.320385] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:19:59.813 11:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:19:59.813 11:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:19:59.813 11:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:19:59.813 11:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:19:59.813 11:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:19:59.813 11:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:19:59.813 11:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:19:59.813 11:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:19:59.813 11:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:19:59.813 11:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:19:59.813 11:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:19:59.813 11:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:19:59.813 11:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:19:59.813 11:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:00.070 11:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:00.070 "name": "Existed_Raid", 00:20:00.070 "uuid": "74ac41db-f972-47ef-aa08-7dfed434ef55", 00:20:00.070 "strip_size_kb": 64, 00:20:00.070 "state": "configuring", 00:20:00.070 "raid_level": "concat", 00:20:00.070 "superblock": true, 00:20:00.070 "num_base_bdevs": 4, 00:20:00.070 "num_base_bdevs_discovered": 1, 00:20:00.070 "num_base_bdevs_operational": 4, 00:20:00.070 "base_bdevs_list": [ 00:20:00.070 { 00:20:00.070 "name": "BaseBdev1", 00:20:00.070 "uuid": "4fc773d7-0611-4b07-bfe5-bd437438a12d", 00:20:00.070 "is_configured": true, 00:20:00.070 "data_offset": 2048, 00:20:00.070 "data_size": 63488 00:20:00.070 }, 00:20:00.070 { 00:20:00.070 "name": "BaseBdev2", 00:20:00.070 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:00.070 "is_configured": false, 00:20:00.070 "data_offset": 0, 00:20:00.070 "data_size": 0 00:20:00.070 }, 00:20:00.070 { 00:20:00.070 "name": "BaseBdev3", 00:20:00.070 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:00.070 "is_configured": false, 00:20:00.070 "data_offset": 0, 00:20:00.070 "data_size": 0 00:20:00.070 }, 00:20:00.070 { 00:20:00.070 "name": "BaseBdev4", 00:20:00.070 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:00.070 "is_configured": false, 00:20:00.070 "data_offset": 0, 00:20:00.071 "data_size": 0 00:20:00.071 } 00:20:00.071 ] 00:20:00.071 }' 00:20:00.071 11:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:00.071 11:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:00.636 11:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:20:00.895 [2024-07-15 11:30:44.405261] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:20:00.895 BaseBdev2 00:20:00.895 11:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:20:00.895 11:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:20:00.895 11:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:20:00.895 11:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:20:00.895 11:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:20:00.895 11:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:20:00.895 11:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:20:01.153 11:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:20:01.412 [ 00:20:01.412 { 00:20:01.412 "name": "BaseBdev2", 00:20:01.412 "aliases": [ 00:20:01.412 "cb6d9024-3bcd-494c-a8cd-5058e57e7486" 00:20:01.412 ], 00:20:01.412 "product_name": "Malloc disk", 00:20:01.412 "block_size": 512, 00:20:01.412 "num_blocks": 65536, 00:20:01.412 "uuid": "cb6d9024-3bcd-494c-a8cd-5058e57e7486", 00:20:01.412 "assigned_rate_limits": { 00:20:01.412 "rw_ios_per_sec": 0, 00:20:01.412 "rw_mbytes_per_sec": 0, 00:20:01.412 "r_mbytes_per_sec": 0, 00:20:01.412 "w_mbytes_per_sec": 0 00:20:01.412 }, 00:20:01.412 "claimed": true, 00:20:01.412 "claim_type": "exclusive_write", 00:20:01.412 "zoned": false, 00:20:01.412 "supported_io_types": { 00:20:01.412 "read": true, 00:20:01.412 "write": true, 00:20:01.412 "unmap": true, 00:20:01.412 "flush": true, 00:20:01.412 "reset": true, 00:20:01.412 "nvme_admin": false, 00:20:01.412 "nvme_io": false, 00:20:01.412 "nvme_io_md": false, 00:20:01.412 "write_zeroes": true, 00:20:01.412 "zcopy": true, 00:20:01.412 "get_zone_info": false, 00:20:01.412 "zone_management": false, 00:20:01.412 "zone_append": false, 00:20:01.412 "compare": false, 00:20:01.412 "compare_and_write": false, 00:20:01.412 "abort": true, 00:20:01.412 "seek_hole": false, 00:20:01.412 "seek_data": false, 00:20:01.412 "copy": true, 00:20:01.412 "nvme_iov_md": false 00:20:01.412 }, 00:20:01.412 "memory_domains": [ 00:20:01.412 { 00:20:01.412 "dma_device_id": "system", 00:20:01.412 "dma_device_type": 1 00:20:01.412 }, 00:20:01.412 { 00:20:01.412 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:01.412 "dma_device_type": 2 00:20:01.412 } 00:20:01.412 ], 00:20:01.412 "driver_specific": {} 00:20:01.412 } 00:20:01.412 ] 00:20:01.412 11:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:20:01.412 11:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:20:01.412 11:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:20:01.412 11:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:20:01.412 11:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:20:01.412 11:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:20:01.412 11:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:20:01.412 11:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:20:01.412 11:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:20:01.412 11:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:20:01.412 11:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:20:01.412 11:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:20:01.412 11:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:20:01.412 11:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:01.413 11:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:01.671 11:30:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:01.671 "name": "Existed_Raid", 00:20:01.671 "uuid": "74ac41db-f972-47ef-aa08-7dfed434ef55", 00:20:01.671 "strip_size_kb": 64, 00:20:01.671 "state": "configuring", 00:20:01.671 "raid_level": "concat", 00:20:01.671 "superblock": true, 00:20:01.671 "num_base_bdevs": 4, 00:20:01.671 "num_base_bdevs_discovered": 2, 00:20:01.671 "num_base_bdevs_operational": 4, 00:20:01.671 "base_bdevs_list": [ 00:20:01.671 { 00:20:01.671 "name": "BaseBdev1", 00:20:01.671 "uuid": "4fc773d7-0611-4b07-bfe5-bd437438a12d", 00:20:01.671 "is_configured": true, 00:20:01.671 "data_offset": 2048, 00:20:01.671 "data_size": 63488 00:20:01.671 }, 00:20:01.671 { 00:20:01.671 "name": "BaseBdev2", 00:20:01.671 "uuid": "cb6d9024-3bcd-494c-a8cd-5058e57e7486", 00:20:01.671 "is_configured": true, 00:20:01.671 "data_offset": 2048, 00:20:01.671 "data_size": 63488 00:20:01.671 }, 00:20:01.671 { 00:20:01.671 "name": "BaseBdev3", 00:20:01.671 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:01.671 "is_configured": false, 00:20:01.671 "data_offset": 0, 00:20:01.671 "data_size": 0 00:20:01.671 }, 00:20:01.671 { 00:20:01.671 "name": "BaseBdev4", 00:20:01.671 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:01.671 "is_configured": false, 00:20:01.671 "data_offset": 0, 00:20:01.671 "data_size": 0 00:20:01.671 } 00:20:01.671 ] 00:20:01.671 }' 00:20:01.671 11:30:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:01.671 11:30:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:02.237 11:30:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:20:02.495 [2024-07-15 11:30:46.000968] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:20:02.495 BaseBdev3 00:20:02.495 11:30:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev3 00:20:02.495 11:30:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:20:02.495 11:30:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:20:02.495 11:30:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:20:02.495 11:30:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:20:02.495 11:30:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:20:02.495 11:30:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:20:02.754 11:30:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:20:03.013 [ 00:20:03.013 { 00:20:03.013 "name": "BaseBdev3", 00:20:03.013 "aliases": [ 00:20:03.013 "5ced5b85-39b5-4152-ab0e-adb93705a2f7" 00:20:03.013 ], 00:20:03.013 "product_name": "Malloc disk", 00:20:03.013 "block_size": 512, 00:20:03.013 "num_blocks": 65536, 00:20:03.013 "uuid": "5ced5b85-39b5-4152-ab0e-adb93705a2f7", 00:20:03.013 "assigned_rate_limits": { 00:20:03.013 "rw_ios_per_sec": 0, 00:20:03.013 "rw_mbytes_per_sec": 0, 00:20:03.013 "r_mbytes_per_sec": 0, 00:20:03.013 "w_mbytes_per_sec": 0 00:20:03.013 }, 00:20:03.013 "claimed": true, 00:20:03.013 "claim_type": "exclusive_write", 00:20:03.013 "zoned": false, 00:20:03.013 "supported_io_types": { 00:20:03.013 "read": true, 00:20:03.013 "write": true, 00:20:03.013 "unmap": true, 00:20:03.013 "flush": true, 00:20:03.013 "reset": true, 00:20:03.013 "nvme_admin": false, 00:20:03.013 "nvme_io": false, 00:20:03.013 "nvme_io_md": false, 00:20:03.013 "write_zeroes": true, 00:20:03.013 "zcopy": true, 00:20:03.013 "get_zone_info": false, 00:20:03.013 "zone_management": false, 00:20:03.013 "zone_append": false, 00:20:03.013 "compare": false, 00:20:03.013 "compare_and_write": false, 00:20:03.013 "abort": true, 00:20:03.013 "seek_hole": false, 00:20:03.013 "seek_data": false, 00:20:03.013 "copy": true, 00:20:03.013 "nvme_iov_md": false 00:20:03.013 }, 00:20:03.013 "memory_domains": [ 00:20:03.013 { 00:20:03.013 "dma_device_id": "system", 00:20:03.013 "dma_device_type": 1 00:20:03.013 }, 00:20:03.013 { 00:20:03.013 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:03.013 "dma_device_type": 2 00:20:03.013 } 00:20:03.013 ], 00:20:03.013 "driver_specific": {} 00:20:03.013 } 00:20:03.013 ] 00:20:03.013 11:30:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:20:03.013 11:30:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:20:03.013 11:30:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:20:03.013 11:30:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:20:03.013 11:30:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:20:03.013 11:30:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:20:03.013 11:30:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:20:03.013 11:30:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:20:03.013 11:30:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:20:03.013 11:30:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:20:03.013 11:30:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:20:03.013 11:30:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:20:03.013 11:30:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:20:03.013 11:30:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:03.013 11:30:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:03.271 11:30:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:03.271 "name": "Existed_Raid", 00:20:03.271 "uuid": "74ac41db-f972-47ef-aa08-7dfed434ef55", 00:20:03.271 "strip_size_kb": 64, 00:20:03.271 "state": "configuring", 00:20:03.271 "raid_level": "concat", 00:20:03.271 "superblock": true, 00:20:03.271 "num_base_bdevs": 4, 00:20:03.271 "num_base_bdevs_discovered": 3, 00:20:03.271 "num_base_bdevs_operational": 4, 00:20:03.271 "base_bdevs_list": [ 00:20:03.271 { 00:20:03.271 "name": "BaseBdev1", 00:20:03.271 "uuid": "4fc773d7-0611-4b07-bfe5-bd437438a12d", 00:20:03.271 "is_configured": true, 00:20:03.271 "data_offset": 2048, 00:20:03.271 "data_size": 63488 00:20:03.271 }, 00:20:03.271 { 00:20:03.271 "name": "BaseBdev2", 00:20:03.271 "uuid": "cb6d9024-3bcd-494c-a8cd-5058e57e7486", 00:20:03.271 "is_configured": true, 00:20:03.271 "data_offset": 2048, 00:20:03.271 "data_size": 63488 00:20:03.271 }, 00:20:03.271 { 00:20:03.271 "name": "BaseBdev3", 00:20:03.271 "uuid": "5ced5b85-39b5-4152-ab0e-adb93705a2f7", 00:20:03.271 "is_configured": true, 00:20:03.271 "data_offset": 2048, 00:20:03.271 "data_size": 63488 00:20:03.271 }, 00:20:03.271 { 00:20:03.271 "name": "BaseBdev4", 00:20:03.271 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:03.271 "is_configured": false, 00:20:03.271 "data_offset": 0, 00:20:03.271 "data_size": 0 00:20:03.271 } 00:20:03.271 ] 00:20:03.271 }' 00:20:03.272 11:30:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:03.272 11:30:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:03.838 11:30:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4 00:20:04.097 [2024-07-15 11:30:47.588466] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:20:04.097 [2024-07-15 11:30:47.588632] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xcf2350 00:20:04.097 [2024-07-15 11:30:47.588645] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:20:04.097 [2024-07-15 11:30:47.588818] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xcf2020 00:20:04.097 [2024-07-15 11:30:47.588949] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xcf2350 00:20:04.097 [2024-07-15 11:30:47.588960] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0xcf2350 00:20:04.097 [2024-07-15 11:30:47.589051] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:04.097 BaseBdev4 00:20:04.097 11:30:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev4 00:20:04.097 11:30:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev4 00:20:04.097 11:30:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:20:04.097 11:30:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:20:04.097 11:30:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:20:04.097 11:30:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:20:04.097 11:30:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:20:04.355 11:30:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 -t 2000 00:20:04.613 [ 00:20:04.613 { 00:20:04.613 "name": "BaseBdev4", 00:20:04.613 "aliases": [ 00:20:04.613 "736ca049-221b-4e73-8172-b6eb0c46826f" 00:20:04.613 ], 00:20:04.613 "product_name": "Malloc disk", 00:20:04.613 "block_size": 512, 00:20:04.613 "num_blocks": 65536, 00:20:04.613 "uuid": "736ca049-221b-4e73-8172-b6eb0c46826f", 00:20:04.613 "assigned_rate_limits": { 00:20:04.613 "rw_ios_per_sec": 0, 00:20:04.613 "rw_mbytes_per_sec": 0, 00:20:04.613 "r_mbytes_per_sec": 0, 00:20:04.613 "w_mbytes_per_sec": 0 00:20:04.613 }, 00:20:04.613 "claimed": true, 00:20:04.613 "claim_type": "exclusive_write", 00:20:04.613 "zoned": false, 00:20:04.613 "supported_io_types": { 00:20:04.613 "read": true, 00:20:04.613 "write": true, 00:20:04.613 "unmap": true, 00:20:04.613 "flush": true, 00:20:04.613 "reset": true, 00:20:04.613 "nvme_admin": false, 00:20:04.613 "nvme_io": false, 00:20:04.613 "nvme_io_md": false, 00:20:04.613 "write_zeroes": true, 00:20:04.613 "zcopy": true, 00:20:04.613 "get_zone_info": false, 00:20:04.613 "zone_management": false, 00:20:04.613 "zone_append": false, 00:20:04.613 "compare": false, 00:20:04.613 "compare_and_write": false, 00:20:04.613 "abort": true, 00:20:04.613 "seek_hole": false, 00:20:04.613 "seek_data": false, 00:20:04.613 "copy": true, 00:20:04.613 "nvme_iov_md": false 00:20:04.613 }, 00:20:04.613 "memory_domains": [ 00:20:04.613 { 00:20:04.613 "dma_device_id": "system", 00:20:04.613 "dma_device_type": 1 00:20:04.613 }, 00:20:04.613 { 00:20:04.613 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:04.613 "dma_device_type": 2 00:20:04.613 } 00:20:04.613 ], 00:20:04.613 "driver_specific": {} 00:20:04.613 } 00:20:04.613 ] 00:20:04.613 11:30:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:20:04.613 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:20:04.613 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:20:04.613 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:20:04.613 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:20:04.613 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:20:04.613 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:20:04.613 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:20:04.613 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:20:04.613 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:20:04.613 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:20:04.613 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:20:04.613 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:20:04.613 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:04.613 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:04.872 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:04.872 "name": "Existed_Raid", 00:20:04.872 "uuid": "74ac41db-f972-47ef-aa08-7dfed434ef55", 00:20:04.872 "strip_size_kb": 64, 00:20:04.872 "state": "online", 00:20:04.872 "raid_level": "concat", 00:20:04.872 "superblock": true, 00:20:04.872 "num_base_bdevs": 4, 00:20:04.872 "num_base_bdevs_discovered": 4, 00:20:04.872 "num_base_bdevs_operational": 4, 00:20:04.872 "base_bdevs_list": [ 00:20:04.872 { 00:20:04.872 "name": "BaseBdev1", 00:20:04.872 "uuid": "4fc773d7-0611-4b07-bfe5-bd437438a12d", 00:20:04.872 "is_configured": true, 00:20:04.872 "data_offset": 2048, 00:20:04.872 "data_size": 63488 00:20:04.872 }, 00:20:04.872 { 00:20:04.872 "name": "BaseBdev2", 00:20:04.872 "uuid": "cb6d9024-3bcd-494c-a8cd-5058e57e7486", 00:20:04.872 "is_configured": true, 00:20:04.872 "data_offset": 2048, 00:20:04.872 "data_size": 63488 00:20:04.872 }, 00:20:04.872 { 00:20:04.872 "name": "BaseBdev3", 00:20:04.872 "uuid": "5ced5b85-39b5-4152-ab0e-adb93705a2f7", 00:20:04.872 "is_configured": true, 00:20:04.872 "data_offset": 2048, 00:20:04.872 "data_size": 63488 00:20:04.872 }, 00:20:04.872 { 00:20:04.872 "name": "BaseBdev4", 00:20:04.872 "uuid": "736ca049-221b-4e73-8172-b6eb0c46826f", 00:20:04.872 "is_configured": true, 00:20:04.872 "data_offset": 2048, 00:20:04.872 "data_size": 63488 00:20:04.872 } 00:20:04.872 ] 00:20:04.872 }' 00:20:04.872 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:04.872 11:30:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:05.438 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:20:05.438 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:20:05.438 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:20:05.438 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:20:05.438 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:20:05.438 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # local name 00:20:05.438 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:20:05.438 11:30:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:20:05.697 [2024-07-15 11:30:49.164975] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:20:05.697 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:20:05.697 "name": "Existed_Raid", 00:20:05.697 "aliases": [ 00:20:05.697 "74ac41db-f972-47ef-aa08-7dfed434ef55" 00:20:05.697 ], 00:20:05.697 "product_name": "Raid Volume", 00:20:05.697 "block_size": 512, 00:20:05.697 "num_blocks": 253952, 00:20:05.697 "uuid": "74ac41db-f972-47ef-aa08-7dfed434ef55", 00:20:05.697 "assigned_rate_limits": { 00:20:05.697 "rw_ios_per_sec": 0, 00:20:05.697 "rw_mbytes_per_sec": 0, 00:20:05.697 "r_mbytes_per_sec": 0, 00:20:05.697 "w_mbytes_per_sec": 0 00:20:05.697 }, 00:20:05.697 "claimed": false, 00:20:05.697 "zoned": false, 00:20:05.697 "supported_io_types": { 00:20:05.697 "read": true, 00:20:05.697 "write": true, 00:20:05.697 "unmap": true, 00:20:05.697 "flush": true, 00:20:05.697 "reset": true, 00:20:05.697 "nvme_admin": false, 00:20:05.697 "nvme_io": false, 00:20:05.697 "nvme_io_md": false, 00:20:05.697 "write_zeroes": true, 00:20:05.697 "zcopy": false, 00:20:05.697 "get_zone_info": false, 00:20:05.697 "zone_management": false, 00:20:05.697 "zone_append": false, 00:20:05.697 "compare": false, 00:20:05.697 "compare_and_write": false, 00:20:05.697 "abort": false, 00:20:05.697 "seek_hole": false, 00:20:05.697 "seek_data": false, 00:20:05.697 "copy": false, 00:20:05.697 "nvme_iov_md": false 00:20:05.697 }, 00:20:05.697 "memory_domains": [ 00:20:05.697 { 00:20:05.697 "dma_device_id": "system", 00:20:05.697 "dma_device_type": 1 00:20:05.697 }, 00:20:05.697 { 00:20:05.697 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:05.697 "dma_device_type": 2 00:20:05.697 }, 00:20:05.697 { 00:20:05.697 "dma_device_id": "system", 00:20:05.697 "dma_device_type": 1 00:20:05.697 }, 00:20:05.697 { 00:20:05.697 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:05.697 "dma_device_type": 2 00:20:05.697 }, 00:20:05.697 { 00:20:05.697 "dma_device_id": "system", 00:20:05.697 "dma_device_type": 1 00:20:05.697 }, 00:20:05.697 { 00:20:05.697 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:05.697 "dma_device_type": 2 00:20:05.697 }, 00:20:05.697 { 00:20:05.697 "dma_device_id": "system", 00:20:05.697 "dma_device_type": 1 00:20:05.697 }, 00:20:05.697 { 00:20:05.697 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:05.697 "dma_device_type": 2 00:20:05.697 } 00:20:05.697 ], 00:20:05.697 "driver_specific": { 00:20:05.697 "raid": { 00:20:05.697 "uuid": "74ac41db-f972-47ef-aa08-7dfed434ef55", 00:20:05.697 "strip_size_kb": 64, 00:20:05.697 "state": "online", 00:20:05.697 "raid_level": "concat", 00:20:05.697 "superblock": true, 00:20:05.697 "num_base_bdevs": 4, 00:20:05.697 "num_base_bdevs_discovered": 4, 00:20:05.697 "num_base_bdevs_operational": 4, 00:20:05.697 "base_bdevs_list": [ 00:20:05.697 { 00:20:05.697 "name": "BaseBdev1", 00:20:05.697 "uuid": "4fc773d7-0611-4b07-bfe5-bd437438a12d", 00:20:05.697 "is_configured": true, 00:20:05.697 "data_offset": 2048, 00:20:05.697 "data_size": 63488 00:20:05.697 }, 00:20:05.697 { 00:20:05.697 "name": "BaseBdev2", 00:20:05.697 "uuid": "cb6d9024-3bcd-494c-a8cd-5058e57e7486", 00:20:05.697 "is_configured": true, 00:20:05.697 "data_offset": 2048, 00:20:05.697 "data_size": 63488 00:20:05.697 }, 00:20:05.697 { 00:20:05.697 "name": "BaseBdev3", 00:20:05.697 "uuid": "5ced5b85-39b5-4152-ab0e-adb93705a2f7", 00:20:05.697 "is_configured": true, 00:20:05.697 "data_offset": 2048, 00:20:05.697 "data_size": 63488 00:20:05.697 }, 00:20:05.697 { 00:20:05.697 "name": "BaseBdev4", 00:20:05.697 "uuid": "736ca049-221b-4e73-8172-b6eb0c46826f", 00:20:05.697 "is_configured": true, 00:20:05.697 "data_offset": 2048, 00:20:05.697 "data_size": 63488 00:20:05.697 } 00:20:05.697 ] 00:20:05.697 } 00:20:05.697 } 00:20:05.697 }' 00:20:05.697 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:20:05.697 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:20:05.697 BaseBdev2 00:20:05.697 BaseBdev3 00:20:05.697 BaseBdev4' 00:20:05.697 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:20:05.697 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:20:05.697 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:20:05.958 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:20:05.958 "name": "BaseBdev1", 00:20:05.958 "aliases": [ 00:20:05.958 "4fc773d7-0611-4b07-bfe5-bd437438a12d" 00:20:05.958 ], 00:20:05.958 "product_name": "Malloc disk", 00:20:05.958 "block_size": 512, 00:20:05.958 "num_blocks": 65536, 00:20:05.958 "uuid": "4fc773d7-0611-4b07-bfe5-bd437438a12d", 00:20:05.958 "assigned_rate_limits": { 00:20:05.958 "rw_ios_per_sec": 0, 00:20:05.958 "rw_mbytes_per_sec": 0, 00:20:05.958 "r_mbytes_per_sec": 0, 00:20:05.958 "w_mbytes_per_sec": 0 00:20:05.958 }, 00:20:05.958 "claimed": true, 00:20:05.958 "claim_type": "exclusive_write", 00:20:05.958 "zoned": false, 00:20:05.958 "supported_io_types": { 00:20:05.958 "read": true, 00:20:05.958 "write": true, 00:20:05.958 "unmap": true, 00:20:05.958 "flush": true, 00:20:05.958 "reset": true, 00:20:05.958 "nvme_admin": false, 00:20:05.958 "nvme_io": false, 00:20:05.958 "nvme_io_md": false, 00:20:05.958 "write_zeroes": true, 00:20:05.958 "zcopy": true, 00:20:05.958 "get_zone_info": false, 00:20:05.958 "zone_management": false, 00:20:05.958 "zone_append": false, 00:20:05.958 "compare": false, 00:20:05.958 "compare_and_write": false, 00:20:05.958 "abort": true, 00:20:05.958 "seek_hole": false, 00:20:05.958 "seek_data": false, 00:20:05.958 "copy": true, 00:20:05.958 "nvme_iov_md": false 00:20:05.958 }, 00:20:05.958 "memory_domains": [ 00:20:05.958 { 00:20:05.958 "dma_device_id": "system", 00:20:05.958 "dma_device_type": 1 00:20:05.958 }, 00:20:05.958 { 00:20:05.958 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:05.958 "dma_device_type": 2 00:20:05.958 } 00:20:05.958 ], 00:20:05.958 "driver_specific": {} 00:20:05.958 }' 00:20:05.958 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:05.958 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:06.257 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:20:06.257 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:06.257 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:06.257 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:20:06.257 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:06.257 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:06.257 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:20:06.257 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:06.257 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:06.257 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:20:06.257 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:20:06.257 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:20:06.257 11:30:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:20:06.525 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:20:06.525 "name": "BaseBdev2", 00:20:06.525 "aliases": [ 00:20:06.525 "cb6d9024-3bcd-494c-a8cd-5058e57e7486" 00:20:06.525 ], 00:20:06.525 "product_name": "Malloc disk", 00:20:06.525 "block_size": 512, 00:20:06.525 "num_blocks": 65536, 00:20:06.525 "uuid": "cb6d9024-3bcd-494c-a8cd-5058e57e7486", 00:20:06.525 "assigned_rate_limits": { 00:20:06.525 "rw_ios_per_sec": 0, 00:20:06.525 "rw_mbytes_per_sec": 0, 00:20:06.525 "r_mbytes_per_sec": 0, 00:20:06.525 "w_mbytes_per_sec": 0 00:20:06.525 }, 00:20:06.525 "claimed": true, 00:20:06.525 "claim_type": "exclusive_write", 00:20:06.525 "zoned": false, 00:20:06.525 "supported_io_types": { 00:20:06.525 "read": true, 00:20:06.525 "write": true, 00:20:06.525 "unmap": true, 00:20:06.525 "flush": true, 00:20:06.525 "reset": true, 00:20:06.525 "nvme_admin": false, 00:20:06.525 "nvme_io": false, 00:20:06.525 "nvme_io_md": false, 00:20:06.525 "write_zeroes": true, 00:20:06.525 "zcopy": true, 00:20:06.525 "get_zone_info": false, 00:20:06.525 "zone_management": false, 00:20:06.525 "zone_append": false, 00:20:06.525 "compare": false, 00:20:06.525 "compare_and_write": false, 00:20:06.525 "abort": true, 00:20:06.525 "seek_hole": false, 00:20:06.525 "seek_data": false, 00:20:06.525 "copy": true, 00:20:06.525 "nvme_iov_md": false 00:20:06.525 }, 00:20:06.525 "memory_domains": [ 00:20:06.525 { 00:20:06.525 "dma_device_id": "system", 00:20:06.525 "dma_device_type": 1 00:20:06.525 }, 00:20:06.525 { 00:20:06.525 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:06.525 "dma_device_type": 2 00:20:06.525 } 00:20:06.525 ], 00:20:06.525 "driver_specific": {} 00:20:06.525 }' 00:20:06.525 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:06.784 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:06.784 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:20:06.784 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:06.784 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:06.784 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:20:06.784 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:06.784 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:06.784 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:20:06.784 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:07.043 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:07.043 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:20:07.043 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:20:07.043 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:20:07.043 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:20:07.043 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:20:07.043 "name": "BaseBdev3", 00:20:07.043 "aliases": [ 00:20:07.043 "5ced5b85-39b5-4152-ab0e-adb93705a2f7" 00:20:07.043 ], 00:20:07.043 "product_name": "Malloc disk", 00:20:07.043 "block_size": 512, 00:20:07.043 "num_blocks": 65536, 00:20:07.043 "uuid": "5ced5b85-39b5-4152-ab0e-adb93705a2f7", 00:20:07.043 "assigned_rate_limits": { 00:20:07.043 "rw_ios_per_sec": 0, 00:20:07.043 "rw_mbytes_per_sec": 0, 00:20:07.043 "r_mbytes_per_sec": 0, 00:20:07.043 "w_mbytes_per_sec": 0 00:20:07.043 }, 00:20:07.043 "claimed": true, 00:20:07.043 "claim_type": "exclusive_write", 00:20:07.043 "zoned": false, 00:20:07.043 "supported_io_types": { 00:20:07.043 "read": true, 00:20:07.043 "write": true, 00:20:07.043 "unmap": true, 00:20:07.043 "flush": true, 00:20:07.043 "reset": true, 00:20:07.043 "nvme_admin": false, 00:20:07.043 "nvme_io": false, 00:20:07.043 "nvme_io_md": false, 00:20:07.043 "write_zeroes": true, 00:20:07.043 "zcopy": true, 00:20:07.043 "get_zone_info": false, 00:20:07.043 "zone_management": false, 00:20:07.043 "zone_append": false, 00:20:07.043 "compare": false, 00:20:07.043 "compare_and_write": false, 00:20:07.043 "abort": true, 00:20:07.043 "seek_hole": false, 00:20:07.043 "seek_data": false, 00:20:07.043 "copy": true, 00:20:07.043 "nvme_iov_md": false 00:20:07.043 }, 00:20:07.043 "memory_domains": [ 00:20:07.043 { 00:20:07.043 "dma_device_id": "system", 00:20:07.043 "dma_device_type": 1 00:20:07.043 }, 00:20:07.043 { 00:20:07.043 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:07.043 "dma_device_type": 2 00:20:07.043 } 00:20:07.043 ], 00:20:07.043 "driver_specific": {} 00:20:07.043 }' 00:20:07.043 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:07.302 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:07.302 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:20:07.302 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:07.302 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:07.302 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:20:07.302 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:07.302 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:07.302 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:20:07.302 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:07.560 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:07.560 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:20:07.560 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:20:07.560 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 00:20:07.560 11:30:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:20:07.818 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:20:07.818 "name": "BaseBdev4", 00:20:07.818 "aliases": [ 00:20:07.818 "736ca049-221b-4e73-8172-b6eb0c46826f" 00:20:07.818 ], 00:20:07.818 "product_name": "Malloc disk", 00:20:07.818 "block_size": 512, 00:20:07.818 "num_blocks": 65536, 00:20:07.818 "uuid": "736ca049-221b-4e73-8172-b6eb0c46826f", 00:20:07.818 "assigned_rate_limits": { 00:20:07.818 "rw_ios_per_sec": 0, 00:20:07.818 "rw_mbytes_per_sec": 0, 00:20:07.818 "r_mbytes_per_sec": 0, 00:20:07.818 "w_mbytes_per_sec": 0 00:20:07.818 }, 00:20:07.818 "claimed": true, 00:20:07.818 "claim_type": "exclusive_write", 00:20:07.818 "zoned": false, 00:20:07.818 "supported_io_types": { 00:20:07.818 "read": true, 00:20:07.818 "write": true, 00:20:07.818 "unmap": true, 00:20:07.818 "flush": true, 00:20:07.818 "reset": true, 00:20:07.818 "nvme_admin": false, 00:20:07.818 "nvme_io": false, 00:20:07.818 "nvme_io_md": false, 00:20:07.818 "write_zeroes": true, 00:20:07.818 "zcopy": true, 00:20:07.818 "get_zone_info": false, 00:20:07.818 "zone_management": false, 00:20:07.818 "zone_append": false, 00:20:07.818 "compare": false, 00:20:07.818 "compare_and_write": false, 00:20:07.818 "abort": true, 00:20:07.818 "seek_hole": false, 00:20:07.818 "seek_data": false, 00:20:07.818 "copy": true, 00:20:07.818 "nvme_iov_md": false 00:20:07.818 }, 00:20:07.818 "memory_domains": [ 00:20:07.818 { 00:20:07.818 "dma_device_id": "system", 00:20:07.818 "dma_device_type": 1 00:20:07.818 }, 00:20:07.818 { 00:20:07.818 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:07.818 "dma_device_type": 2 00:20:07.818 } 00:20:07.818 ], 00:20:07.818 "driver_specific": {} 00:20:07.818 }' 00:20:07.818 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:07.818 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:07.818 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:20:07.818 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:07.818 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:07.818 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:20:07.818 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:07.818 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:08.077 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:20:08.077 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:08.077 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:08.077 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:20:08.077 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:20:08.335 [2024-07-15 11:30:51.715488] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:20:08.335 [2024-07-15 11:30:51.715523] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:20:08.335 [2024-07-15 11:30:51.715572] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:20:08.335 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@275 -- # local expected_state 00:20:08.335 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # has_redundancy concat 00:20:08.335 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # case $1 in 00:20:08.335 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # return 1 00:20:08.335 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@277 -- # expected_state=offline 00:20:08.335 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:20:08.335 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:20:08.335 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=offline 00:20:08.335 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:20:08.335 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:20:08.335 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:20:08.335 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:20:08.335 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:20:08.335 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:20:08.335 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:20:08.335 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:08.335 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:08.594 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:08.594 "name": "Existed_Raid", 00:20:08.594 "uuid": "74ac41db-f972-47ef-aa08-7dfed434ef55", 00:20:08.594 "strip_size_kb": 64, 00:20:08.594 "state": "offline", 00:20:08.594 "raid_level": "concat", 00:20:08.594 "superblock": true, 00:20:08.594 "num_base_bdevs": 4, 00:20:08.594 "num_base_bdevs_discovered": 3, 00:20:08.594 "num_base_bdevs_operational": 3, 00:20:08.594 "base_bdevs_list": [ 00:20:08.594 { 00:20:08.594 "name": null, 00:20:08.594 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:08.594 "is_configured": false, 00:20:08.594 "data_offset": 2048, 00:20:08.594 "data_size": 63488 00:20:08.594 }, 00:20:08.594 { 00:20:08.594 "name": "BaseBdev2", 00:20:08.594 "uuid": "cb6d9024-3bcd-494c-a8cd-5058e57e7486", 00:20:08.594 "is_configured": true, 00:20:08.594 "data_offset": 2048, 00:20:08.594 "data_size": 63488 00:20:08.594 }, 00:20:08.594 { 00:20:08.594 "name": "BaseBdev3", 00:20:08.594 "uuid": "5ced5b85-39b5-4152-ab0e-adb93705a2f7", 00:20:08.594 "is_configured": true, 00:20:08.594 "data_offset": 2048, 00:20:08.594 "data_size": 63488 00:20:08.594 }, 00:20:08.594 { 00:20:08.594 "name": "BaseBdev4", 00:20:08.594 "uuid": "736ca049-221b-4e73-8172-b6eb0c46826f", 00:20:08.594 "is_configured": true, 00:20:08.594 "data_offset": 2048, 00:20:08.594 "data_size": 63488 00:20:08.594 } 00:20:08.594 ] 00:20:08.594 }' 00:20:08.594 11:30:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:08.594 11:30:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:09.160 11:30:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:20:09.160 11:30:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:20:09.160 11:30:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:09.160 11:30:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:20:09.419 11:30:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:20:09.419 11:30:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:20:09.419 11:30:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:20:09.419 [2024-07-15 11:30:52.984804] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:20:09.676 11:30:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:20:09.676 11:30:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:20:09.676 11:30:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:20:09.676 11:30:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:09.676 11:30:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:20:09.677 11:30:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:20:09.677 11:30:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev3 00:20:09.933 [2024-07-15 11:30:53.484694] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:20:09.933 11:30:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:20:09.933 11:30:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:20:09.933 11:30:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:09.933 11:30:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:20:10.191 11:30:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:20:10.191 11:30:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:20:10.191 11:30:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev4 00:20:10.449 [2024-07-15 11:30:53.989506] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:20:10.449 [2024-07-15 11:30:53.989551] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xcf2350 name Existed_Raid, state offline 00:20:10.449 11:30:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:20:10.449 11:30:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:20:10.449 11:30:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:10.449 11:30:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:20:10.707 11:30:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:20:10.707 11:30:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:20:10.707 11:30:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # '[' 4 -gt 2 ']' 00:20:10.707 11:30:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i = 1 )) 00:20:10.707 11:30:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:20:10.707 11:30:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:20:10.966 BaseBdev2 00:20:10.966 11:30:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev2 00:20:10.966 11:30:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:20:10.966 11:30:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:20:10.966 11:30:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:20:10.966 11:30:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:20:10.966 11:30:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:20:10.966 11:30:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:20:11.225 11:30:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:20:11.483 [ 00:20:11.483 { 00:20:11.483 "name": "BaseBdev2", 00:20:11.483 "aliases": [ 00:20:11.483 "8d6e3416-5b01-4367-a1f6-d092d3a176be" 00:20:11.483 ], 00:20:11.483 "product_name": "Malloc disk", 00:20:11.483 "block_size": 512, 00:20:11.483 "num_blocks": 65536, 00:20:11.483 "uuid": "8d6e3416-5b01-4367-a1f6-d092d3a176be", 00:20:11.483 "assigned_rate_limits": { 00:20:11.483 "rw_ios_per_sec": 0, 00:20:11.483 "rw_mbytes_per_sec": 0, 00:20:11.483 "r_mbytes_per_sec": 0, 00:20:11.483 "w_mbytes_per_sec": 0 00:20:11.483 }, 00:20:11.483 "claimed": false, 00:20:11.483 "zoned": false, 00:20:11.483 "supported_io_types": { 00:20:11.483 "read": true, 00:20:11.483 "write": true, 00:20:11.483 "unmap": true, 00:20:11.483 "flush": true, 00:20:11.483 "reset": true, 00:20:11.483 "nvme_admin": false, 00:20:11.483 "nvme_io": false, 00:20:11.483 "nvme_io_md": false, 00:20:11.483 "write_zeroes": true, 00:20:11.483 "zcopy": true, 00:20:11.483 "get_zone_info": false, 00:20:11.483 "zone_management": false, 00:20:11.483 "zone_append": false, 00:20:11.483 "compare": false, 00:20:11.483 "compare_and_write": false, 00:20:11.483 "abort": true, 00:20:11.483 "seek_hole": false, 00:20:11.483 "seek_data": false, 00:20:11.483 "copy": true, 00:20:11.483 "nvme_iov_md": false 00:20:11.483 }, 00:20:11.483 "memory_domains": [ 00:20:11.483 { 00:20:11.483 "dma_device_id": "system", 00:20:11.483 "dma_device_type": 1 00:20:11.483 }, 00:20:11.483 { 00:20:11.483 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:11.483 "dma_device_type": 2 00:20:11.483 } 00:20:11.483 ], 00:20:11.483 "driver_specific": {} 00:20:11.483 } 00:20:11.483 ] 00:20:11.483 11:30:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:20:11.483 11:30:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:20:11.483 11:30:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:20:11.483 11:30:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:20:11.741 BaseBdev3 00:20:11.741 11:30:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev3 00:20:11.741 11:30:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:20:11.741 11:30:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:20:11.741 11:30:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:20:11.741 11:30:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:20:11.741 11:30:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:20:11.741 11:30:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:20:12.000 11:30:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:20:12.259 [ 00:20:12.259 { 00:20:12.259 "name": "BaseBdev3", 00:20:12.259 "aliases": [ 00:20:12.259 "520833b6-eebb-4cbf-b605-5fd8c5fc6ee6" 00:20:12.259 ], 00:20:12.259 "product_name": "Malloc disk", 00:20:12.259 "block_size": 512, 00:20:12.259 "num_blocks": 65536, 00:20:12.259 "uuid": "520833b6-eebb-4cbf-b605-5fd8c5fc6ee6", 00:20:12.259 "assigned_rate_limits": { 00:20:12.259 "rw_ios_per_sec": 0, 00:20:12.259 "rw_mbytes_per_sec": 0, 00:20:12.259 "r_mbytes_per_sec": 0, 00:20:12.259 "w_mbytes_per_sec": 0 00:20:12.259 }, 00:20:12.259 "claimed": false, 00:20:12.259 "zoned": false, 00:20:12.259 "supported_io_types": { 00:20:12.259 "read": true, 00:20:12.259 "write": true, 00:20:12.259 "unmap": true, 00:20:12.259 "flush": true, 00:20:12.259 "reset": true, 00:20:12.259 "nvme_admin": false, 00:20:12.259 "nvme_io": false, 00:20:12.259 "nvme_io_md": false, 00:20:12.259 "write_zeroes": true, 00:20:12.259 "zcopy": true, 00:20:12.259 "get_zone_info": false, 00:20:12.259 "zone_management": false, 00:20:12.259 "zone_append": false, 00:20:12.259 "compare": false, 00:20:12.259 "compare_and_write": false, 00:20:12.259 "abort": true, 00:20:12.259 "seek_hole": false, 00:20:12.259 "seek_data": false, 00:20:12.259 "copy": true, 00:20:12.259 "nvme_iov_md": false 00:20:12.259 }, 00:20:12.259 "memory_domains": [ 00:20:12.259 { 00:20:12.259 "dma_device_id": "system", 00:20:12.259 "dma_device_type": 1 00:20:12.259 }, 00:20:12.259 { 00:20:12.259 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:12.259 "dma_device_type": 2 00:20:12.259 } 00:20:12.259 ], 00:20:12.259 "driver_specific": {} 00:20:12.259 } 00:20:12.259 ] 00:20:12.259 11:30:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:20:12.259 11:30:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:20:12.259 11:30:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:20:12.259 11:30:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4 00:20:12.517 BaseBdev4 00:20:12.517 11:30:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev4 00:20:12.517 11:30:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev4 00:20:12.517 11:30:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:20:12.517 11:30:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:20:12.517 11:30:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:20:12.517 11:30:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:20:12.518 11:30:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:20:12.777 11:30:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 -t 2000 00:20:13.036 [ 00:20:13.036 { 00:20:13.036 "name": "BaseBdev4", 00:20:13.036 "aliases": [ 00:20:13.036 "51adb667-52ff-4ed5-aa69-b267e6041e14" 00:20:13.036 ], 00:20:13.036 "product_name": "Malloc disk", 00:20:13.036 "block_size": 512, 00:20:13.036 "num_blocks": 65536, 00:20:13.036 "uuid": "51adb667-52ff-4ed5-aa69-b267e6041e14", 00:20:13.036 "assigned_rate_limits": { 00:20:13.036 "rw_ios_per_sec": 0, 00:20:13.036 "rw_mbytes_per_sec": 0, 00:20:13.036 "r_mbytes_per_sec": 0, 00:20:13.036 "w_mbytes_per_sec": 0 00:20:13.036 }, 00:20:13.036 "claimed": false, 00:20:13.036 "zoned": false, 00:20:13.036 "supported_io_types": { 00:20:13.036 "read": true, 00:20:13.036 "write": true, 00:20:13.036 "unmap": true, 00:20:13.036 "flush": true, 00:20:13.036 "reset": true, 00:20:13.036 "nvme_admin": false, 00:20:13.036 "nvme_io": false, 00:20:13.036 "nvme_io_md": false, 00:20:13.036 "write_zeroes": true, 00:20:13.036 "zcopy": true, 00:20:13.036 "get_zone_info": false, 00:20:13.036 "zone_management": false, 00:20:13.036 "zone_append": false, 00:20:13.036 "compare": false, 00:20:13.036 "compare_and_write": false, 00:20:13.036 "abort": true, 00:20:13.036 "seek_hole": false, 00:20:13.036 "seek_data": false, 00:20:13.036 "copy": true, 00:20:13.036 "nvme_iov_md": false 00:20:13.036 }, 00:20:13.036 "memory_domains": [ 00:20:13.036 { 00:20:13.036 "dma_device_id": "system", 00:20:13.036 "dma_device_type": 1 00:20:13.036 }, 00:20:13.036 { 00:20:13.036 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:13.036 "dma_device_type": 2 00:20:13.036 } 00:20:13.036 ], 00:20:13.036 "driver_specific": {} 00:20:13.036 } 00:20:13.036 ] 00:20:13.036 11:30:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:20:13.036 11:30:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:20:13.036 11:30:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:20:13.036 11:30:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@305 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -s -r concat -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:20:13.295 [2024-07-15 11:30:56.679648] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:20:13.295 [2024-07-15 11:30:56.679687] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:20:13.295 [2024-07-15 11:30:56.679709] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:20:13.295 [2024-07-15 11:30:56.681029] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:20:13.295 [2024-07-15 11:30:56.681069] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:20:13.295 11:30:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:20:13.295 11:30:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:20:13.295 11:30:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:20:13.295 11:30:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:20:13.295 11:30:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:20:13.295 11:30:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:20:13.295 11:30:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:20:13.295 11:30:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:20:13.295 11:30:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:20:13.295 11:30:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:20:13.295 11:30:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:13.295 11:30:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:13.553 11:30:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:13.553 "name": "Existed_Raid", 00:20:13.553 "uuid": "a09b7cb5-6c0e-46e7-ae81-cb1471a4bfea", 00:20:13.553 "strip_size_kb": 64, 00:20:13.553 "state": "configuring", 00:20:13.553 "raid_level": "concat", 00:20:13.553 "superblock": true, 00:20:13.553 "num_base_bdevs": 4, 00:20:13.553 "num_base_bdevs_discovered": 3, 00:20:13.553 "num_base_bdevs_operational": 4, 00:20:13.553 "base_bdevs_list": [ 00:20:13.553 { 00:20:13.553 "name": "BaseBdev1", 00:20:13.553 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:13.553 "is_configured": false, 00:20:13.553 "data_offset": 0, 00:20:13.553 "data_size": 0 00:20:13.553 }, 00:20:13.553 { 00:20:13.553 "name": "BaseBdev2", 00:20:13.553 "uuid": "8d6e3416-5b01-4367-a1f6-d092d3a176be", 00:20:13.553 "is_configured": true, 00:20:13.554 "data_offset": 2048, 00:20:13.554 "data_size": 63488 00:20:13.554 }, 00:20:13.554 { 00:20:13.554 "name": "BaseBdev3", 00:20:13.554 "uuid": "520833b6-eebb-4cbf-b605-5fd8c5fc6ee6", 00:20:13.554 "is_configured": true, 00:20:13.554 "data_offset": 2048, 00:20:13.554 "data_size": 63488 00:20:13.554 }, 00:20:13.554 { 00:20:13.554 "name": "BaseBdev4", 00:20:13.554 "uuid": "51adb667-52ff-4ed5-aa69-b267e6041e14", 00:20:13.554 "is_configured": true, 00:20:13.554 "data_offset": 2048, 00:20:13.554 "data_size": 63488 00:20:13.554 } 00:20:13.554 ] 00:20:13.554 }' 00:20:13.554 11:30:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:13.554 11:30:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:14.122 11:30:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev2 00:20:14.122 [2024-07-15 11:30:57.654189] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:20:14.122 11:30:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@309 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:20:14.122 11:30:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:20:14.122 11:30:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:20:14.122 11:30:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:20:14.122 11:30:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:20:14.122 11:30:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:20:14.122 11:30:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:20:14.122 11:30:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:20:14.122 11:30:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:20:14.122 11:30:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:20:14.122 11:30:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:14.122 11:30:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:14.381 11:30:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:14.381 "name": "Existed_Raid", 00:20:14.381 "uuid": "a09b7cb5-6c0e-46e7-ae81-cb1471a4bfea", 00:20:14.381 "strip_size_kb": 64, 00:20:14.381 "state": "configuring", 00:20:14.381 "raid_level": "concat", 00:20:14.381 "superblock": true, 00:20:14.381 "num_base_bdevs": 4, 00:20:14.381 "num_base_bdevs_discovered": 2, 00:20:14.381 "num_base_bdevs_operational": 4, 00:20:14.381 "base_bdevs_list": [ 00:20:14.381 { 00:20:14.381 "name": "BaseBdev1", 00:20:14.381 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:14.381 "is_configured": false, 00:20:14.381 "data_offset": 0, 00:20:14.381 "data_size": 0 00:20:14.381 }, 00:20:14.381 { 00:20:14.381 "name": null, 00:20:14.381 "uuid": "8d6e3416-5b01-4367-a1f6-d092d3a176be", 00:20:14.381 "is_configured": false, 00:20:14.381 "data_offset": 2048, 00:20:14.381 "data_size": 63488 00:20:14.381 }, 00:20:14.381 { 00:20:14.381 "name": "BaseBdev3", 00:20:14.381 "uuid": "520833b6-eebb-4cbf-b605-5fd8c5fc6ee6", 00:20:14.381 "is_configured": true, 00:20:14.381 "data_offset": 2048, 00:20:14.381 "data_size": 63488 00:20:14.381 }, 00:20:14.381 { 00:20:14.381 "name": "BaseBdev4", 00:20:14.381 "uuid": "51adb667-52ff-4ed5-aa69-b267e6041e14", 00:20:14.381 "is_configured": true, 00:20:14.381 "data_offset": 2048, 00:20:14.381 "data_size": 63488 00:20:14.381 } 00:20:14.381 ] 00:20:14.381 }' 00:20:14.381 11:30:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:14.381 11:30:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:14.949 11:30:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:14.949 11:30:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:20:15.207 11:30:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # [[ false == \f\a\l\s\e ]] 00:20:15.207 11:30:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:20:15.466 [2024-07-15 11:30:58.861019] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:20:15.466 BaseBdev1 00:20:15.466 11:30:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@313 -- # waitforbdev BaseBdev1 00:20:15.466 11:30:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:20:15.466 11:30:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:20:15.466 11:30:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:20:15.466 11:30:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:20:15.466 11:30:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:20:15.466 11:30:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:20:15.725 11:30:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:20:15.725 [ 00:20:15.725 { 00:20:15.725 "name": "BaseBdev1", 00:20:15.725 "aliases": [ 00:20:15.725 "98aa2c04-1df4-4b92-a864-887d0de265fe" 00:20:15.725 ], 00:20:15.725 "product_name": "Malloc disk", 00:20:15.725 "block_size": 512, 00:20:15.725 "num_blocks": 65536, 00:20:15.725 "uuid": "98aa2c04-1df4-4b92-a864-887d0de265fe", 00:20:15.725 "assigned_rate_limits": { 00:20:15.725 "rw_ios_per_sec": 0, 00:20:15.725 "rw_mbytes_per_sec": 0, 00:20:15.725 "r_mbytes_per_sec": 0, 00:20:15.725 "w_mbytes_per_sec": 0 00:20:15.725 }, 00:20:15.725 "claimed": true, 00:20:15.725 "claim_type": "exclusive_write", 00:20:15.725 "zoned": false, 00:20:15.725 "supported_io_types": { 00:20:15.725 "read": true, 00:20:15.725 "write": true, 00:20:15.725 "unmap": true, 00:20:15.725 "flush": true, 00:20:15.725 "reset": true, 00:20:15.725 "nvme_admin": false, 00:20:15.725 "nvme_io": false, 00:20:15.725 "nvme_io_md": false, 00:20:15.725 "write_zeroes": true, 00:20:15.725 "zcopy": true, 00:20:15.725 "get_zone_info": false, 00:20:15.725 "zone_management": false, 00:20:15.725 "zone_append": false, 00:20:15.725 "compare": false, 00:20:15.725 "compare_and_write": false, 00:20:15.725 "abort": true, 00:20:15.725 "seek_hole": false, 00:20:15.725 "seek_data": false, 00:20:15.725 "copy": true, 00:20:15.725 "nvme_iov_md": false 00:20:15.725 }, 00:20:15.725 "memory_domains": [ 00:20:15.725 { 00:20:15.725 "dma_device_id": "system", 00:20:15.725 "dma_device_type": 1 00:20:15.725 }, 00:20:15.725 { 00:20:15.725 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:15.725 "dma_device_type": 2 00:20:15.725 } 00:20:15.725 ], 00:20:15.725 "driver_specific": {} 00:20:15.725 } 00:20:15.725 ] 00:20:15.725 11:30:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:20:15.725 11:30:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:20:15.725 11:30:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:20:15.725 11:30:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:20:15.725 11:30:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:20:15.725 11:30:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:20:15.725 11:30:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:20:15.725 11:30:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:20:15.725 11:30:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:20:15.725 11:30:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:20:15.725 11:30:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:20:15.725 11:30:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:15.725 11:30:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:15.983 11:30:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:15.983 "name": "Existed_Raid", 00:20:15.983 "uuid": "a09b7cb5-6c0e-46e7-ae81-cb1471a4bfea", 00:20:15.983 "strip_size_kb": 64, 00:20:15.983 "state": "configuring", 00:20:15.983 "raid_level": "concat", 00:20:15.983 "superblock": true, 00:20:15.983 "num_base_bdevs": 4, 00:20:15.983 "num_base_bdevs_discovered": 3, 00:20:15.983 "num_base_bdevs_operational": 4, 00:20:15.983 "base_bdevs_list": [ 00:20:15.983 { 00:20:15.983 "name": "BaseBdev1", 00:20:15.983 "uuid": "98aa2c04-1df4-4b92-a864-887d0de265fe", 00:20:15.983 "is_configured": true, 00:20:15.983 "data_offset": 2048, 00:20:15.983 "data_size": 63488 00:20:15.983 }, 00:20:15.983 { 00:20:15.983 "name": null, 00:20:15.983 "uuid": "8d6e3416-5b01-4367-a1f6-d092d3a176be", 00:20:15.983 "is_configured": false, 00:20:15.983 "data_offset": 2048, 00:20:15.983 "data_size": 63488 00:20:15.983 }, 00:20:15.983 { 00:20:15.983 "name": "BaseBdev3", 00:20:15.983 "uuid": "520833b6-eebb-4cbf-b605-5fd8c5fc6ee6", 00:20:15.983 "is_configured": true, 00:20:15.983 "data_offset": 2048, 00:20:15.983 "data_size": 63488 00:20:15.983 }, 00:20:15.983 { 00:20:15.983 "name": "BaseBdev4", 00:20:15.983 "uuid": "51adb667-52ff-4ed5-aa69-b267e6041e14", 00:20:15.983 "is_configured": true, 00:20:15.983 "data_offset": 2048, 00:20:15.983 "data_size": 63488 00:20:15.983 } 00:20:15.983 ] 00:20:15.983 }' 00:20:15.983 11:30:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:15.983 11:30:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:16.917 11:31:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:16.917 11:31:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:20:16.917 11:31:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # [[ true == \t\r\u\e ]] 00:20:16.917 11:31:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@317 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev3 00:20:17.175 [2024-07-15 11:31:00.641770] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:20:17.175 11:31:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:20:17.175 11:31:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:20:17.175 11:31:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:20:17.175 11:31:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:20:17.175 11:31:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:20:17.175 11:31:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:20:17.175 11:31:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:20:17.175 11:31:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:20:17.175 11:31:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:20:17.175 11:31:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:20:17.175 11:31:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:17.175 11:31:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:17.434 11:31:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:17.434 "name": "Existed_Raid", 00:20:17.434 "uuid": "a09b7cb5-6c0e-46e7-ae81-cb1471a4bfea", 00:20:17.434 "strip_size_kb": 64, 00:20:17.434 "state": "configuring", 00:20:17.434 "raid_level": "concat", 00:20:17.434 "superblock": true, 00:20:17.434 "num_base_bdevs": 4, 00:20:17.434 "num_base_bdevs_discovered": 2, 00:20:17.434 "num_base_bdevs_operational": 4, 00:20:17.434 "base_bdevs_list": [ 00:20:17.434 { 00:20:17.434 "name": "BaseBdev1", 00:20:17.434 "uuid": "98aa2c04-1df4-4b92-a864-887d0de265fe", 00:20:17.434 "is_configured": true, 00:20:17.434 "data_offset": 2048, 00:20:17.434 "data_size": 63488 00:20:17.434 }, 00:20:17.434 { 00:20:17.434 "name": null, 00:20:17.434 "uuid": "8d6e3416-5b01-4367-a1f6-d092d3a176be", 00:20:17.434 "is_configured": false, 00:20:17.434 "data_offset": 2048, 00:20:17.434 "data_size": 63488 00:20:17.434 }, 00:20:17.434 { 00:20:17.434 "name": null, 00:20:17.434 "uuid": "520833b6-eebb-4cbf-b605-5fd8c5fc6ee6", 00:20:17.434 "is_configured": false, 00:20:17.434 "data_offset": 2048, 00:20:17.434 "data_size": 63488 00:20:17.434 }, 00:20:17.434 { 00:20:17.434 "name": "BaseBdev4", 00:20:17.434 "uuid": "51adb667-52ff-4ed5-aa69-b267e6041e14", 00:20:17.434 "is_configured": true, 00:20:17.434 "data_offset": 2048, 00:20:17.434 "data_size": 63488 00:20:17.434 } 00:20:17.434 ] 00:20:17.434 }' 00:20:17.434 11:31:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:17.434 11:31:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:17.999 11:31:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:17.999 11:31:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:20:18.257 11:31:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # [[ false == \f\a\l\s\e ]] 00:20:18.257 11:31:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:20:18.515 [2024-07-15 11:31:01.901274] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:20:18.515 11:31:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@322 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:20:18.515 11:31:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:20:18.515 11:31:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:20:18.515 11:31:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:20:18.515 11:31:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:20:18.515 11:31:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:20:18.515 11:31:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:20:18.515 11:31:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:20:18.515 11:31:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:20:18.515 11:31:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:20:18.515 11:31:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:18.515 11:31:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:18.774 11:31:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:18.774 "name": "Existed_Raid", 00:20:18.774 "uuid": "a09b7cb5-6c0e-46e7-ae81-cb1471a4bfea", 00:20:18.774 "strip_size_kb": 64, 00:20:18.774 "state": "configuring", 00:20:18.774 "raid_level": "concat", 00:20:18.774 "superblock": true, 00:20:18.774 "num_base_bdevs": 4, 00:20:18.774 "num_base_bdevs_discovered": 3, 00:20:18.774 "num_base_bdevs_operational": 4, 00:20:18.774 "base_bdevs_list": [ 00:20:18.774 { 00:20:18.774 "name": "BaseBdev1", 00:20:18.774 "uuid": "98aa2c04-1df4-4b92-a864-887d0de265fe", 00:20:18.774 "is_configured": true, 00:20:18.774 "data_offset": 2048, 00:20:18.774 "data_size": 63488 00:20:18.774 }, 00:20:18.774 { 00:20:18.774 "name": null, 00:20:18.774 "uuid": "8d6e3416-5b01-4367-a1f6-d092d3a176be", 00:20:18.774 "is_configured": false, 00:20:18.774 "data_offset": 2048, 00:20:18.774 "data_size": 63488 00:20:18.774 }, 00:20:18.774 { 00:20:18.774 "name": "BaseBdev3", 00:20:18.774 "uuid": "520833b6-eebb-4cbf-b605-5fd8c5fc6ee6", 00:20:18.774 "is_configured": true, 00:20:18.774 "data_offset": 2048, 00:20:18.774 "data_size": 63488 00:20:18.774 }, 00:20:18.774 { 00:20:18.774 "name": "BaseBdev4", 00:20:18.774 "uuid": "51adb667-52ff-4ed5-aa69-b267e6041e14", 00:20:18.774 "is_configured": true, 00:20:18.774 "data_offset": 2048, 00:20:18.774 "data_size": 63488 00:20:18.774 } 00:20:18.774 ] 00:20:18.774 }' 00:20:18.774 11:31:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:18.774 11:31:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:19.713 11:31:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:19.713 11:31:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:20:19.713 11:31:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # [[ true == \t\r\u\e ]] 00:20:19.713 11:31:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@325 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:20:19.978 [2024-07-15 11:31:03.413303] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:20:19.978 11:31:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:20:19.978 11:31:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:20:19.978 11:31:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:20:19.978 11:31:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:20:19.978 11:31:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:20:19.978 11:31:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:20:19.978 11:31:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:20:19.978 11:31:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:20:19.978 11:31:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:20:19.978 11:31:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:20:19.978 11:31:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:19.978 11:31:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:20.236 11:31:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:20.236 "name": "Existed_Raid", 00:20:20.236 "uuid": "a09b7cb5-6c0e-46e7-ae81-cb1471a4bfea", 00:20:20.236 "strip_size_kb": 64, 00:20:20.236 "state": "configuring", 00:20:20.236 "raid_level": "concat", 00:20:20.236 "superblock": true, 00:20:20.236 "num_base_bdevs": 4, 00:20:20.236 "num_base_bdevs_discovered": 2, 00:20:20.236 "num_base_bdevs_operational": 4, 00:20:20.236 "base_bdevs_list": [ 00:20:20.236 { 00:20:20.236 "name": null, 00:20:20.236 "uuid": "98aa2c04-1df4-4b92-a864-887d0de265fe", 00:20:20.236 "is_configured": false, 00:20:20.236 "data_offset": 2048, 00:20:20.236 "data_size": 63488 00:20:20.236 }, 00:20:20.236 { 00:20:20.236 "name": null, 00:20:20.236 "uuid": "8d6e3416-5b01-4367-a1f6-d092d3a176be", 00:20:20.236 "is_configured": false, 00:20:20.236 "data_offset": 2048, 00:20:20.236 "data_size": 63488 00:20:20.236 }, 00:20:20.236 { 00:20:20.236 "name": "BaseBdev3", 00:20:20.236 "uuid": "520833b6-eebb-4cbf-b605-5fd8c5fc6ee6", 00:20:20.236 "is_configured": true, 00:20:20.236 "data_offset": 2048, 00:20:20.236 "data_size": 63488 00:20:20.236 }, 00:20:20.236 { 00:20:20.236 "name": "BaseBdev4", 00:20:20.236 "uuid": "51adb667-52ff-4ed5-aa69-b267e6041e14", 00:20:20.236 "is_configured": true, 00:20:20.236 "data_offset": 2048, 00:20:20.236 "data_size": 63488 00:20:20.236 } 00:20:20.236 ] 00:20:20.236 }' 00:20:20.236 11:31:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:20.236 11:31:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:20.830 11:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@327 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:20.830 11:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@327 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:20:20.830 11:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@327 -- # [[ false == \f\a\l\s\e ]] 00:20:20.830 11:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@329 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:20:21.089 [2024-07-15 11:31:04.612850] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:20:21.089 11:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@330 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:20:21.089 11:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:20:21.089 11:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:20:21.089 11:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:20:21.089 11:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:20:21.089 11:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:20:21.089 11:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:20:21.089 11:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:20:21.089 11:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:20:21.089 11:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:20:21.089 11:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:21.089 11:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:21.654 11:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:21.654 "name": "Existed_Raid", 00:20:21.654 "uuid": "a09b7cb5-6c0e-46e7-ae81-cb1471a4bfea", 00:20:21.654 "strip_size_kb": 64, 00:20:21.654 "state": "configuring", 00:20:21.654 "raid_level": "concat", 00:20:21.654 "superblock": true, 00:20:21.654 "num_base_bdevs": 4, 00:20:21.654 "num_base_bdevs_discovered": 3, 00:20:21.654 "num_base_bdevs_operational": 4, 00:20:21.654 "base_bdevs_list": [ 00:20:21.654 { 00:20:21.654 "name": null, 00:20:21.654 "uuid": "98aa2c04-1df4-4b92-a864-887d0de265fe", 00:20:21.654 "is_configured": false, 00:20:21.654 "data_offset": 2048, 00:20:21.654 "data_size": 63488 00:20:21.654 }, 00:20:21.654 { 00:20:21.654 "name": "BaseBdev2", 00:20:21.654 "uuid": "8d6e3416-5b01-4367-a1f6-d092d3a176be", 00:20:21.654 "is_configured": true, 00:20:21.654 "data_offset": 2048, 00:20:21.654 "data_size": 63488 00:20:21.654 }, 00:20:21.654 { 00:20:21.654 "name": "BaseBdev3", 00:20:21.654 "uuid": "520833b6-eebb-4cbf-b605-5fd8c5fc6ee6", 00:20:21.654 "is_configured": true, 00:20:21.654 "data_offset": 2048, 00:20:21.654 "data_size": 63488 00:20:21.654 }, 00:20:21.654 { 00:20:21.654 "name": "BaseBdev4", 00:20:21.654 "uuid": "51adb667-52ff-4ed5-aa69-b267e6041e14", 00:20:21.654 "is_configured": true, 00:20:21.654 "data_offset": 2048, 00:20:21.654 "data_size": 63488 00:20:21.654 } 00:20:21.654 ] 00:20:21.654 }' 00:20:21.654 11:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:21.654 11:31:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:22.217 11:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@331 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:22.217 11:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@331 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:20:22.474 11:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@331 -- # [[ true == \t\r\u\e ]] 00:20:22.474 11:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:22.474 11:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@333 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:20:22.731 11:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b NewBaseBdev -u 98aa2c04-1df4-4b92-a864-887d0de265fe 00:20:23.296 [2024-07-15 11:31:06.649556] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:20:23.296 [2024-07-15 11:31:06.649712] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xcf4850 00:20:23.296 [2024-07-15 11:31:06.649725] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:20:23.296 [2024-07-15 11:31:06.649898] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xcead80 00:20:23.296 [2024-07-15 11:31:06.650024] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xcf4850 00:20:23.296 [2024-07-15 11:31:06.650034] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0xcf4850 00:20:23.296 [2024-07-15 11:31:06.650123] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:23.296 NewBaseBdev 00:20:23.296 11:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@334 -- # waitforbdev NewBaseBdev 00:20:23.296 11:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=NewBaseBdev 00:20:23.296 11:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:20:23.297 11:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:20:23.297 11:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:20:23.297 11:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:20:23.297 11:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:20:23.554 11:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev -t 2000 00:20:23.811 [ 00:20:23.811 { 00:20:23.811 "name": "NewBaseBdev", 00:20:23.811 "aliases": [ 00:20:23.811 "98aa2c04-1df4-4b92-a864-887d0de265fe" 00:20:23.811 ], 00:20:23.811 "product_name": "Malloc disk", 00:20:23.811 "block_size": 512, 00:20:23.811 "num_blocks": 65536, 00:20:23.811 "uuid": "98aa2c04-1df4-4b92-a864-887d0de265fe", 00:20:23.811 "assigned_rate_limits": { 00:20:23.811 "rw_ios_per_sec": 0, 00:20:23.811 "rw_mbytes_per_sec": 0, 00:20:23.812 "r_mbytes_per_sec": 0, 00:20:23.812 "w_mbytes_per_sec": 0 00:20:23.812 }, 00:20:23.812 "claimed": true, 00:20:23.812 "claim_type": "exclusive_write", 00:20:23.812 "zoned": false, 00:20:23.812 "supported_io_types": { 00:20:23.812 "read": true, 00:20:23.812 "write": true, 00:20:23.812 "unmap": true, 00:20:23.812 "flush": true, 00:20:23.812 "reset": true, 00:20:23.812 "nvme_admin": false, 00:20:23.812 "nvme_io": false, 00:20:23.812 "nvme_io_md": false, 00:20:23.812 "write_zeroes": true, 00:20:23.812 "zcopy": true, 00:20:23.812 "get_zone_info": false, 00:20:23.812 "zone_management": false, 00:20:23.812 "zone_append": false, 00:20:23.812 "compare": false, 00:20:23.812 "compare_and_write": false, 00:20:23.812 "abort": true, 00:20:23.812 "seek_hole": false, 00:20:23.812 "seek_data": false, 00:20:23.812 "copy": true, 00:20:23.812 "nvme_iov_md": false 00:20:23.812 }, 00:20:23.812 "memory_domains": [ 00:20:23.812 { 00:20:23.812 "dma_device_id": "system", 00:20:23.812 "dma_device_type": 1 00:20:23.812 }, 00:20:23.812 { 00:20:23.812 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:23.812 "dma_device_type": 2 00:20:23.812 } 00:20:23.812 ], 00:20:23.812 "driver_specific": {} 00:20:23.812 } 00:20:23.812 ] 00:20:23.812 11:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:20:23.812 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@335 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:20:23.812 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:20:23.812 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:20:23.812 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:20:23.812 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:20:23.812 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:20:23.812 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:20:23.812 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:20:23.812 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:20:23.812 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:20:23.812 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:23.812 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:24.070 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:24.070 "name": "Existed_Raid", 00:20:24.070 "uuid": "a09b7cb5-6c0e-46e7-ae81-cb1471a4bfea", 00:20:24.070 "strip_size_kb": 64, 00:20:24.070 "state": "online", 00:20:24.070 "raid_level": "concat", 00:20:24.070 "superblock": true, 00:20:24.070 "num_base_bdevs": 4, 00:20:24.070 "num_base_bdevs_discovered": 4, 00:20:24.070 "num_base_bdevs_operational": 4, 00:20:24.070 "base_bdevs_list": [ 00:20:24.070 { 00:20:24.070 "name": "NewBaseBdev", 00:20:24.070 "uuid": "98aa2c04-1df4-4b92-a864-887d0de265fe", 00:20:24.070 "is_configured": true, 00:20:24.070 "data_offset": 2048, 00:20:24.070 "data_size": 63488 00:20:24.070 }, 00:20:24.070 { 00:20:24.070 "name": "BaseBdev2", 00:20:24.070 "uuid": "8d6e3416-5b01-4367-a1f6-d092d3a176be", 00:20:24.070 "is_configured": true, 00:20:24.070 "data_offset": 2048, 00:20:24.070 "data_size": 63488 00:20:24.070 }, 00:20:24.070 { 00:20:24.070 "name": "BaseBdev3", 00:20:24.070 "uuid": "520833b6-eebb-4cbf-b605-5fd8c5fc6ee6", 00:20:24.070 "is_configured": true, 00:20:24.070 "data_offset": 2048, 00:20:24.070 "data_size": 63488 00:20:24.070 }, 00:20:24.070 { 00:20:24.070 "name": "BaseBdev4", 00:20:24.070 "uuid": "51adb667-52ff-4ed5-aa69-b267e6041e14", 00:20:24.070 "is_configured": true, 00:20:24.070 "data_offset": 2048, 00:20:24.070 "data_size": 63488 00:20:24.070 } 00:20:24.070 ] 00:20:24.070 }' 00:20:24.070 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:24.070 11:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:24.637 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@336 -- # verify_raid_bdev_properties Existed_Raid 00:20:24.637 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:20:24.637 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:20:24.637 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:20:24.637 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:20:24.637 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # local name 00:20:24.637 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:20:24.637 11:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:20:24.637 [2024-07-15 11:31:08.210036] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:20:24.894 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:20:24.894 "name": "Existed_Raid", 00:20:24.894 "aliases": [ 00:20:24.894 "a09b7cb5-6c0e-46e7-ae81-cb1471a4bfea" 00:20:24.894 ], 00:20:24.894 "product_name": "Raid Volume", 00:20:24.894 "block_size": 512, 00:20:24.894 "num_blocks": 253952, 00:20:24.894 "uuid": "a09b7cb5-6c0e-46e7-ae81-cb1471a4bfea", 00:20:24.894 "assigned_rate_limits": { 00:20:24.894 "rw_ios_per_sec": 0, 00:20:24.894 "rw_mbytes_per_sec": 0, 00:20:24.894 "r_mbytes_per_sec": 0, 00:20:24.894 "w_mbytes_per_sec": 0 00:20:24.894 }, 00:20:24.894 "claimed": false, 00:20:24.894 "zoned": false, 00:20:24.894 "supported_io_types": { 00:20:24.894 "read": true, 00:20:24.894 "write": true, 00:20:24.894 "unmap": true, 00:20:24.895 "flush": true, 00:20:24.895 "reset": true, 00:20:24.895 "nvme_admin": false, 00:20:24.895 "nvme_io": false, 00:20:24.895 "nvme_io_md": false, 00:20:24.895 "write_zeroes": true, 00:20:24.895 "zcopy": false, 00:20:24.895 "get_zone_info": false, 00:20:24.895 "zone_management": false, 00:20:24.895 "zone_append": false, 00:20:24.895 "compare": false, 00:20:24.895 "compare_and_write": false, 00:20:24.895 "abort": false, 00:20:24.895 "seek_hole": false, 00:20:24.895 "seek_data": false, 00:20:24.895 "copy": false, 00:20:24.895 "nvme_iov_md": false 00:20:24.895 }, 00:20:24.895 "memory_domains": [ 00:20:24.895 { 00:20:24.895 "dma_device_id": "system", 00:20:24.895 "dma_device_type": 1 00:20:24.895 }, 00:20:24.895 { 00:20:24.895 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:24.895 "dma_device_type": 2 00:20:24.895 }, 00:20:24.895 { 00:20:24.895 "dma_device_id": "system", 00:20:24.895 "dma_device_type": 1 00:20:24.895 }, 00:20:24.895 { 00:20:24.895 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:24.895 "dma_device_type": 2 00:20:24.895 }, 00:20:24.895 { 00:20:24.895 "dma_device_id": "system", 00:20:24.895 "dma_device_type": 1 00:20:24.895 }, 00:20:24.895 { 00:20:24.895 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:24.895 "dma_device_type": 2 00:20:24.895 }, 00:20:24.895 { 00:20:24.895 "dma_device_id": "system", 00:20:24.895 "dma_device_type": 1 00:20:24.895 }, 00:20:24.895 { 00:20:24.895 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:24.895 "dma_device_type": 2 00:20:24.895 } 00:20:24.895 ], 00:20:24.895 "driver_specific": { 00:20:24.895 "raid": { 00:20:24.895 "uuid": "a09b7cb5-6c0e-46e7-ae81-cb1471a4bfea", 00:20:24.895 "strip_size_kb": 64, 00:20:24.895 "state": "online", 00:20:24.895 "raid_level": "concat", 00:20:24.895 "superblock": true, 00:20:24.895 "num_base_bdevs": 4, 00:20:24.895 "num_base_bdevs_discovered": 4, 00:20:24.895 "num_base_bdevs_operational": 4, 00:20:24.895 "base_bdevs_list": [ 00:20:24.895 { 00:20:24.895 "name": "NewBaseBdev", 00:20:24.895 "uuid": "98aa2c04-1df4-4b92-a864-887d0de265fe", 00:20:24.895 "is_configured": true, 00:20:24.895 "data_offset": 2048, 00:20:24.895 "data_size": 63488 00:20:24.895 }, 00:20:24.895 { 00:20:24.895 "name": "BaseBdev2", 00:20:24.895 "uuid": "8d6e3416-5b01-4367-a1f6-d092d3a176be", 00:20:24.895 "is_configured": true, 00:20:24.895 "data_offset": 2048, 00:20:24.895 "data_size": 63488 00:20:24.895 }, 00:20:24.895 { 00:20:24.895 "name": "BaseBdev3", 00:20:24.895 "uuid": "520833b6-eebb-4cbf-b605-5fd8c5fc6ee6", 00:20:24.895 "is_configured": true, 00:20:24.895 "data_offset": 2048, 00:20:24.895 "data_size": 63488 00:20:24.895 }, 00:20:24.895 { 00:20:24.895 "name": "BaseBdev4", 00:20:24.895 "uuid": "51adb667-52ff-4ed5-aa69-b267e6041e14", 00:20:24.895 "is_configured": true, 00:20:24.895 "data_offset": 2048, 00:20:24.895 "data_size": 63488 00:20:24.895 } 00:20:24.895 ] 00:20:24.895 } 00:20:24.895 } 00:20:24.895 }' 00:20:24.895 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:20:24.895 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # base_bdev_names='NewBaseBdev 00:20:24.895 BaseBdev2 00:20:24.895 BaseBdev3 00:20:24.895 BaseBdev4' 00:20:24.895 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:20:24.895 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev 00:20:24.895 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:20:25.153 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:20:25.153 "name": "NewBaseBdev", 00:20:25.153 "aliases": [ 00:20:25.153 "98aa2c04-1df4-4b92-a864-887d0de265fe" 00:20:25.153 ], 00:20:25.153 "product_name": "Malloc disk", 00:20:25.153 "block_size": 512, 00:20:25.153 "num_blocks": 65536, 00:20:25.153 "uuid": "98aa2c04-1df4-4b92-a864-887d0de265fe", 00:20:25.153 "assigned_rate_limits": { 00:20:25.153 "rw_ios_per_sec": 0, 00:20:25.153 "rw_mbytes_per_sec": 0, 00:20:25.153 "r_mbytes_per_sec": 0, 00:20:25.153 "w_mbytes_per_sec": 0 00:20:25.153 }, 00:20:25.153 "claimed": true, 00:20:25.153 "claim_type": "exclusive_write", 00:20:25.153 "zoned": false, 00:20:25.153 "supported_io_types": { 00:20:25.153 "read": true, 00:20:25.153 "write": true, 00:20:25.153 "unmap": true, 00:20:25.153 "flush": true, 00:20:25.153 "reset": true, 00:20:25.153 "nvme_admin": false, 00:20:25.153 "nvme_io": false, 00:20:25.153 "nvme_io_md": false, 00:20:25.153 "write_zeroes": true, 00:20:25.153 "zcopy": true, 00:20:25.153 "get_zone_info": false, 00:20:25.153 "zone_management": false, 00:20:25.153 "zone_append": false, 00:20:25.153 "compare": false, 00:20:25.153 "compare_and_write": false, 00:20:25.153 "abort": true, 00:20:25.153 "seek_hole": false, 00:20:25.153 "seek_data": false, 00:20:25.153 "copy": true, 00:20:25.153 "nvme_iov_md": false 00:20:25.153 }, 00:20:25.153 "memory_domains": [ 00:20:25.153 { 00:20:25.153 "dma_device_id": "system", 00:20:25.153 "dma_device_type": 1 00:20:25.153 }, 00:20:25.153 { 00:20:25.153 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:25.153 "dma_device_type": 2 00:20:25.153 } 00:20:25.153 ], 00:20:25.153 "driver_specific": {} 00:20:25.153 }' 00:20:25.153 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:25.153 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:25.153 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:20:25.153 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:25.153 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:25.153 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:20:25.153 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:25.153 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:25.411 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:20:25.411 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:25.411 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:25.411 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:20:25.411 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:20:25.411 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:20:25.411 11:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:20:25.977 11:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:20:25.977 "name": "BaseBdev2", 00:20:25.977 "aliases": [ 00:20:25.977 "8d6e3416-5b01-4367-a1f6-d092d3a176be" 00:20:25.977 ], 00:20:25.977 "product_name": "Malloc disk", 00:20:25.977 "block_size": 512, 00:20:25.977 "num_blocks": 65536, 00:20:25.977 "uuid": "8d6e3416-5b01-4367-a1f6-d092d3a176be", 00:20:25.977 "assigned_rate_limits": { 00:20:25.977 "rw_ios_per_sec": 0, 00:20:25.977 "rw_mbytes_per_sec": 0, 00:20:25.977 "r_mbytes_per_sec": 0, 00:20:25.977 "w_mbytes_per_sec": 0 00:20:25.977 }, 00:20:25.977 "claimed": true, 00:20:25.977 "claim_type": "exclusive_write", 00:20:25.977 "zoned": false, 00:20:25.977 "supported_io_types": { 00:20:25.977 "read": true, 00:20:25.977 "write": true, 00:20:25.977 "unmap": true, 00:20:25.977 "flush": true, 00:20:25.977 "reset": true, 00:20:25.977 "nvme_admin": false, 00:20:25.977 "nvme_io": false, 00:20:25.977 "nvme_io_md": false, 00:20:25.977 "write_zeroes": true, 00:20:25.977 "zcopy": true, 00:20:25.977 "get_zone_info": false, 00:20:25.977 "zone_management": false, 00:20:25.977 "zone_append": false, 00:20:25.977 "compare": false, 00:20:25.977 "compare_and_write": false, 00:20:25.977 "abort": true, 00:20:25.977 "seek_hole": false, 00:20:25.977 "seek_data": false, 00:20:25.977 "copy": true, 00:20:25.977 "nvme_iov_md": false 00:20:25.977 }, 00:20:25.977 "memory_domains": [ 00:20:25.977 { 00:20:25.977 "dma_device_id": "system", 00:20:25.977 "dma_device_type": 1 00:20:25.977 }, 00:20:25.977 { 00:20:25.977 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:25.977 "dma_device_type": 2 00:20:25.977 } 00:20:25.977 ], 00:20:25.977 "driver_specific": {} 00:20:25.977 }' 00:20:25.977 11:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:25.977 11:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:25.977 11:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:20:25.977 11:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:25.977 11:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:26.236 11:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:20:26.236 11:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:26.236 11:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:26.236 11:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:20:26.236 11:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:26.236 11:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:26.236 11:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:20:26.236 11:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:20:26.236 11:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:20:26.236 11:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:20:26.493 11:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:20:26.493 "name": "BaseBdev3", 00:20:26.493 "aliases": [ 00:20:26.493 "520833b6-eebb-4cbf-b605-5fd8c5fc6ee6" 00:20:26.493 ], 00:20:26.493 "product_name": "Malloc disk", 00:20:26.493 "block_size": 512, 00:20:26.493 "num_blocks": 65536, 00:20:26.493 "uuid": "520833b6-eebb-4cbf-b605-5fd8c5fc6ee6", 00:20:26.493 "assigned_rate_limits": { 00:20:26.493 "rw_ios_per_sec": 0, 00:20:26.493 "rw_mbytes_per_sec": 0, 00:20:26.493 "r_mbytes_per_sec": 0, 00:20:26.493 "w_mbytes_per_sec": 0 00:20:26.493 }, 00:20:26.493 "claimed": true, 00:20:26.493 "claim_type": "exclusive_write", 00:20:26.493 "zoned": false, 00:20:26.493 "supported_io_types": { 00:20:26.493 "read": true, 00:20:26.493 "write": true, 00:20:26.493 "unmap": true, 00:20:26.493 "flush": true, 00:20:26.493 "reset": true, 00:20:26.493 "nvme_admin": false, 00:20:26.493 "nvme_io": false, 00:20:26.493 "nvme_io_md": false, 00:20:26.493 "write_zeroes": true, 00:20:26.493 "zcopy": true, 00:20:26.493 "get_zone_info": false, 00:20:26.493 "zone_management": false, 00:20:26.493 "zone_append": false, 00:20:26.493 "compare": false, 00:20:26.493 "compare_and_write": false, 00:20:26.493 "abort": true, 00:20:26.493 "seek_hole": false, 00:20:26.493 "seek_data": false, 00:20:26.493 "copy": true, 00:20:26.493 "nvme_iov_md": false 00:20:26.493 }, 00:20:26.493 "memory_domains": [ 00:20:26.493 { 00:20:26.493 "dma_device_id": "system", 00:20:26.493 "dma_device_type": 1 00:20:26.493 }, 00:20:26.493 { 00:20:26.493 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:26.493 "dma_device_type": 2 00:20:26.493 } 00:20:26.493 ], 00:20:26.493 "driver_specific": {} 00:20:26.493 }' 00:20:26.493 11:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:26.493 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:26.493 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:20:26.493 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:26.750 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:26.750 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:20:26.750 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:26.750 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:26.750 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:20:26.750 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:26.750 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:27.008 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:20:27.008 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:20:27.008 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 00:20:27.008 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:20:27.008 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:20:27.008 "name": "BaseBdev4", 00:20:27.008 "aliases": [ 00:20:27.008 "51adb667-52ff-4ed5-aa69-b267e6041e14" 00:20:27.008 ], 00:20:27.008 "product_name": "Malloc disk", 00:20:27.008 "block_size": 512, 00:20:27.008 "num_blocks": 65536, 00:20:27.008 "uuid": "51adb667-52ff-4ed5-aa69-b267e6041e14", 00:20:27.008 "assigned_rate_limits": { 00:20:27.008 "rw_ios_per_sec": 0, 00:20:27.008 "rw_mbytes_per_sec": 0, 00:20:27.008 "r_mbytes_per_sec": 0, 00:20:27.008 "w_mbytes_per_sec": 0 00:20:27.008 }, 00:20:27.008 "claimed": true, 00:20:27.008 "claim_type": "exclusive_write", 00:20:27.008 "zoned": false, 00:20:27.008 "supported_io_types": { 00:20:27.008 "read": true, 00:20:27.008 "write": true, 00:20:27.008 "unmap": true, 00:20:27.008 "flush": true, 00:20:27.009 "reset": true, 00:20:27.009 "nvme_admin": false, 00:20:27.009 "nvme_io": false, 00:20:27.009 "nvme_io_md": false, 00:20:27.009 "write_zeroes": true, 00:20:27.009 "zcopy": true, 00:20:27.009 "get_zone_info": false, 00:20:27.009 "zone_management": false, 00:20:27.009 "zone_append": false, 00:20:27.009 "compare": false, 00:20:27.009 "compare_and_write": false, 00:20:27.009 "abort": true, 00:20:27.009 "seek_hole": false, 00:20:27.009 "seek_data": false, 00:20:27.009 "copy": true, 00:20:27.009 "nvme_iov_md": false 00:20:27.009 }, 00:20:27.009 "memory_domains": [ 00:20:27.009 { 00:20:27.009 "dma_device_id": "system", 00:20:27.009 "dma_device_type": 1 00:20:27.009 }, 00:20:27.009 { 00:20:27.009 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:27.009 "dma_device_type": 2 00:20:27.009 } 00:20:27.009 ], 00:20:27.009 "driver_specific": {} 00:20:27.009 }' 00:20:27.009 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:27.266 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:27.267 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:20:27.267 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:27.267 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:27.267 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:20:27.267 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:27.267 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:27.267 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:20:27.267 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:27.523 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:27.523 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:20:27.523 11:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@338 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:20:27.523 [2024-07-15 11:31:11.085339] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:20:27.523 [2024-07-15 11:31:11.085364] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:20:27.523 [2024-07-15 11:31:11.085417] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:20:27.523 [2024-07-15 11:31:11.085479] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:20:27.523 [2024-07-15 11:31:11.085491] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xcf4850 name Existed_Raid, state offline 00:20:27.523 11:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@341 -- # killprocess 942599 00:20:27.523 11:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@948 -- # '[' -z 942599 ']' 00:20:27.523 11:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@952 -- # kill -0 942599 00:20:27.523 11:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@953 -- # uname 00:20:27.523 11:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:20:27.523 11:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 942599 00:20:27.779 11:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:20:27.779 11:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:20:27.779 11:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@966 -- # echo 'killing process with pid 942599' 00:20:27.779 killing process with pid 942599 00:20:27.779 11:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@967 -- # kill 942599 00:20:27.779 [2024-07-15 11:31:11.153163] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:20:27.779 11:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # wait 942599 00:20:27.779 [2024-07-15 11:31:11.191340] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:20:28.037 11:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@343 -- # return 0 00:20:28.037 00:20:28.037 real 0m32.595s 00:20:28.037 user 0m59.877s 00:20:28.037 sys 0m5.767s 00:20:28.037 11:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1124 -- # xtrace_disable 00:20:28.038 11:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:28.038 ************************************ 00:20:28.038 END TEST raid_state_function_test_sb 00:20:28.038 ************************************ 00:20:28.038 11:31:11 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:20:28.038 11:31:11 bdev_raid -- bdev/bdev_raid.sh@869 -- # run_test raid_superblock_test raid_superblock_test concat 4 00:20:28.038 11:31:11 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:20:28.038 11:31:11 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:20:28.038 11:31:11 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:20:28.038 ************************************ 00:20:28.038 START TEST raid_superblock_test 00:20:28.038 ************************************ 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1123 -- # raid_superblock_test concat 4 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@392 -- # local raid_level=concat 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local num_base_bdevs=4 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # base_bdevs_malloc=() 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local base_bdevs_malloc 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_pt=() 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_pt 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt_uuid=() 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt_uuid 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local raid_bdev_name=raid_bdev1 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local strip_size 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size_create_arg 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local raid_bdev_uuid 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@403 -- # '[' concat '!=' raid1 ']' 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # strip_size=64 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size_create_arg='-z 64' 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # raid_pid=947546 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # waitforlisten 947546 /var/tmp/spdk-raid.sock 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@410 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -L bdev_raid 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@829 -- # '[' -z 947546 ']' 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:20:28.038 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:20:28.038 11:31:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:28.038 [2024-07-15 11:31:11.560558] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:20:28.038 [2024-07-15 11:31:11.560612] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid947546 ] 00:20:28.295 [2024-07-15 11:31:11.673121] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:28.295 [2024-07-15 11:31:11.770819] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:20:28.295 [2024-07-15 11:31:11.829161] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:28.295 [2024-07-15 11:31:11.829200] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:28.859 11:31:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:20:28.859 11:31:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@862 -- # return 0 00:20:28.859 11:31:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i = 1 )) 00:20:28.859 11:31:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:20:28.859 11:31:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc1 00:20:28.859 11:31:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt1 00:20:28.859 11:31:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:20:28.859 11:31:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:20:28.859 11:31:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:20:28.859 11:31:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:20:28.859 11:31:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc1 00:20:29.117 malloc1 00:20:29.117 11:31:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:20:29.374 [2024-07-15 11:31:12.916884] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:20:29.374 [2024-07-15 11:31:12.916938] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:29.374 [2024-07-15 11:31:12.916959] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1caf570 00:20:29.374 [2024-07-15 11:31:12.916971] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:29.374 [2024-07-15 11:31:12.918523] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:29.374 [2024-07-15 11:31:12.918551] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:20:29.374 pt1 00:20:29.374 11:31:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:20:29.374 11:31:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:20:29.374 11:31:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc2 00:20:29.374 11:31:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt2 00:20:29.374 11:31:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:20:29.374 11:31:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:20:29.374 11:31:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:20:29.374 11:31:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:20:29.374 11:31:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc2 00:20:29.631 malloc2 00:20:29.631 11:31:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:20:29.887 [2024-07-15 11:31:13.415016] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:20:29.887 [2024-07-15 11:31:13.415058] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:29.887 [2024-07-15 11:31:13.415075] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1cb0970 00:20:29.887 [2024-07-15 11:31:13.415087] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:29.887 [2024-07-15 11:31:13.416520] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:29.887 [2024-07-15 11:31:13.416552] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:20:29.887 pt2 00:20:29.887 11:31:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:20:29.887 11:31:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:20:29.887 11:31:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc3 00:20:29.887 11:31:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt3 00:20:29.887 11:31:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:20:29.887 11:31:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:20:29.887 11:31:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:20:29.887 11:31:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:20:29.887 11:31:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc3 00:20:30.144 malloc3 00:20:30.144 11:31:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:20:30.402 [2024-07-15 11:31:13.900780] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:20:30.402 [2024-07-15 11:31:13.900821] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:30.402 [2024-07-15 11:31:13.900838] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1e47340 00:20:30.402 [2024-07-15 11:31:13.900850] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:30.402 [2024-07-15 11:31:13.902223] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:30.402 [2024-07-15 11:31:13.902250] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:20:30.402 pt3 00:20:30.402 11:31:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:20:30.402 11:31:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:20:30.402 11:31:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc4 00:20:30.402 11:31:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt4 00:20:30.402 11:31:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:20:30.402 11:31:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:20:30.402 11:31:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:20:30.402 11:31:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:20:30.402 11:31:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc4 00:20:30.660 malloc4 00:20:30.660 11:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:20:31.225 [2024-07-15 11:31:14.651564] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:20:31.225 [2024-07-15 11:31:14.651617] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:31.225 [2024-07-15 11:31:14.651639] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1e49c60 00:20:31.225 [2024-07-15 11:31:14.651651] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:31.225 [2024-07-15 11:31:14.653276] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:31.225 [2024-07-15 11:31:14.653305] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:20:31.225 pt4 00:20:31.225 11:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:20:31.225 11:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:20:31.225 11:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@429 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'pt1 pt2 pt3 pt4' -n raid_bdev1 -s 00:20:31.483 [2024-07-15 11:31:14.908265] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:20:31.483 [2024-07-15 11:31:14.909617] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:20:31.483 [2024-07-15 11:31:14.909674] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:20:31.483 [2024-07-15 11:31:14.909717] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:20:31.483 [2024-07-15 11:31:14.909891] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1ca7530 00:20:31.483 [2024-07-15 11:31:14.909902] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:20:31.483 [2024-07-15 11:31:14.910126] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1ca5770 00:20:31.483 [2024-07-15 11:31:14.910280] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1ca7530 00:20:31.483 [2024-07-15 11:31:14.910290] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x1ca7530 00:20:31.483 [2024-07-15 11:31:14.910388] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:31.483 11:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:20:31.483 11:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:20:31.483 11:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:20:31.483 11:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:20:31.483 11:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:20:31.483 11:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:20:31.483 11:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:20:31.483 11:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:20:31.483 11:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:20:31.483 11:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:20:31.483 11:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:31.483 11:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:31.740 11:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:31.740 "name": "raid_bdev1", 00:20:31.740 "uuid": "41788dc0-696c-47e4-bdf7-43df2b427c2b", 00:20:31.740 "strip_size_kb": 64, 00:20:31.740 "state": "online", 00:20:31.740 "raid_level": "concat", 00:20:31.740 "superblock": true, 00:20:31.740 "num_base_bdevs": 4, 00:20:31.740 "num_base_bdevs_discovered": 4, 00:20:31.740 "num_base_bdevs_operational": 4, 00:20:31.740 "base_bdevs_list": [ 00:20:31.740 { 00:20:31.740 "name": "pt1", 00:20:31.740 "uuid": "00000000-0000-0000-0000-000000000001", 00:20:31.740 "is_configured": true, 00:20:31.740 "data_offset": 2048, 00:20:31.740 "data_size": 63488 00:20:31.740 }, 00:20:31.740 { 00:20:31.740 "name": "pt2", 00:20:31.740 "uuid": "00000000-0000-0000-0000-000000000002", 00:20:31.740 "is_configured": true, 00:20:31.740 "data_offset": 2048, 00:20:31.740 "data_size": 63488 00:20:31.740 }, 00:20:31.740 { 00:20:31.740 "name": "pt3", 00:20:31.740 "uuid": "00000000-0000-0000-0000-000000000003", 00:20:31.740 "is_configured": true, 00:20:31.740 "data_offset": 2048, 00:20:31.740 "data_size": 63488 00:20:31.740 }, 00:20:31.740 { 00:20:31.740 "name": "pt4", 00:20:31.740 "uuid": "00000000-0000-0000-0000-000000000004", 00:20:31.740 "is_configured": true, 00:20:31.740 "data_offset": 2048, 00:20:31.740 "data_size": 63488 00:20:31.740 } 00:20:31.740 ] 00:20:31.740 }' 00:20:31.740 11:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:31.740 11:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:32.303 11:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_properties raid_bdev1 00:20:32.303 11:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:20:32.303 11:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:20:32.303 11:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:20:32.303 11:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:20:32.303 11:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # local name 00:20:32.303 11:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:20:32.303 11:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:20:32.561 [2024-07-15 11:31:16.007435] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:20:32.561 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:20:32.561 "name": "raid_bdev1", 00:20:32.561 "aliases": [ 00:20:32.561 "41788dc0-696c-47e4-bdf7-43df2b427c2b" 00:20:32.561 ], 00:20:32.561 "product_name": "Raid Volume", 00:20:32.561 "block_size": 512, 00:20:32.561 "num_blocks": 253952, 00:20:32.561 "uuid": "41788dc0-696c-47e4-bdf7-43df2b427c2b", 00:20:32.561 "assigned_rate_limits": { 00:20:32.561 "rw_ios_per_sec": 0, 00:20:32.561 "rw_mbytes_per_sec": 0, 00:20:32.561 "r_mbytes_per_sec": 0, 00:20:32.561 "w_mbytes_per_sec": 0 00:20:32.561 }, 00:20:32.561 "claimed": false, 00:20:32.561 "zoned": false, 00:20:32.561 "supported_io_types": { 00:20:32.561 "read": true, 00:20:32.561 "write": true, 00:20:32.561 "unmap": true, 00:20:32.561 "flush": true, 00:20:32.561 "reset": true, 00:20:32.561 "nvme_admin": false, 00:20:32.561 "nvme_io": false, 00:20:32.561 "nvme_io_md": false, 00:20:32.561 "write_zeroes": true, 00:20:32.561 "zcopy": false, 00:20:32.561 "get_zone_info": false, 00:20:32.561 "zone_management": false, 00:20:32.561 "zone_append": false, 00:20:32.561 "compare": false, 00:20:32.561 "compare_and_write": false, 00:20:32.561 "abort": false, 00:20:32.561 "seek_hole": false, 00:20:32.561 "seek_data": false, 00:20:32.561 "copy": false, 00:20:32.561 "nvme_iov_md": false 00:20:32.562 }, 00:20:32.562 "memory_domains": [ 00:20:32.562 { 00:20:32.562 "dma_device_id": "system", 00:20:32.562 "dma_device_type": 1 00:20:32.562 }, 00:20:32.562 { 00:20:32.562 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:32.562 "dma_device_type": 2 00:20:32.562 }, 00:20:32.562 { 00:20:32.562 "dma_device_id": "system", 00:20:32.562 "dma_device_type": 1 00:20:32.562 }, 00:20:32.562 { 00:20:32.562 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:32.562 "dma_device_type": 2 00:20:32.562 }, 00:20:32.562 { 00:20:32.562 "dma_device_id": "system", 00:20:32.562 "dma_device_type": 1 00:20:32.562 }, 00:20:32.562 { 00:20:32.562 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:32.562 "dma_device_type": 2 00:20:32.562 }, 00:20:32.562 { 00:20:32.562 "dma_device_id": "system", 00:20:32.562 "dma_device_type": 1 00:20:32.562 }, 00:20:32.562 { 00:20:32.562 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:32.562 "dma_device_type": 2 00:20:32.562 } 00:20:32.562 ], 00:20:32.562 "driver_specific": { 00:20:32.562 "raid": { 00:20:32.562 "uuid": "41788dc0-696c-47e4-bdf7-43df2b427c2b", 00:20:32.562 "strip_size_kb": 64, 00:20:32.562 "state": "online", 00:20:32.562 "raid_level": "concat", 00:20:32.562 "superblock": true, 00:20:32.562 "num_base_bdevs": 4, 00:20:32.562 "num_base_bdevs_discovered": 4, 00:20:32.562 "num_base_bdevs_operational": 4, 00:20:32.562 "base_bdevs_list": [ 00:20:32.562 { 00:20:32.562 "name": "pt1", 00:20:32.562 "uuid": "00000000-0000-0000-0000-000000000001", 00:20:32.562 "is_configured": true, 00:20:32.562 "data_offset": 2048, 00:20:32.562 "data_size": 63488 00:20:32.562 }, 00:20:32.562 { 00:20:32.562 "name": "pt2", 00:20:32.562 "uuid": "00000000-0000-0000-0000-000000000002", 00:20:32.562 "is_configured": true, 00:20:32.562 "data_offset": 2048, 00:20:32.562 "data_size": 63488 00:20:32.562 }, 00:20:32.562 { 00:20:32.562 "name": "pt3", 00:20:32.562 "uuid": "00000000-0000-0000-0000-000000000003", 00:20:32.562 "is_configured": true, 00:20:32.562 "data_offset": 2048, 00:20:32.562 "data_size": 63488 00:20:32.562 }, 00:20:32.562 { 00:20:32.562 "name": "pt4", 00:20:32.562 "uuid": "00000000-0000-0000-0000-000000000004", 00:20:32.562 "is_configured": true, 00:20:32.562 "data_offset": 2048, 00:20:32.562 "data_size": 63488 00:20:32.562 } 00:20:32.562 ] 00:20:32.562 } 00:20:32.562 } 00:20:32.562 }' 00:20:32.562 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:20:32.562 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:20:32.562 pt2 00:20:32.562 pt3 00:20:32.562 pt4' 00:20:32.562 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:20:32.562 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:20:32.562 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:20:32.819 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:20:32.819 "name": "pt1", 00:20:32.820 "aliases": [ 00:20:32.820 "00000000-0000-0000-0000-000000000001" 00:20:32.820 ], 00:20:32.820 "product_name": "passthru", 00:20:32.820 "block_size": 512, 00:20:32.820 "num_blocks": 65536, 00:20:32.820 "uuid": "00000000-0000-0000-0000-000000000001", 00:20:32.820 "assigned_rate_limits": { 00:20:32.820 "rw_ios_per_sec": 0, 00:20:32.820 "rw_mbytes_per_sec": 0, 00:20:32.820 "r_mbytes_per_sec": 0, 00:20:32.820 "w_mbytes_per_sec": 0 00:20:32.820 }, 00:20:32.820 "claimed": true, 00:20:32.820 "claim_type": "exclusive_write", 00:20:32.820 "zoned": false, 00:20:32.820 "supported_io_types": { 00:20:32.820 "read": true, 00:20:32.820 "write": true, 00:20:32.820 "unmap": true, 00:20:32.820 "flush": true, 00:20:32.820 "reset": true, 00:20:32.820 "nvme_admin": false, 00:20:32.820 "nvme_io": false, 00:20:32.820 "nvme_io_md": false, 00:20:32.820 "write_zeroes": true, 00:20:32.820 "zcopy": true, 00:20:32.820 "get_zone_info": false, 00:20:32.820 "zone_management": false, 00:20:32.820 "zone_append": false, 00:20:32.820 "compare": false, 00:20:32.820 "compare_and_write": false, 00:20:32.820 "abort": true, 00:20:32.820 "seek_hole": false, 00:20:32.820 "seek_data": false, 00:20:32.820 "copy": true, 00:20:32.820 "nvme_iov_md": false 00:20:32.820 }, 00:20:32.820 "memory_domains": [ 00:20:32.820 { 00:20:32.820 "dma_device_id": "system", 00:20:32.820 "dma_device_type": 1 00:20:32.820 }, 00:20:32.820 { 00:20:32.820 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:32.820 "dma_device_type": 2 00:20:32.820 } 00:20:32.820 ], 00:20:32.820 "driver_specific": { 00:20:32.820 "passthru": { 00:20:32.820 "name": "pt1", 00:20:32.820 "base_bdev_name": "malloc1" 00:20:32.820 } 00:20:32.820 } 00:20:32.820 }' 00:20:32.820 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:32.820 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:32.820 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:20:32.820 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:32.820 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:33.078 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:20:33.078 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:33.078 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:33.078 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:20:33.078 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:33.078 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:33.078 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:20:33.078 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:20:33.078 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:20:33.078 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:20:33.335 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:20:33.335 "name": "pt2", 00:20:33.335 "aliases": [ 00:20:33.335 "00000000-0000-0000-0000-000000000002" 00:20:33.335 ], 00:20:33.335 "product_name": "passthru", 00:20:33.335 "block_size": 512, 00:20:33.335 "num_blocks": 65536, 00:20:33.335 "uuid": "00000000-0000-0000-0000-000000000002", 00:20:33.335 "assigned_rate_limits": { 00:20:33.335 "rw_ios_per_sec": 0, 00:20:33.335 "rw_mbytes_per_sec": 0, 00:20:33.335 "r_mbytes_per_sec": 0, 00:20:33.335 "w_mbytes_per_sec": 0 00:20:33.335 }, 00:20:33.335 "claimed": true, 00:20:33.335 "claim_type": "exclusive_write", 00:20:33.335 "zoned": false, 00:20:33.336 "supported_io_types": { 00:20:33.336 "read": true, 00:20:33.336 "write": true, 00:20:33.336 "unmap": true, 00:20:33.336 "flush": true, 00:20:33.336 "reset": true, 00:20:33.336 "nvme_admin": false, 00:20:33.336 "nvme_io": false, 00:20:33.336 "nvme_io_md": false, 00:20:33.336 "write_zeroes": true, 00:20:33.336 "zcopy": true, 00:20:33.336 "get_zone_info": false, 00:20:33.336 "zone_management": false, 00:20:33.336 "zone_append": false, 00:20:33.336 "compare": false, 00:20:33.336 "compare_and_write": false, 00:20:33.336 "abort": true, 00:20:33.336 "seek_hole": false, 00:20:33.336 "seek_data": false, 00:20:33.336 "copy": true, 00:20:33.336 "nvme_iov_md": false 00:20:33.336 }, 00:20:33.336 "memory_domains": [ 00:20:33.336 { 00:20:33.336 "dma_device_id": "system", 00:20:33.336 "dma_device_type": 1 00:20:33.336 }, 00:20:33.336 { 00:20:33.336 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:33.336 "dma_device_type": 2 00:20:33.336 } 00:20:33.336 ], 00:20:33.336 "driver_specific": { 00:20:33.336 "passthru": { 00:20:33.336 "name": "pt2", 00:20:33.336 "base_bdev_name": "malloc2" 00:20:33.336 } 00:20:33.336 } 00:20:33.336 }' 00:20:33.336 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:33.336 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:33.336 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:20:33.336 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:33.593 11:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:33.593 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:20:33.593 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:33.593 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:33.593 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:20:33.593 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:33.593 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:33.593 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:20:33.593 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:20:33.593 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt3 00:20:33.593 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:20:33.850 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:20:33.850 "name": "pt3", 00:20:33.850 "aliases": [ 00:20:33.850 "00000000-0000-0000-0000-000000000003" 00:20:33.850 ], 00:20:33.850 "product_name": "passthru", 00:20:33.850 "block_size": 512, 00:20:33.850 "num_blocks": 65536, 00:20:33.850 "uuid": "00000000-0000-0000-0000-000000000003", 00:20:33.850 "assigned_rate_limits": { 00:20:33.850 "rw_ios_per_sec": 0, 00:20:33.850 "rw_mbytes_per_sec": 0, 00:20:33.850 "r_mbytes_per_sec": 0, 00:20:33.850 "w_mbytes_per_sec": 0 00:20:33.850 }, 00:20:33.850 "claimed": true, 00:20:33.850 "claim_type": "exclusive_write", 00:20:33.850 "zoned": false, 00:20:33.850 "supported_io_types": { 00:20:33.850 "read": true, 00:20:33.850 "write": true, 00:20:33.850 "unmap": true, 00:20:33.850 "flush": true, 00:20:33.850 "reset": true, 00:20:33.850 "nvme_admin": false, 00:20:33.850 "nvme_io": false, 00:20:33.850 "nvme_io_md": false, 00:20:33.850 "write_zeroes": true, 00:20:33.850 "zcopy": true, 00:20:33.850 "get_zone_info": false, 00:20:33.850 "zone_management": false, 00:20:33.850 "zone_append": false, 00:20:33.850 "compare": false, 00:20:33.850 "compare_and_write": false, 00:20:33.850 "abort": true, 00:20:33.850 "seek_hole": false, 00:20:33.850 "seek_data": false, 00:20:33.850 "copy": true, 00:20:33.850 "nvme_iov_md": false 00:20:33.850 }, 00:20:33.850 "memory_domains": [ 00:20:33.850 { 00:20:33.850 "dma_device_id": "system", 00:20:33.850 "dma_device_type": 1 00:20:33.850 }, 00:20:33.850 { 00:20:33.850 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:33.850 "dma_device_type": 2 00:20:33.850 } 00:20:33.850 ], 00:20:33.850 "driver_specific": { 00:20:33.850 "passthru": { 00:20:33.850 "name": "pt3", 00:20:33.850 "base_bdev_name": "malloc3" 00:20:33.850 } 00:20:33.850 } 00:20:33.850 }' 00:20:33.850 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:33.850 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:34.107 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:20:34.107 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:34.107 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:34.107 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:20:34.107 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:34.107 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:34.107 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:20:34.107 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:34.107 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:34.365 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:20:34.365 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:20:34.365 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:20:34.365 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt4 00:20:34.690 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:20:34.690 "name": "pt4", 00:20:34.690 "aliases": [ 00:20:34.690 "00000000-0000-0000-0000-000000000004" 00:20:34.690 ], 00:20:34.690 "product_name": "passthru", 00:20:34.690 "block_size": 512, 00:20:34.690 "num_blocks": 65536, 00:20:34.690 "uuid": "00000000-0000-0000-0000-000000000004", 00:20:34.690 "assigned_rate_limits": { 00:20:34.690 "rw_ios_per_sec": 0, 00:20:34.690 "rw_mbytes_per_sec": 0, 00:20:34.690 "r_mbytes_per_sec": 0, 00:20:34.690 "w_mbytes_per_sec": 0 00:20:34.690 }, 00:20:34.690 "claimed": true, 00:20:34.690 "claim_type": "exclusive_write", 00:20:34.690 "zoned": false, 00:20:34.690 "supported_io_types": { 00:20:34.690 "read": true, 00:20:34.690 "write": true, 00:20:34.690 "unmap": true, 00:20:34.690 "flush": true, 00:20:34.690 "reset": true, 00:20:34.690 "nvme_admin": false, 00:20:34.690 "nvme_io": false, 00:20:34.690 "nvme_io_md": false, 00:20:34.690 "write_zeroes": true, 00:20:34.690 "zcopy": true, 00:20:34.690 "get_zone_info": false, 00:20:34.690 "zone_management": false, 00:20:34.690 "zone_append": false, 00:20:34.690 "compare": false, 00:20:34.690 "compare_and_write": false, 00:20:34.690 "abort": true, 00:20:34.690 "seek_hole": false, 00:20:34.690 "seek_data": false, 00:20:34.690 "copy": true, 00:20:34.690 "nvme_iov_md": false 00:20:34.690 }, 00:20:34.690 "memory_domains": [ 00:20:34.690 { 00:20:34.690 "dma_device_id": "system", 00:20:34.690 "dma_device_type": 1 00:20:34.690 }, 00:20:34.690 { 00:20:34.690 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:34.690 "dma_device_type": 2 00:20:34.690 } 00:20:34.690 ], 00:20:34.690 "driver_specific": { 00:20:34.690 "passthru": { 00:20:34.690 "name": "pt4", 00:20:34.690 "base_bdev_name": "malloc4" 00:20:34.690 } 00:20:34.690 } 00:20:34.690 }' 00:20:34.690 11:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:34.690 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:34.690 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:20:34.690 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:34.690 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:34.690 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:20:34.690 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:34.690 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:34.690 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:20:34.690 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:34.690 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:34.947 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:20:34.947 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:20:34.947 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # jq -r '.[] | .uuid' 00:20:34.947 [2024-07-15 11:31:18.473961] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:20:34.947 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # raid_bdev_uuid=41788dc0-696c-47e4-bdf7-43df2b427c2b 00:20:34.947 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # '[' -z 41788dc0-696c-47e4-bdf7-43df2b427c2b ']' 00:20:34.947 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@440 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:20:35.204 [2024-07-15 11:31:18.722310] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:20:35.204 [2024-07-15 11:31:18.722338] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:20:35.204 [2024-07-15 11:31:18.722399] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:20:35.204 [2024-07-15 11:31:18.722464] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:20:35.204 [2024-07-15 11:31:18.722476] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1ca7530 name raid_bdev1, state offline 00:20:35.204 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:35.204 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # jq -r '.[]' 00:20:35.461 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # raid_bdev= 00:20:35.461 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # '[' -n '' ']' 00:20:35.461 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:20:35.461 11:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt1 00:20:35.717 11:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:20:35.717 11:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:20:35.974 11:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:20:35.974 11:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt3 00:20:36.231 11:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:20:36.231 11:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt4 00:20:36.488 11:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs 00:20:36.488 11:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:20:36.745 11:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # '[' false == true ']' 00:20:36.745 11:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@456 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'malloc1 malloc2 malloc3 malloc4' -n raid_bdev1 00:20:36.745 11:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@648 -- # local es=0 00:20:36.745 11:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'malloc1 malloc2 malloc3 malloc4' -n raid_bdev1 00:20:36.745 11:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:20:36.745 11:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:20:36.745 11:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:20:36.745 11:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:20:36.745 11:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:20:36.745 11:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:20:36.745 11:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:20:36.745 11:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py ]] 00:20:36.745 11:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'malloc1 malloc2 malloc3 malloc4' -n raid_bdev1 00:20:37.002 [2024-07-15 11:31:20.434771] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:20:37.002 [2024-07-15 11:31:20.436176] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:20:37.002 [2024-07-15 11:31:20.436223] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:20:37.002 [2024-07-15 11:31:20.436256] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:20:37.002 [2024-07-15 11:31:20.436303] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:20:37.002 [2024-07-15 11:31:20.436350] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:20:37.002 [2024-07-15 11:31:20.436373] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:20:37.002 [2024-07-15 11:31:20.436395] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:20:37.002 [2024-07-15 11:31:20.436413] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:20:37.002 [2024-07-15 11:31:20.436424] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1e52ff0 name raid_bdev1, state configuring 00:20:37.002 request: 00:20:37.002 { 00:20:37.002 "name": "raid_bdev1", 00:20:37.002 "raid_level": "concat", 00:20:37.002 "base_bdevs": [ 00:20:37.002 "malloc1", 00:20:37.002 "malloc2", 00:20:37.002 "malloc3", 00:20:37.002 "malloc4" 00:20:37.002 ], 00:20:37.002 "strip_size_kb": 64, 00:20:37.002 "superblock": false, 00:20:37.002 "method": "bdev_raid_create", 00:20:37.002 "req_id": 1 00:20:37.002 } 00:20:37.002 Got JSON-RPC error response 00:20:37.002 response: 00:20:37.002 { 00:20:37.002 "code": -17, 00:20:37.002 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:20:37.002 } 00:20:37.002 11:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@651 -- # es=1 00:20:37.002 11:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:20:37.002 11:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:20:37.002 11:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:20:37.002 11:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:37.002 11:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # jq -r '.[]' 00:20:37.260 11:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # raid_bdev= 00:20:37.260 11:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # '[' -n '' ']' 00:20:37.260 11:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@464 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:20:37.517 [2024-07-15 11:31:20.920131] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:20:37.517 [2024-07-15 11:31:20.920182] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:37.517 [2024-07-15 11:31:20.920204] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1caf7a0 00:20:37.517 [2024-07-15 11:31:20.920217] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:37.517 [2024-07-15 11:31:20.921891] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:37.517 [2024-07-15 11:31:20.921921] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:20:37.517 [2024-07-15 11:31:20.922004] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:20:37.517 [2024-07-15 11:31:20.922035] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:20:37.517 pt1 00:20:37.517 11:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@467 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:20:37.517 11:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:20:37.517 11:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:20:37.517 11:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:20:37.517 11:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:20:37.517 11:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:20:37.517 11:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:20:37.517 11:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:20:37.517 11:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:20:37.517 11:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:20:37.517 11:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:37.517 11:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:37.774 11:31:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:37.774 "name": "raid_bdev1", 00:20:37.774 "uuid": "41788dc0-696c-47e4-bdf7-43df2b427c2b", 00:20:37.774 "strip_size_kb": 64, 00:20:37.774 "state": "configuring", 00:20:37.774 "raid_level": "concat", 00:20:37.774 "superblock": true, 00:20:37.774 "num_base_bdevs": 4, 00:20:37.774 "num_base_bdevs_discovered": 1, 00:20:37.774 "num_base_bdevs_operational": 4, 00:20:37.774 "base_bdevs_list": [ 00:20:37.774 { 00:20:37.774 "name": "pt1", 00:20:37.774 "uuid": "00000000-0000-0000-0000-000000000001", 00:20:37.774 "is_configured": true, 00:20:37.774 "data_offset": 2048, 00:20:37.774 "data_size": 63488 00:20:37.774 }, 00:20:37.774 { 00:20:37.774 "name": null, 00:20:37.774 "uuid": "00000000-0000-0000-0000-000000000002", 00:20:37.774 "is_configured": false, 00:20:37.774 "data_offset": 2048, 00:20:37.774 "data_size": 63488 00:20:37.774 }, 00:20:37.774 { 00:20:37.774 "name": null, 00:20:37.774 "uuid": "00000000-0000-0000-0000-000000000003", 00:20:37.774 "is_configured": false, 00:20:37.774 "data_offset": 2048, 00:20:37.774 "data_size": 63488 00:20:37.774 }, 00:20:37.774 { 00:20:37.774 "name": null, 00:20:37.774 "uuid": "00000000-0000-0000-0000-000000000004", 00:20:37.774 "is_configured": false, 00:20:37.774 "data_offset": 2048, 00:20:37.774 "data_size": 63488 00:20:37.774 } 00:20:37.774 ] 00:20:37.774 }' 00:20:37.774 11:31:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:37.774 11:31:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:38.337 11:31:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@469 -- # '[' 4 -gt 2 ']' 00:20:38.337 11:31:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@471 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:20:38.594 [2024-07-15 11:31:21.942851] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:20:38.594 [2024-07-15 11:31:21.942900] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:38.594 [2024-07-15 11:31:21.942918] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1ca6ea0 00:20:38.594 [2024-07-15 11:31:21.942938] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:38.594 [2024-07-15 11:31:21.943290] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:38.594 [2024-07-15 11:31:21.943308] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:20:38.594 [2024-07-15 11:31:21.943371] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:20:38.594 [2024-07-15 11:31:21.943391] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:20:38.594 pt2 00:20:38.594 11:31:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:20:38.594 [2024-07-15 11:31:22.131349] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:20:38.594 11:31:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:20:38.594 11:31:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:20:38.594 11:31:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:20:38.594 11:31:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:20:38.594 11:31:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:20:38.594 11:31:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:20:38.594 11:31:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:20:38.594 11:31:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:20:38.594 11:31:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:20:38.594 11:31:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:20:38.594 11:31:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:38.594 11:31:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:38.851 11:31:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:38.851 "name": "raid_bdev1", 00:20:38.851 "uuid": "41788dc0-696c-47e4-bdf7-43df2b427c2b", 00:20:38.851 "strip_size_kb": 64, 00:20:38.851 "state": "configuring", 00:20:38.851 "raid_level": "concat", 00:20:38.851 "superblock": true, 00:20:38.851 "num_base_bdevs": 4, 00:20:38.851 "num_base_bdevs_discovered": 1, 00:20:38.851 "num_base_bdevs_operational": 4, 00:20:38.851 "base_bdevs_list": [ 00:20:38.851 { 00:20:38.851 "name": "pt1", 00:20:38.851 "uuid": "00000000-0000-0000-0000-000000000001", 00:20:38.851 "is_configured": true, 00:20:38.851 "data_offset": 2048, 00:20:38.851 "data_size": 63488 00:20:38.851 }, 00:20:38.851 { 00:20:38.851 "name": null, 00:20:38.851 "uuid": "00000000-0000-0000-0000-000000000002", 00:20:38.851 "is_configured": false, 00:20:38.851 "data_offset": 2048, 00:20:38.851 "data_size": 63488 00:20:38.851 }, 00:20:38.851 { 00:20:38.851 "name": null, 00:20:38.851 "uuid": "00000000-0000-0000-0000-000000000003", 00:20:38.851 "is_configured": false, 00:20:38.851 "data_offset": 2048, 00:20:38.851 "data_size": 63488 00:20:38.851 }, 00:20:38.851 { 00:20:38.851 "name": null, 00:20:38.851 "uuid": "00000000-0000-0000-0000-000000000004", 00:20:38.851 "is_configured": false, 00:20:38.851 "data_offset": 2048, 00:20:38.851 "data_size": 63488 00:20:38.851 } 00:20:38.851 ] 00:20:38.851 }' 00:20:38.851 11:31:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:38.851 11:31:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:39.417 11:31:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i = 1 )) 00:20:39.417 11:31:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:20:39.417 11:31:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:20:39.675 [2024-07-15 11:31:23.222226] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:20:39.675 [2024-07-15 11:31:23.222276] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:39.675 [2024-07-15 11:31:23.222294] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1ca5ec0 00:20:39.675 [2024-07-15 11:31:23.222306] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:39.675 [2024-07-15 11:31:23.222654] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:39.675 [2024-07-15 11:31:23.222672] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:20:39.675 [2024-07-15 11:31:23.222734] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:20:39.675 [2024-07-15 11:31:23.222754] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:20:39.675 pt2 00:20:39.675 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:20:39.675 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:20:39.675 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:20:39.934 [2024-07-15 11:31:23.466871] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:20:39.934 [2024-07-15 11:31:23.466910] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:39.934 [2024-07-15 11:31:23.466932] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1ca60f0 00:20:39.934 [2024-07-15 11:31:23.466944] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:39.934 [2024-07-15 11:31:23.467248] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:39.934 [2024-07-15 11:31:23.467266] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:20:39.934 [2024-07-15 11:31:23.467319] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:20:39.934 [2024-07-15 11:31:23.467337] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:20:39.934 pt3 00:20:39.934 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:20:39.934 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:20:39.934 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:20:40.192 [2024-07-15 11:31:23.711514] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:20:40.192 [2024-07-15 11:31:23.711555] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:40.192 [2024-07-15 11:31:23.711571] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1caeaf0 00:20:40.192 [2024-07-15 11:31:23.711583] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:40.192 [2024-07-15 11:31:23.711886] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:40.192 [2024-07-15 11:31:23.711903] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:20:40.192 [2024-07-15 11:31:23.711966] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:20:40.192 [2024-07-15 11:31:23.711985] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:20:40.192 [2024-07-15 11:31:23.712106] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1ca88f0 00:20:40.192 [2024-07-15 11:31:23.712117] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:20:40.192 [2024-07-15 11:31:23.712286] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1ca8150 00:20:40.192 [2024-07-15 11:31:23.712413] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1ca88f0 00:20:40.192 [2024-07-15 11:31:23.712423] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x1ca88f0 00:20:40.192 [2024-07-15 11:31:23.712518] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:40.192 pt4 00:20:40.192 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:20:40.192 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:20:40.192 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@482 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:20:40.192 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:20:40.192 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:20:40.192 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:20:40.192 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:20:40.192 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:20:40.192 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:20:40.192 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:20:40.192 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:20:40.192 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:20:40.192 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:40.192 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:40.450 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:40.450 "name": "raid_bdev1", 00:20:40.450 "uuid": "41788dc0-696c-47e4-bdf7-43df2b427c2b", 00:20:40.450 "strip_size_kb": 64, 00:20:40.450 "state": "online", 00:20:40.450 "raid_level": "concat", 00:20:40.450 "superblock": true, 00:20:40.450 "num_base_bdevs": 4, 00:20:40.450 "num_base_bdevs_discovered": 4, 00:20:40.450 "num_base_bdevs_operational": 4, 00:20:40.450 "base_bdevs_list": [ 00:20:40.450 { 00:20:40.450 "name": "pt1", 00:20:40.450 "uuid": "00000000-0000-0000-0000-000000000001", 00:20:40.450 "is_configured": true, 00:20:40.450 "data_offset": 2048, 00:20:40.450 "data_size": 63488 00:20:40.450 }, 00:20:40.450 { 00:20:40.450 "name": "pt2", 00:20:40.450 "uuid": "00000000-0000-0000-0000-000000000002", 00:20:40.450 "is_configured": true, 00:20:40.450 "data_offset": 2048, 00:20:40.451 "data_size": 63488 00:20:40.451 }, 00:20:40.451 { 00:20:40.451 "name": "pt3", 00:20:40.451 "uuid": "00000000-0000-0000-0000-000000000003", 00:20:40.451 "is_configured": true, 00:20:40.451 "data_offset": 2048, 00:20:40.451 "data_size": 63488 00:20:40.451 }, 00:20:40.451 { 00:20:40.451 "name": "pt4", 00:20:40.451 "uuid": "00000000-0000-0000-0000-000000000004", 00:20:40.451 "is_configured": true, 00:20:40.451 "data_offset": 2048, 00:20:40.451 "data_size": 63488 00:20:40.451 } 00:20:40.451 ] 00:20:40.451 }' 00:20:40.451 11:31:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:40.451 11:31:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:41.016 11:31:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_properties raid_bdev1 00:20:41.016 11:31:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:20:41.016 11:31:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:20:41.016 11:31:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:20:41.016 11:31:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:20:41.016 11:31:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # local name 00:20:41.016 11:31:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:20:41.016 11:31:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:20:41.274 [2024-07-15 11:31:24.746578] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:20:41.274 11:31:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:20:41.274 "name": "raid_bdev1", 00:20:41.274 "aliases": [ 00:20:41.274 "41788dc0-696c-47e4-bdf7-43df2b427c2b" 00:20:41.274 ], 00:20:41.274 "product_name": "Raid Volume", 00:20:41.274 "block_size": 512, 00:20:41.274 "num_blocks": 253952, 00:20:41.274 "uuid": "41788dc0-696c-47e4-bdf7-43df2b427c2b", 00:20:41.274 "assigned_rate_limits": { 00:20:41.274 "rw_ios_per_sec": 0, 00:20:41.274 "rw_mbytes_per_sec": 0, 00:20:41.274 "r_mbytes_per_sec": 0, 00:20:41.274 "w_mbytes_per_sec": 0 00:20:41.274 }, 00:20:41.274 "claimed": false, 00:20:41.274 "zoned": false, 00:20:41.274 "supported_io_types": { 00:20:41.274 "read": true, 00:20:41.274 "write": true, 00:20:41.274 "unmap": true, 00:20:41.274 "flush": true, 00:20:41.274 "reset": true, 00:20:41.274 "nvme_admin": false, 00:20:41.274 "nvme_io": false, 00:20:41.274 "nvme_io_md": false, 00:20:41.274 "write_zeroes": true, 00:20:41.274 "zcopy": false, 00:20:41.274 "get_zone_info": false, 00:20:41.274 "zone_management": false, 00:20:41.274 "zone_append": false, 00:20:41.274 "compare": false, 00:20:41.274 "compare_and_write": false, 00:20:41.274 "abort": false, 00:20:41.274 "seek_hole": false, 00:20:41.274 "seek_data": false, 00:20:41.274 "copy": false, 00:20:41.274 "nvme_iov_md": false 00:20:41.274 }, 00:20:41.274 "memory_domains": [ 00:20:41.274 { 00:20:41.274 "dma_device_id": "system", 00:20:41.274 "dma_device_type": 1 00:20:41.274 }, 00:20:41.274 { 00:20:41.275 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:41.275 "dma_device_type": 2 00:20:41.275 }, 00:20:41.275 { 00:20:41.275 "dma_device_id": "system", 00:20:41.275 "dma_device_type": 1 00:20:41.275 }, 00:20:41.275 { 00:20:41.275 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:41.275 "dma_device_type": 2 00:20:41.275 }, 00:20:41.275 { 00:20:41.275 "dma_device_id": "system", 00:20:41.275 "dma_device_type": 1 00:20:41.275 }, 00:20:41.275 { 00:20:41.275 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:41.275 "dma_device_type": 2 00:20:41.275 }, 00:20:41.275 { 00:20:41.275 "dma_device_id": "system", 00:20:41.275 "dma_device_type": 1 00:20:41.275 }, 00:20:41.275 { 00:20:41.275 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:41.275 "dma_device_type": 2 00:20:41.275 } 00:20:41.275 ], 00:20:41.275 "driver_specific": { 00:20:41.275 "raid": { 00:20:41.275 "uuid": "41788dc0-696c-47e4-bdf7-43df2b427c2b", 00:20:41.275 "strip_size_kb": 64, 00:20:41.275 "state": "online", 00:20:41.275 "raid_level": "concat", 00:20:41.275 "superblock": true, 00:20:41.275 "num_base_bdevs": 4, 00:20:41.275 "num_base_bdevs_discovered": 4, 00:20:41.275 "num_base_bdevs_operational": 4, 00:20:41.275 "base_bdevs_list": [ 00:20:41.275 { 00:20:41.275 "name": "pt1", 00:20:41.275 "uuid": "00000000-0000-0000-0000-000000000001", 00:20:41.275 "is_configured": true, 00:20:41.275 "data_offset": 2048, 00:20:41.275 "data_size": 63488 00:20:41.275 }, 00:20:41.275 { 00:20:41.275 "name": "pt2", 00:20:41.275 "uuid": "00000000-0000-0000-0000-000000000002", 00:20:41.275 "is_configured": true, 00:20:41.275 "data_offset": 2048, 00:20:41.275 "data_size": 63488 00:20:41.275 }, 00:20:41.275 { 00:20:41.275 "name": "pt3", 00:20:41.275 "uuid": "00000000-0000-0000-0000-000000000003", 00:20:41.275 "is_configured": true, 00:20:41.275 "data_offset": 2048, 00:20:41.275 "data_size": 63488 00:20:41.275 }, 00:20:41.275 { 00:20:41.275 "name": "pt4", 00:20:41.275 "uuid": "00000000-0000-0000-0000-000000000004", 00:20:41.275 "is_configured": true, 00:20:41.275 "data_offset": 2048, 00:20:41.275 "data_size": 63488 00:20:41.275 } 00:20:41.275 ] 00:20:41.275 } 00:20:41.275 } 00:20:41.275 }' 00:20:41.275 11:31:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:20:41.275 11:31:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:20:41.275 pt2 00:20:41.275 pt3 00:20:41.275 pt4' 00:20:41.275 11:31:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:20:41.275 11:31:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:20:41.275 11:31:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:20:41.533 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:20:41.533 "name": "pt1", 00:20:41.533 "aliases": [ 00:20:41.533 "00000000-0000-0000-0000-000000000001" 00:20:41.533 ], 00:20:41.533 "product_name": "passthru", 00:20:41.533 "block_size": 512, 00:20:41.533 "num_blocks": 65536, 00:20:41.533 "uuid": "00000000-0000-0000-0000-000000000001", 00:20:41.533 "assigned_rate_limits": { 00:20:41.533 "rw_ios_per_sec": 0, 00:20:41.533 "rw_mbytes_per_sec": 0, 00:20:41.533 "r_mbytes_per_sec": 0, 00:20:41.533 "w_mbytes_per_sec": 0 00:20:41.533 }, 00:20:41.533 "claimed": true, 00:20:41.533 "claim_type": "exclusive_write", 00:20:41.533 "zoned": false, 00:20:41.533 "supported_io_types": { 00:20:41.533 "read": true, 00:20:41.533 "write": true, 00:20:41.533 "unmap": true, 00:20:41.533 "flush": true, 00:20:41.533 "reset": true, 00:20:41.533 "nvme_admin": false, 00:20:41.533 "nvme_io": false, 00:20:41.533 "nvme_io_md": false, 00:20:41.533 "write_zeroes": true, 00:20:41.533 "zcopy": true, 00:20:41.533 "get_zone_info": false, 00:20:41.533 "zone_management": false, 00:20:41.533 "zone_append": false, 00:20:41.533 "compare": false, 00:20:41.533 "compare_and_write": false, 00:20:41.533 "abort": true, 00:20:41.533 "seek_hole": false, 00:20:41.533 "seek_data": false, 00:20:41.533 "copy": true, 00:20:41.533 "nvme_iov_md": false 00:20:41.533 }, 00:20:41.533 "memory_domains": [ 00:20:41.533 { 00:20:41.533 "dma_device_id": "system", 00:20:41.533 "dma_device_type": 1 00:20:41.533 }, 00:20:41.533 { 00:20:41.533 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:41.533 "dma_device_type": 2 00:20:41.533 } 00:20:41.533 ], 00:20:41.533 "driver_specific": { 00:20:41.533 "passthru": { 00:20:41.533 "name": "pt1", 00:20:41.533 "base_bdev_name": "malloc1" 00:20:41.533 } 00:20:41.533 } 00:20:41.533 }' 00:20:41.533 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:41.533 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:41.792 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:20:41.792 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:41.792 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:41.792 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:20:41.792 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:41.792 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:41.792 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:20:41.792 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:41.792 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:42.051 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:20:42.051 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:20:42.051 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:20:42.051 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:20:42.309 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:20:42.309 "name": "pt2", 00:20:42.309 "aliases": [ 00:20:42.309 "00000000-0000-0000-0000-000000000002" 00:20:42.309 ], 00:20:42.309 "product_name": "passthru", 00:20:42.309 "block_size": 512, 00:20:42.309 "num_blocks": 65536, 00:20:42.309 "uuid": "00000000-0000-0000-0000-000000000002", 00:20:42.309 "assigned_rate_limits": { 00:20:42.309 "rw_ios_per_sec": 0, 00:20:42.309 "rw_mbytes_per_sec": 0, 00:20:42.309 "r_mbytes_per_sec": 0, 00:20:42.309 "w_mbytes_per_sec": 0 00:20:42.309 }, 00:20:42.309 "claimed": true, 00:20:42.309 "claim_type": "exclusive_write", 00:20:42.309 "zoned": false, 00:20:42.309 "supported_io_types": { 00:20:42.309 "read": true, 00:20:42.309 "write": true, 00:20:42.309 "unmap": true, 00:20:42.309 "flush": true, 00:20:42.309 "reset": true, 00:20:42.309 "nvme_admin": false, 00:20:42.309 "nvme_io": false, 00:20:42.309 "nvme_io_md": false, 00:20:42.309 "write_zeroes": true, 00:20:42.309 "zcopy": true, 00:20:42.309 "get_zone_info": false, 00:20:42.309 "zone_management": false, 00:20:42.309 "zone_append": false, 00:20:42.309 "compare": false, 00:20:42.309 "compare_and_write": false, 00:20:42.309 "abort": true, 00:20:42.309 "seek_hole": false, 00:20:42.309 "seek_data": false, 00:20:42.309 "copy": true, 00:20:42.309 "nvme_iov_md": false 00:20:42.309 }, 00:20:42.309 "memory_domains": [ 00:20:42.309 { 00:20:42.309 "dma_device_id": "system", 00:20:42.309 "dma_device_type": 1 00:20:42.309 }, 00:20:42.309 { 00:20:42.309 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:42.309 "dma_device_type": 2 00:20:42.309 } 00:20:42.309 ], 00:20:42.309 "driver_specific": { 00:20:42.309 "passthru": { 00:20:42.309 "name": "pt2", 00:20:42.309 "base_bdev_name": "malloc2" 00:20:42.309 } 00:20:42.309 } 00:20:42.309 }' 00:20:42.309 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:42.309 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:42.309 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:20:42.309 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:42.309 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:42.309 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:20:42.309 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:42.309 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:42.567 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:20:42.567 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:42.567 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:42.567 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:20:42.567 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:20:42.567 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt3 00:20:42.567 11:31:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:20:42.825 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:20:42.825 "name": "pt3", 00:20:42.825 "aliases": [ 00:20:42.825 "00000000-0000-0000-0000-000000000003" 00:20:42.825 ], 00:20:42.825 "product_name": "passthru", 00:20:42.825 "block_size": 512, 00:20:42.825 "num_blocks": 65536, 00:20:42.825 "uuid": "00000000-0000-0000-0000-000000000003", 00:20:42.825 "assigned_rate_limits": { 00:20:42.825 "rw_ios_per_sec": 0, 00:20:42.825 "rw_mbytes_per_sec": 0, 00:20:42.825 "r_mbytes_per_sec": 0, 00:20:42.825 "w_mbytes_per_sec": 0 00:20:42.825 }, 00:20:42.825 "claimed": true, 00:20:42.825 "claim_type": "exclusive_write", 00:20:42.825 "zoned": false, 00:20:42.825 "supported_io_types": { 00:20:42.825 "read": true, 00:20:42.825 "write": true, 00:20:42.825 "unmap": true, 00:20:42.825 "flush": true, 00:20:42.825 "reset": true, 00:20:42.825 "nvme_admin": false, 00:20:42.825 "nvme_io": false, 00:20:42.825 "nvme_io_md": false, 00:20:42.825 "write_zeroes": true, 00:20:42.825 "zcopy": true, 00:20:42.825 "get_zone_info": false, 00:20:42.825 "zone_management": false, 00:20:42.825 "zone_append": false, 00:20:42.825 "compare": false, 00:20:42.825 "compare_and_write": false, 00:20:42.825 "abort": true, 00:20:42.825 "seek_hole": false, 00:20:42.825 "seek_data": false, 00:20:42.825 "copy": true, 00:20:42.825 "nvme_iov_md": false 00:20:42.825 }, 00:20:42.825 "memory_domains": [ 00:20:42.825 { 00:20:42.825 "dma_device_id": "system", 00:20:42.825 "dma_device_type": 1 00:20:42.825 }, 00:20:42.825 { 00:20:42.825 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:42.825 "dma_device_type": 2 00:20:42.825 } 00:20:42.825 ], 00:20:42.825 "driver_specific": { 00:20:42.825 "passthru": { 00:20:42.825 "name": "pt3", 00:20:42.825 "base_bdev_name": "malloc3" 00:20:42.825 } 00:20:42.825 } 00:20:42.825 }' 00:20:42.825 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:42.825 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:42.825 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:20:42.825 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:42.825 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:42.825 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:20:42.825 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:43.108 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:43.108 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:20:43.108 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:43.108 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:43.108 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:20:43.108 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:20:43.108 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt4 00:20:43.108 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:20:43.366 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:20:43.366 "name": "pt4", 00:20:43.366 "aliases": [ 00:20:43.366 "00000000-0000-0000-0000-000000000004" 00:20:43.366 ], 00:20:43.366 "product_name": "passthru", 00:20:43.366 "block_size": 512, 00:20:43.366 "num_blocks": 65536, 00:20:43.366 "uuid": "00000000-0000-0000-0000-000000000004", 00:20:43.366 "assigned_rate_limits": { 00:20:43.366 "rw_ios_per_sec": 0, 00:20:43.366 "rw_mbytes_per_sec": 0, 00:20:43.366 "r_mbytes_per_sec": 0, 00:20:43.366 "w_mbytes_per_sec": 0 00:20:43.366 }, 00:20:43.366 "claimed": true, 00:20:43.366 "claim_type": "exclusive_write", 00:20:43.366 "zoned": false, 00:20:43.366 "supported_io_types": { 00:20:43.366 "read": true, 00:20:43.366 "write": true, 00:20:43.366 "unmap": true, 00:20:43.366 "flush": true, 00:20:43.366 "reset": true, 00:20:43.366 "nvme_admin": false, 00:20:43.366 "nvme_io": false, 00:20:43.366 "nvme_io_md": false, 00:20:43.366 "write_zeroes": true, 00:20:43.366 "zcopy": true, 00:20:43.366 "get_zone_info": false, 00:20:43.366 "zone_management": false, 00:20:43.366 "zone_append": false, 00:20:43.366 "compare": false, 00:20:43.366 "compare_and_write": false, 00:20:43.366 "abort": true, 00:20:43.366 "seek_hole": false, 00:20:43.366 "seek_data": false, 00:20:43.366 "copy": true, 00:20:43.366 "nvme_iov_md": false 00:20:43.366 }, 00:20:43.366 "memory_domains": [ 00:20:43.366 { 00:20:43.366 "dma_device_id": "system", 00:20:43.366 "dma_device_type": 1 00:20:43.366 }, 00:20:43.366 { 00:20:43.366 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:43.366 "dma_device_type": 2 00:20:43.366 } 00:20:43.366 ], 00:20:43.366 "driver_specific": { 00:20:43.366 "passthru": { 00:20:43.366 "name": "pt4", 00:20:43.366 "base_bdev_name": "malloc4" 00:20:43.366 } 00:20:43.366 } 00:20:43.366 }' 00:20:43.366 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:43.366 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:20:43.366 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:20:43.366 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:43.366 11:31:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:20:43.624 11:31:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:20:43.624 11:31:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:43.624 11:31:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:20:43.624 11:31:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:20:43.625 11:31:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:43.625 11:31:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:20:43.625 11:31:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:20:43.625 11:31:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:20:43.625 11:31:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # jq -r '.[] | .uuid' 00:20:43.883 [2024-07-15 11:31:27.397591] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:20:43.883 11:31:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # '[' 41788dc0-696c-47e4-bdf7-43df2b427c2b '!=' 41788dc0-696c-47e4-bdf7-43df2b427c2b ']' 00:20:43.883 11:31:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@490 -- # has_redundancy concat 00:20:43.883 11:31:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:20:43.883 11:31:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@215 -- # return 1 00:20:43.883 11:31:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@562 -- # killprocess 947546 00:20:43.883 11:31:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@948 -- # '[' -z 947546 ']' 00:20:43.883 11:31:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@952 -- # kill -0 947546 00:20:43.883 11:31:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@953 -- # uname 00:20:43.883 11:31:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:20:43.883 11:31:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 947546 00:20:43.883 11:31:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:20:43.883 11:31:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:20:43.883 11:31:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 947546' 00:20:43.883 killing process with pid 947546 00:20:43.883 11:31:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@967 -- # kill 947546 00:20:43.883 [2024-07-15 11:31:27.466455] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:20:43.883 [2024-07-15 11:31:27.466517] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:20:43.883 [2024-07-15 11:31:27.466578] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:20:43.883 [2024-07-15 11:31:27.466590] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1ca88f0 name raid_bdev1, state offline 00:20:43.883 11:31:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # wait 947546 00:20:44.141 [2024-07-15 11:31:27.503503] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:20:44.141 11:31:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@564 -- # return 0 00:20:44.141 00:20:44.141 real 0m16.204s 00:20:44.141 user 0m29.269s 00:20:44.141 sys 0m2.899s 00:20:44.141 11:31:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:20:44.141 11:31:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:44.141 ************************************ 00:20:44.141 END TEST raid_superblock_test 00:20:44.141 ************************************ 00:20:44.401 11:31:27 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:20:44.401 11:31:27 bdev_raid -- bdev/bdev_raid.sh@870 -- # run_test raid_read_error_test raid_io_error_test concat 4 read 00:20:44.401 11:31:27 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:20:44.401 11:31:27 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:20:44.401 11:31:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:20:44.401 ************************************ 00:20:44.401 START TEST raid_read_error_test 00:20:44.401 ************************************ 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1123 -- # raid_io_error_test concat 4 read 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@788 -- # local raid_level=concat 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@789 -- # local num_base_bdevs=4 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local error_io_type=read 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i = 1 )) 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev1 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev2 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev3 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev4 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local base_bdevs 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local raid_bdev_name=raid_bdev1 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local strip_size 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local create_arg 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local bdevperf_log 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local fail_per_s 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # '[' concat '!=' raid1 ']' 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@799 -- # strip_size=64 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # create_arg+=' -z 64' 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@805 -- # mktemp -p /raidtest 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@805 -- # bdevperf_log=/raidtest/tmp.lfT8GBlCy2 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@808 -- # raid_pid=949991 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # waitforlisten 949991 /var/tmp/spdk-raid.sock 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@829 -- # '[' -z 949991 ']' 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:20:44.401 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:20:44.401 11:31:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:44.401 [2024-07-15 11:31:27.859219] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:20:44.401 [2024-07-15 11:31:27.859283] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid949991 ] 00:20:44.401 [2024-07-15 11:31:27.990599] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:44.660 [2024-07-15 11:31:28.096395] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:20:44.660 [2024-07-15 11:31:28.163126] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:44.660 [2024-07-15 11:31:28.163167] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:45.229 11:31:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:20:45.229 11:31:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@862 -- # return 0 00:20:45.229 11:31:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:20:45.229 11:31:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:20:45.488 BaseBdev1_malloc 00:20:45.488 11:31:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev1_malloc 00:20:45.747 true 00:20:45.747 11:31:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:20:46.006 [2024-07-15 11:31:29.486305] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:20:46.006 [2024-07-15 11:31:29.486349] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:46.006 [2024-07-15 11:31:29.486369] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x20cd0d0 00:20:46.006 [2024-07-15 11:31:29.486387] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:46.006 [2024-07-15 11:31:29.488209] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:46.006 [2024-07-15 11:31:29.488239] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:20:46.006 BaseBdev1 00:20:46.006 11:31:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:20:46.006 11:31:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:20:46.266 BaseBdev2_malloc 00:20:46.266 11:31:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev2_malloc 00:20:46.525 true 00:20:46.525 11:31:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:20:46.784 [2024-07-15 11:31:30.225350] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:20:46.784 [2024-07-15 11:31:30.225395] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:46.784 [2024-07-15 11:31:30.225416] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x20d1910 00:20:46.784 [2024-07-15 11:31:30.225429] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:46.784 [2024-07-15 11:31:30.227049] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:46.784 [2024-07-15 11:31:30.227079] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:20:46.784 BaseBdev2 00:20:46.784 11:31:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:20:46.784 11:31:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:20:47.043 BaseBdev3_malloc 00:20:47.043 11:31:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev3_malloc 00:20:47.302 true 00:20:47.302 11:31:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:20:47.561 [2024-07-15 11:31:30.969322] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:20:47.561 [2024-07-15 11:31:30.969364] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:47.561 [2024-07-15 11:31:30.969385] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x20d3bd0 00:20:47.561 [2024-07-15 11:31:30.969398] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:47.561 [2024-07-15 11:31:30.970966] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:47.561 [2024-07-15 11:31:30.970992] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:20:47.561 BaseBdev3 00:20:47.561 11:31:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:20:47.562 11:31:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:20:47.820 BaseBdev4_malloc 00:20:47.820 11:31:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev4_malloc 00:20:48.080 true 00:20:48.080 11:31:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:20:48.341 [2024-07-15 11:31:31.691792] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:20:48.341 [2024-07-15 11:31:31.691842] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:48.341 [2024-07-15 11:31:31.691864] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x20d4aa0 00:20:48.341 [2024-07-15 11:31:31.691877] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:48.341 [2024-07-15 11:31:31.693477] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:48.341 [2024-07-15 11:31:31.693505] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:20:48.341 BaseBdev4 00:20:48.341 11:31:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@819 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n raid_bdev1 -s 00:20:48.634 [2024-07-15 11:31:31.940489] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:20:48.634 [2024-07-15 11:31:31.941855] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:20:48.634 [2024-07-15 11:31:31.941924] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:20:48.634 [2024-07-15 11:31:31.941995] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:20:48.634 [2024-07-15 11:31:31.942228] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x20cec20 00:20:48.634 [2024-07-15 11:31:31.942239] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:20:48.634 [2024-07-15 11:31:31.942443] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1f23260 00:20:48.634 [2024-07-15 11:31:31.942595] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x20cec20 00:20:48.634 [2024-07-15 11:31:31.942605] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x20cec20 00:20:48.634 [2024-07-15 11:31:31.942711] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:48.634 11:31:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@820 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:20:48.634 11:31:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:20:48.634 11:31:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:20:48.634 11:31:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:20:48.634 11:31:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:20:48.634 11:31:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:20:48.634 11:31:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:20:48.634 11:31:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:20:48.634 11:31:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:20:48.634 11:31:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:20:48.634 11:31:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:48.634 11:31:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:48.634 11:31:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:48.634 "name": "raid_bdev1", 00:20:48.634 "uuid": "1cbd1075-fddb-4209-aab8-1b02279f0169", 00:20:48.634 "strip_size_kb": 64, 00:20:48.634 "state": "online", 00:20:48.634 "raid_level": "concat", 00:20:48.634 "superblock": true, 00:20:48.634 "num_base_bdevs": 4, 00:20:48.634 "num_base_bdevs_discovered": 4, 00:20:48.634 "num_base_bdevs_operational": 4, 00:20:48.634 "base_bdevs_list": [ 00:20:48.634 { 00:20:48.634 "name": "BaseBdev1", 00:20:48.634 "uuid": "8a46cc73-2ec8-5245-a453-5c2ba22ce327", 00:20:48.634 "is_configured": true, 00:20:48.634 "data_offset": 2048, 00:20:48.634 "data_size": 63488 00:20:48.634 }, 00:20:48.634 { 00:20:48.634 "name": "BaseBdev2", 00:20:48.635 "uuid": "2f4c39f8-d22b-5428-96ce-fe8e6fbc2128", 00:20:48.635 "is_configured": true, 00:20:48.635 "data_offset": 2048, 00:20:48.635 "data_size": 63488 00:20:48.635 }, 00:20:48.635 { 00:20:48.635 "name": "BaseBdev3", 00:20:48.635 "uuid": "fd53cf4d-3f15-5480-a518-2fe0d8e55275", 00:20:48.635 "is_configured": true, 00:20:48.635 "data_offset": 2048, 00:20:48.635 "data_size": 63488 00:20:48.635 }, 00:20:48.635 { 00:20:48.635 "name": "BaseBdev4", 00:20:48.635 "uuid": "30e536a1-3aeb-582d-b225-bee1a96515ae", 00:20:48.635 "is_configured": true, 00:20:48.635 "data_offset": 2048, 00:20:48.635 "data_size": 63488 00:20:48.635 } 00:20:48.635 ] 00:20:48.635 }' 00:20:48.635 11:31:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:48.635 11:31:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:49.202 11:31:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@824 -- # sleep 1 00:20:49.202 11:31:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@823 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:20:49.460 [2024-07-15 11:31:32.899318] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x20c0fc0 00:20:50.395 11:31:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@827 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:20:50.653 11:31:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # local expected_num_base_bdevs 00:20:50.653 11:31:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@830 -- # [[ concat = \r\a\i\d\1 ]] 00:20:50.653 11:31:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=4 00:20:50.653 11:31:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:20:50.653 11:31:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:20:50.653 11:31:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:20:50.653 11:31:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:20:50.653 11:31:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:20:50.653 11:31:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:20:50.653 11:31:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:20:50.653 11:31:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:20:50.653 11:31:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:20:50.653 11:31:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:20:50.653 11:31:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:50.653 11:31:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:50.912 11:31:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:50.912 "name": "raid_bdev1", 00:20:50.912 "uuid": "1cbd1075-fddb-4209-aab8-1b02279f0169", 00:20:50.912 "strip_size_kb": 64, 00:20:50.912 "state": "online", 00:20:50.912 "raid_level": "concat", 00:20:50.912 "superblock": true, 00:20:50.912 "num_base_bdevs": 4, 00:20:50.912 "num_base_bdevs_discovered": 4, 00:20:50.912 "num_base_bdevs_operational": 4, 00:20:50.912 "base_bdevs_list": [ 00:20:50.912 { 00:20:50.912 "name": "BaseBdev1", 00:20:50.912 "uuid": "8a46cc73-2ec8-5245-a453-5c2ba22ce327", 00:20:50.912 "is_configured": true, 00:20:50.912 "data_offset": 2048, 00:20:50.912 "data_size": 63488 00:20:50.912 }, 00:20:50.912 { 00:20:50.912 "name": "BaseBdev2", 00:20:50.912 "uuid": "2f4c39f8-d22b-5428-96ce-fe8e6fbc2128", 00:20:50.912 "is_configured": true, 00:20:50.912 "data_offset": 2048, 00:20:50.912 "data_size": 63488 00:20:50.912 }, 00:20:50.912 { 00:20:50.912 "name": "BaseBdev3", 00:20:50.912 "uuid": "fd53cf4d-3f15-5480-a518-2fe0d8e55275", 00:20:50.912 "is_configured": true, 00:20:50.912 "data_offset": 2048, 00:20:50.912 "data_size": 63488 00:20:50.912 }, 00:20:50.912 { 00:20:50.912 "name": "BaseBdev4", 00:20:50.912 "uuid": "30e536a1-3aeb-582d-b225-bee1a96515ae", 00:20:50.912 "is_configured": true, 00:20:50.912 "data_offset": 2048, 00:20:50.912 "data_size": 63488 00:20:50.912 } 00:20:50.912 ] 00:20:50.912 }' 00:20:50.912 11:31:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:50.912 11:31:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:51.576 11:31:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:20:51.576 [2024-07-15 11:31:35.117343] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:20:51.576 [2024-07-15 11:31:35.117386] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:20:51.576 [2024-07-15 11:31:35.120549] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:20:51.576 [2024-07-15 11:31:35.120589] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:51.576 [2024-07-15 11:31:35.120630] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:20:51.576 [2024-07-15 11:31:35.120641] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x20cec20 name raid_bdev1, state offline 00:20:51.576 0 00:20:51.576 11:31:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # killprocess 949991 00:20:51.576 11:31:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@948 -- # '[' -z 949991 ']' 00:20:51.576 11:31:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@952 -- # kill -0 949991 00:20:51.576 11:31:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@953 -- # uname 00:20:51.576 11:31:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:20:51.576 11:31:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 949991 00:20:51.835 11:31:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:20:51.835 11:31:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:20:51.835 11:31:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 949991' 00:20:51.835 killing process with pid 949991 00:20:51.835 11:31:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@967 -- # kill 949991 00:20:51.835 [2024-07-15 11:31:35.186032] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:20:51.835 11:31:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # wait 949991 00:20:51.835 [2024-07-15 11:31:35.218010] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:20:52.094 11:31:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # grep -v Job /raidtest/tmp.lfT8GBlCy2 00:20:52.094 11:31:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # grep raid_bdev1 00:20:52.094 11:31:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # awk '{print $6}' 00:20:52.094 11:31:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # fail_per_s=0.45 00:20:52.094 11:31:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@844 -- # has_redundancy concat 00:20:52.094 11:31:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:20:52.094 11:31:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@215 -- # return 1 00:20:52.094 11:31:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.45 != \0\.\0\0 ]] 00:20:52.094 00:20:52.094 real 0m7.679s 00:20:52.094 user 0m12.303s 00:20:52.094 sys 0m1.341s 00:20:52.094 11:31:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:20:52.094 11:31:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:52.094 ************************************ 00:20:52.094 END TEST raid_read_error_test 00:20:52.094 ************************************ 00:20:52.094 11:31:35 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:20:52.094 11:31:35 bdev_raid -- bdev/bdev_raid.sh@871 -- # run_test raid_write_error_test raid_io_error_test concat 4 write 00:20:52.094 11:31:35 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:20:52.094 11:31:35 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:20:52.094 11:31:35 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:20:52.094 ************************************ 00:20:52.094 START TEST raid_write_error_test 00:20:52.094 ************************************ 00:20:52.094 11:31:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1123 -- # raid_io_error_test concat 4 write 00:20:52.094 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@788 -- # local raid_level=concat 00:20:52.094 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@789 -- # local num_base_bdevs=4 00:20:52.094 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local error_io_type=write 00:20:52.094 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i = 1 )) 00:20:52.094 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:20:52.094 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev1 00:20:52.094 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:20:52.094 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:20:52.094 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev2 00:20:52.094 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:20:52.094 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:20:52.094 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev3 00:20:52.094 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:20:52.094 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev4 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local base_bdevs 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local raid_bdev_name=raid_bdev1 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local strip_size 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local create_arg 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local bdevperf_log 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local fail_per_s 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # '[' concat '!=' raid1 ']' 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@799 -- # strip_size=64 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # create_arg+=' -z 64' 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@805 -- # mktemp -p /raidtest 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@805 -- # bdevperf_log=/raidtest/tmp.0Dqhf1uyTs 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@808 -- # raid_pid=950993 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # waitforlisten 950993 /var/tmp/spdk-raid.sock 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@829 -- # '[' -z 950993 ']' 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:20:52.095 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:20:52.095 11:31:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:52.095 [2024-07-15 11:31:35.632073] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:20:52.095 [2024-07-15 11:31:35.632145] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid950993 ] 00:20:52.353 [2024-07-15 11:31:35.771014] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:52.353 [2024-07-15 11:31:35.868172] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:20:52.353 [2024-07-15 11:31:35.931001] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:52.353 [2024-07-15 11:31:35.931054] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:53.289 11:31:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:20:53.289 11:31:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@862 -- # return 0 00:20:53.289 11:31:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:20:53.289 11:31:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:20:53.289 BaseBdev1_malloc 00:20:53.289 11:31:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev1_malloc 00:20:53.547 true 00:20:53.547 11:31:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:20:53.805 [2024-07-15 11:31:37.281683] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:20:53.805 [2024-07-15 11:31:37.281735] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:53.805 [2024-07-15 11:31:37.281757] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x24900d0 00:20:53.805 [2024-07-15 11:31:37.281770] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:53.805 [2024-07-15 11:31:37.283568] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:53.805 [2024-07-15 11:31:37.283598] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:20:53.805 BaseBdev1 00:20:53.805 11:31:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:20:53.805 11:31:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:20:54.063 BaseBdev2_malloc 00:20:54.063 11:31:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev2_malloc 00:20:54.320 true 00:20:54.320 11:31:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:20:54.578 [2024-07-15 11:31:38.012249] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:20:54.578 [2024-07-15 11:31:38.012294] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:54.578 [2024-07-15 11:31:38.012314] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2494910 00:20:54.578 [2024-07-15 11:31:38.012326] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:54.578 [2024-07-15 11:31:38.013708] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:54.578 [2024-07-15 11:31:38.013736] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:20:54.578 BaseBdev2 00:20:54.578 11:31:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:20:54.578 11:31:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:20:54.835 BaseBdev3_malloc 00:20:54.835 11:31:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev3_malloc 00:20:55.093 true 00:20:55.093 11:31:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:20:55.351 [2024-07-15 11:31:38.790898] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:20:55.351 [2024-07-15 11:31:38.790949] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:55.351 [2024-07-15 11:31:38.790976] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2496bd0 00:20:55.351 [2024-07-15 11:31:38.790989] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:55.351 [2024-07-15 11:31:38.792578] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:55.351 [2024-07-15 11:31:38.792607] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:20:55.351 BaseBdev3 00:20:55.351 11:31:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:20:55.351 11:31:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:20:55.609 BaseBdev4_malloc 00:20:55.609 11:31:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev4_malloc 00:20:56.175 true 00:20:56.175 11:31:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:20:56.741 [2024-07-15 11:31:40.054778] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:20:56.741 [2024-07-15 11:31:40.054828] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:56.741 [2024-07-15 11:31:40.054851] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2497aa0 00:20:56.741 [2024-07-15 11:31:40.054864] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:56.741 [2024-07-15 11:31:40.056492] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:56.741 [2024-07-15 11:31:40.056521] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:20:56.741 BaseBdev4 00:20:56.741 11:31:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@819 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -z 64 -r concat -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n raid_bdev1 -s 00:20:56.741 [2024-07-15 11:31:40.311499] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:20:56.741 [2024-07-15 11:31:40.312882] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:20:56.741 [2024-07-15 11:31:40.312964] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:20:56.741 [2024-07-15 11:31:40.313027] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:20:56.741 [2024-07-15 11:31:40.313269] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x2491c20 00:20:56.741 [2024-07-15 11:31:40.313281] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:20:56.741 [2024-07-15 11:31:40.313489] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x22e6260 00:20:56.741 [2024-07-15 11:31:40.313646] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x2491c20 00:20:56.741 [2024-07-15 11:31:40.313657] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x2491c20 00:20:56.741 [2024-07-15 11:31:40.313764] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:56.741 11:31:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@820 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:20:56.741 11:31:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:20:56.741 11:31:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:20:56.741 11:31:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:20:56.742 11:31:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:20:56.742 11:31:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:20:56.742 11:31:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:20:56.742 11:31:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:20:56.742 11:31:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:20:56.742 11:31:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:20:56.742 11:31:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:56.742 11:31:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:57.309 11:31:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:57.309 "name": "raid_bdev1", 00:20:57.309 "uuid": "392e4996-802c-4250-b2c8-f53aa91d0b1d", 00:20:57.309 "strip_size_kb": 64, 00:20:57.309 "state": "online", 00:20:57.309 "raid_level": "concat", 00:20:57.309 "superblock": true, 00:20:57.309 "num_base_bdevs": 4, 00:20:57.309 "num_base_bdevs_discovered": 4, 00:20:57.309 "num_base_bdevs_operational": 4, 00:20:57.309 "base_bdevs_list": [ 00:20:57.309 { 00:20:57.309 "name": "BaseBdev1", 00:20:57.309 "uuid": "be2be12d-dc0f-559e-9082-b06465a593d7", 00:20:57.309 "is_configured": true, 00:20:57.309 "data_offset": 2048, 00:20:57.309 "data_size": 63488 00:20:57.309 }, 00:20:57.309 { 00:20:57.309 "name": "BaseBdev2", 00:20:57.309 "uuid": "c831bb19-03a2-5e46-8283-5bf8a99aa0a0", 00:20:57.309 "is_configured": true, 00:20:57.309 "data_offset": 2048, 00:20:57.309 "data_size": 63488 00:20:57.309 }, 00:20:57.309 { 00:20:57.309 "name": "BaseBdev3", 00:20:57.309 "uuid": "5ab7da12-95dd-5f9e-846a-a4154f440ef9", 00:20:57.309 "is_configured": true, 00:20:57.309 "data_offset": 2048, 00:20:57.309 "data_size": 63488 00:20:57.309 }, 00:20:57.309 { 00:20:57.309 "name": "BaseBdev4", 00:20:57.309 "uuid": "b067cf96-919f-5b1e-af2b-195c0d393f45", 00:20:57.309 "is_configured": true, 00:20:57.309 "data_offset": 2048, 00:20:57.309 "data_size": 63488 00:20:57.309 } 00:20:57.309 ] 00:20:57.309 }' 00:20:57.309 11:31:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:57.309 11:31:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:57.568 11:31:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@824 -- # sleep 1 00:20:57.568 11:31:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@823 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:20:57.826 [2024-07-15 11:31:41.238223] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x2483fc0 00:20:58.762 11:31:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@827 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:20:59.021 11:31:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # local expected_num_base_bdevs 00:20:59.021 11:31:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@830 -- # [[ concat = \r\a\i\d\1 ]] 00:20:59.021 11:31:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=4 00:20:59.021 11:31:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:20:59.021 11:31:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:20:59.021 11:31:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:20:59.021 11:31:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=concat 00:20:59.021 11:31:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=64 00:20:59.021 11:31:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:20:59.021 11:31:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:20:59.021 11:31:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:20:59.021 11:31:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:20:59.021 11:31:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:20:59.022 11:31:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:59.022 11:31:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:20:59.022 11:31:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:20:59.022 "name": "raid_bdev1", 00:20:59.022 "uuid": "392e4996-802c-4250-b2c8-f53aa91d0b1d", 00:20:59.022 "strip_size_kb": 64, 00:20:59.022 "state": "online", 00:20:59.022 "raid_level": "concat", 00:20:59.022 "superblock": true, 00:20:59.022 "num_base_bdevs": 4, 00:20:59.022 "num_base_bdevs_discovered": 4, 00:20:59.022 "num_base_bdevs_operational": 4, 00:20:59.022 "base_bdevs_list": [ 00:20:59.022 { 00:20:59.022 "name": "BaseBdev1", 00:20:59.022 "uuid": "be2be12d-dc0f-559e-9082-b06465a593d7", 00:20:59.022 "is_configured": true, 00:20:59.022 "data_offset": 2048, 00:20:59.022 "data_size": 63488 00:20:59.022 }, 00:20:59.022 { 00:20:59.022 "name": "BaseBdev2", 00:20:59.022 "uuid": "c831bb19-03a2-5e46-8283-5bf8a99aa0a0", 00:20:59.022 "is_configured": true, 00:20:59.022 "data_offset": 2048, 00:20:59.022 "data_size": 63488 00:20:59.022 }, 00:20:59.022 { 00:20:59.022 "name": "BaseBdev3", 00:20:59.022 "uuid": "5ab7da12-95dd-5f9e-846a-a4154f440ef9", 00:20:59.022 "is_configured": true, 00:20:59.022 "data_offset": 2048, 00:20:59.022 "data_size": 63488 00:20:59.022 }, 00:20:59.022 { 00:20:59.022 "name": "BaseBdev4", 00:20:59.022 "uuid": "b067cf96-919f-5b1e-af2b-195c0d393f45", 00:20:59.022 "is_configured": true, 00:20:59.022 "data_offset": 2048, 00:20:59.022 "data_size": 63488 00:20:59.022 } 00:20:59.022 ] 00:20:59.022 }' 00:20:59.022 11:31:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:20:59.022 11:31:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:59.958 11:31:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:20:59.958 [2024-07-15 11:31:43.411128] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:20:59.958 [2024-07-15 11:31:43.411172] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:20:59.958 [2024-07-15 11:31:43.414418] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:20:59.958 [2024-07-15 11:31:43.414458] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:59.958 [2024-07-15 11:31:43.414498] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:20:59.958 [2024-07-15 11:31:43.414509] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x2491c20 name raid_bdev1, state offline 00:20:59.958 0 00:20:59.958 11:31:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # killprocess 950993 00:20:59.958 11:31:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@948 -- # '[' -z 950993 ']' 00:20:59.958 11:31:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@952 -- # kill -0 950993 00:20:59.958 11:31:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@953 -- # uname 00:20:59.958 11:31:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:20:59.958 11:31:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 950993 00:20:59.958 11:31:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:20:59.958 11:31:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:20:59.958 11:31:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 950993' 00:20:59.958 killing process with pid 950993 00:20:59.958 11:31:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@967 -- # kill 950993 00:20:59.958 [2024-07-15 11:31:43.478234] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:20:59.958 11:31:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # wait 950993 00:20:59.958 [2024-07-15 11:31:43.509995] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:21:00.217 11:31:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # grep -v Job /raidtest/tmp.0Dqhf1uyTs 00:21:00.217 11:31:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # grep raid_bdev1 00:21:00.217 11:31:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # awk '{print $6}' 00:21:00.217 11:31:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # fail_per_s=0.46 00:21:00.217 11:31:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@844 -- # has_redundancy concat 00:21:00.217 11:31:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:21:00.217 11:31:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@215 -- # return 1 00:21:00.217 11:31:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.46 != \0\.\0\0 ]] 00:21:00.217 00:21:00.217 real 0m8.203s 00:21:00.217 user 0m13.219s 00:21:00.217 sys 0m1.445s 00:21:00.217 11:31:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:21:00.217 11:31:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:00.217 ************************************ 00:21:00.217 END TEST raid_write_error_test 00:21:00.217 ************************************ 00:21:00.217 11:31:43 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:21:00.217 11:31:43 bdev_raid -- bdev/bdev_raid.sh@866 -- # for level in raid0 concat raid1 00:21:00.217 11:31:43 bdev_raid -- bdev/bdev_raid.sh@867 -- # run_test raid_state_function_test raid_state_function_test raid1 4 false 00:21:00.217 11:31:43 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:21:00.217 11:31:43 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:21:00.217 11:31:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:21:00.476 ************************************ 00:21:00.476 START TEST raid_state_function_test 00:21:00.476 ************************************ 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1123 -- # raid_state_function_test raid1 4 false 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@220 -- # local raid_level=raid1 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=4 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # local superblock=false 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev3 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # echo BaseBdev4 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@226 -- # local strip_size 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # '[' raid1 '!=' raid1 ']' 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@234 -- # strip_size=0 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # '[' false = true ']' 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@240 -- # superblock_create_arg= 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # raid_pid=952238 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 952238' 00:21:00.476 Process raid pid: 952238 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@246 -- # waitforlisten 952238 /var/tmp/spdk-raid.sock 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@829 -- # '[' -z 952238 ']' 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:21:00.476 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:21:00.476 11:31:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:00.476 [2024-07-15 11:31:43.914503] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:21:00.476 [2024-07-15 11:31:43.914571] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:00.476 [2024-07-15 11:31:44.043058] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:00.734 [2024-07-15 11:31:44.140836] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:21:00.735 [2024-07-15 11:31:44.201227] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:21:00.735 [2024-07-15 11:31:44.201264] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:21:01.300 11:31:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:21:01.300 11:31:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@862 -- # return 0 00:21:01.301 11:31:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:21:01.558 [2024-07-15 11:31:45.075796] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:21:01.558 [2024-07-15 11:31:45.075836] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:21:01.558 [2024-07-15 11:31:45.075848] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:21:01.558 [2024-07-15 11:31:45.075860] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:21:01.558 [2024-07-15 11:31:45.075869] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:21:01.558 [2024-07-15 11:31:45.075880] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:21:01.558 [2024-07-15 11:31:45.075889] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:21:01.558 [2024-07-15 11:31:45.075900] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:21:01.558 11:31:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:01.558 11:31:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:01.558 11:31:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:01.558 11:31:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:01.558 11:31:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:01.558 11:31:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:01.558 11:31:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:01.558 11:31:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:01.558 11:31:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:01.558 11:31:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:01.558 11:31:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:01.558 11:31:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:01.817 11:31:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:01.817 "name": "Existed_Raid", 00:21:01.817 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:01.817 "strip_size_kb": 0, 00:21:01.817 "state": "configuring", 00:21:01.817 "raid_level": "raid1", 00:21:01.817 "superblock": false, 00:21:01.817 "num_base_bdevs": 4, 00:21:01.817 "num_base_bdevs_discovered": 0, 00:21:01.817 "num_base_bdevs_operational": 4, 00:21:01.817 "base_bdevs_list": [ 00:21:01.817 { 00:21:01.817 "name": "BaseBdev1", 00:21:01.817 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:01.817 "is_configured": false, 00:21:01.817 "data_offset": 0, 00:21:01.817 "data_size": 0 00:21:01.817 }, 00:21:01.817 { 00:21:01.817 "name": "BaseBdev2", 00:21:01.817 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:01.817 "is_configured": false, 00:21:01.817 "data_offset": 0, 00:21:01.817 "data_size": 0 00:21:01.817 }, 00:21:01.817 { 00:21:01.817 "name": "BaseBdev3", 00:21:01.817 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:01.817 "is_configured": false, 00:21:01.817 "data_offset": 0, 00:21:01.817 "data_size": 0 00:21:01.817 }, 00:21:01.817 { 00:21:01.817 "name": "BaseBdev4", 00:21:01.817 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:01.817 "is_configured": false, 00:21:01.817 "data_offset": 0, 00:21:01.817 "data_size": 0 00:21:01.817 } 00:21:01.817 ] 00:21:01.817 }' 00:21:01.817 11:31:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:01.817 11:31:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:02.438 11:31:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:21:02.696 [2024-07-15 11:31:46.178600] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:21:02.696 [2024-07-15 11:31:46.178631] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1d7baa0 name Existed_Raid, state configuring 00:21:02.696 11:31:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:21:02.954 [2024-07-15 11:31:46.423262] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:21:02.954 [2024-07-15 11:31:46.423291] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:21:02.954 [2024-07-15 11:31:46.423300] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:21:02.954 [2024-07-15 11:31:46.423312] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:21:02.954 [2024-07-15 11:31:46.423321] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:21:02.954 [2024-07-15 11:31:46.423332] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:21:02.954 [2024-07-15 11:31:46.423341] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:21:02.954 [2024-07-15 11:31:46.423352] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:21:02.954 11:31:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:21:03.212 [2024-07-15 11:31:46.677855] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:21:03.212 BaseBdev1 00:21:03.212 11:31:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:21:03.212 11:31:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:21:03.212 11:31:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:21:03.213 11:31:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:21:03.213 11:31:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:21:03.213 11:31:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:21:03.213 11:31:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:21:03.471 11:31:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:21:03.728 [ 00:21:03.728 { 00:21:03.728 "name": "BaseBdev1", 00:21:03.728 "aliases": [ 00:21:03.728 "e7e220f1-6c42-4f6d-8266-07f4ae2fbb6e" 00:21:03.728 ], 00:21:03.728 "product_name": "Malloc disk", 00:21:03.728 "block_size": 512, 00:21:03.728 "num_blocks": 65536, 00:21:03.728 "uuid": "e7e220f1-6c42-4f6d-8266-07f4ae2fbb6e", 00:21:03.728 "assigned_rate_limits": { 00:21:03.728 "rw_ios_per_sec": 0, 00:21:03.728 "rw_mbytes_per_sec": 0, 00:21:03.728 "r_mbytes_per_sec": 0, 00:21:03.729 "w_mbytes_per_sec": 0 00:21:03.729 }, 00:21:03.729 "claimed": true, 00:21:03.729 "claim_type": "exclusive_write", 00:21:03.729 "zoned": false, 00:21:03.729 "supported_io_types": { 00:21:03.729 "read": true, 00:21:03.729 "write": true, 00:21:03.729 "unmap": true, 00:21:03.729 "flush": true, 00:21:03.729 "reset": true, 00:21:03.729 "nvme_admin": false, 00:21:03.729 "nvme_io": false, 00:21:03.729 "nvme_io_md": false, 00:21:03.729 "write_zeroes": true, 00:21:03.729 "zcopy": true, 00:21:03.729 "get_zone_info": false, 00:21:03.729 "zone_management": false, 00:21:03.729 "zone_append": false, 00:21:03.729 "compare": false, 00:21:03.729 "compare_and_write": false, 00:21:03.729 "abort": true, 00:21:03.729 "seek_hole": false, 00:21:03.729 "seek_data": false, 00:21:03.729 "copy": true, 00:21:03.729 "nvme_iov_md": false 00:21:03.729 }, 00:21:03.729 "memory_domains": [ 00:21:03.729 { 00:21:03.729 "dma_device_id": "system", 00:21:03.729 "dma_device_type": 1 00:21:03.729 }, 00:21:03.729 { 00:21:03.729 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:03.729 "dma_device_type": 2 00:21:03.729 } 00:21:03.729 ], 00:21:03.729 "driver_specific": {} 00:21:03.729 } 00:21:03.729 ] 00:21:03.729 11:31:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:21:03.729 11:31:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:03.729 11:31:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:03.729 11:31:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:03.729 11:31:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:03.729 11:31:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:03.729 11:31:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:03.729 11:31:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:03.729 11:31:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:03.729 11:31:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:03.729 11:31:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:03.729 11:31:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:03.729 11:31:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:03.987 11:31:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:03.987 "name": "Existed_Raid", 00:21:03.987 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:03.987 "strip_size_kb": 0, 00:21:03.987 "state": "configuring", 00:21:03.987 "raid_level": "raid1", 00:21:03.987 "superblock": false, 00:21:03.987 "num_base_bdevs": 4, 00:21:03.987 "num_base_bdevs_discovered": 1, 00:21:03.987 "num_base_bdevs_operational": 4, 00:21:03.987 "base_bdevs_list": [ 00:21:03.987 { 00:21:03.987 "name": "BaseBdev1", 00:21:03.987 "uuid": "e7e220f1-6c42-4f6d-8266-07f4ae2fbb6e", 00:21:03.987 "is_configured": true, 00:21:03.987 "data_offset": 0, 00:21:03.987 "data_size": 65536 00:21:03.987 }, 00:21:03.987 { 00:21:03.987 "name": "BaseBdev2", 00:21:03.987 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:03.987 "is_configured": false, 00:21:03.987 "data_offset": 0, 00:21:03.987 "data_size": 0 00:21:03.987 }, 00:21:03.987 { 00:21:03.987 "name": "BaseBdev3", 00:21:03.987 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:03.987 "is_configured": false, 00:21:03.987 "data_offset": 0, 00:21:03.987 "data_size": 0 00:21:03.987 }, 00:21:03.987 { 00:21:03.987 "name": "BaseBdev4", 00:21:03.987 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:03.987 "is_configured": false, 00:21:03.987 "data_offset": 0, 00:21:03.987 "data_size": 0 00:21:03.987 } 00:21:03.987 ] 00:21:03.987 }' 00:21:03.987 11:31:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:03.987 11:31:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:04.552 11:31:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:21:04.810 [2024-07-15 11:31:48.237996] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:21:04.810 [2024-07-15 11:31:48.238039] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1d7b310 name Existed_Raid, state configuring 00:21:04.810 11:31:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:21:05.068 [2024-07-15 11:31:48.478667] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:21:05.068 [2024-07-15 11:31:48.480163] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:21:05.068 [2024-07-15 11:31:48.480198] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:21:05.068 [2024-07-15 11:31:48.480208] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:21:05.068 [2024-07-15 11:31:48.480220] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:21:05.068 [2024-07-15 11:31:48.480229] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:21:05.068 [2024-07-15 11:31:48.480240] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:21:05.068 11:31:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:21:05.068 11:31:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:21:05.068 11:31:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:05.068 11:31:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:05.068 11:31:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:05.068 11:31:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:05.068 11:31:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:05.068 11:31:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:05.068 11:31:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:05.068 11:31:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:05.068 11:31:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:05.068 11:31:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:05.068 11:31:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:05.068 11:31:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:05.326 11:31:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:05.326 "name": "Existed_Raid", 00:21:05.326 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:05.326 "strip_size_kb": 0, 00:21:05.326 "state": "configuring", 00:21:05.326 "raid_level": "raid1", 00:21:05.326 "superblock": false, 00:21:05.326 "num_base_bdevs": 4, 00:21:05.326 "num_base_bdevs_discovered": 1, 00:21:05.326 "num_base_bdevs_operational": 4, 00:21:05.326 "base_bdevs_list": [ 00:21:05.326 { 00:21:05.326 "name": "BaseBdev1", 00:21:05.326 "uuid": "e7e220f1-6c42-4f6d-8266-07f4ae2fbb6e", 00:21:05.326 "is_configured": true, 00:21:05.326 "data_offset": 0, 00:21:05.326 "data_size": 65536 00:21:05.326 }, 00:21:05.326 { 00:21:05.326 "name": "BaseBdev2", 00:21:05.326 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:05.326 "is_configured": false, 00:21:05.326 "data_offset": 0, 00:21:05.326 "data_size": 0 00:21:05.326 }, 00:21:05.326 { 00:21:05.326 "name": "BaseBdev3", 00:21:05.326 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:05.326 "is_configured": false, 00:21:05.326 "data_offset": 0, 00:21:05.326 "data_size": 0 00:21:05.326 }, 00:21:05.326 { 00:21:05.326 "name": "BaseBdev4", 00:21:05.326 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:05.326 "is_configured": false, 00:21:05.326 "data_offset": 0, 00:21:05.326 "data_size": 0 00:21:05.326 } 00:21:05.326 ] 00:21:05.326 }' 00:21:05.326 11:31:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:05.326 11:31:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:05.891 11:31:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:21:06.149 [2024-07-15 11:31:49.581008] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:21:06.149 BaseBdev2 00:21:06.149 11:31:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:21:06.149 11:31:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:21:06.149 11:31:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:21:06.149 11:31:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:21:06.149 11:31:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:21:06.149 11:31:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:21:06.150 11:31:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:21:06.408 11:31:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:21:06.665 [ 00:21:06.665 { 00:21:06.665 "name": "BaseBdev2", 00:21:06.665 "aliases": [ 00:21:06.665 "fec97ce1-f36e-42c7-bfcd-2d1bf00c1f3e" 00:21:06.665 ], 00:21:06.665 "product_name": "Malloc disk", 00:21:06.665 "block_size": 512, 00:21:06.665 "num_blocks": 65536, 00:21:06.665 "uuid": "fec97ce1-f36e-42c7-bfcd-2d1bf00c1f3e", 00:21:06.665 "assigned_rate_limits": { 00:21:06.665 "rw_ios_per_sec": 0, 00:21:06.665 "rw_mbytes_per_sec": 0, 00:21:06.665 "r_mbytes_per_sec": 0, 00:21:06.665 "w_mbytes_per_sec": 0 00:21:06.665 }, 00:21:06.665 "claimed": true, 00:21:06.665 "claim_type": "exclusive_write", 00:21:06.665 "zoned": false, 00:21:06.665 "supported_io_types": { 00:21:06.665 "read": true, 00:21:06.665 "write": true, 00:21:06.665 "unmap": true, 00:21:06.665 "flush": true, 00:21:06.665 "reset": true, 00:21:06.665 "nvme_admin": false, 00:21:06.665 "nvme_io": false, 00:21:06.665 "nvme_io_md": false, 00:21:06.665 "write_zeroes": true, 00:21:06.665 "zcopy": true, 00:21:06.665 "get_zone_info": false, 00:21:06.665 "zone_management": false, 00:21:06.665 "zone_append": false, 00:21:06.666 "compare": false, 00:21:06.666 "compare_and_write": false, 00:21:06.666 "abort": true, 00:21:06.666 "seek_hole": false, 00:21:06.666 "seek_data": false, 00:21:06.666 "copy": true, 00:21:06.666 "nvme_iov_md": false 00:21:06.666 }, 00:21:06.666 "memory_domains": [ 00:21:06.666 { 00:21:06.666 "dma_device_id": "system", 00:21:06.666 "dma_device_type": 1 00:21:06.666 }, 00:21:06.666 { 00:21:06.666 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:06.666 "dma_device_type": 2 00:21:06.666 } 00:21:06.666 ], 00:21:06.666 "driver_specific": {} 00:21:06.666 } 00:21:06.666 ] 00:21:06.666 11:31:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:21:06.666 11:31:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:21:06.666 11:31:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:21:06.666 11:31:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:06.666 11:31:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:06.666 11:31:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:06.666 11:31:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:06.666 11:31:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:06.666 11:31:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:06.666 11:31:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:06.666 11:31:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:06.666 11:31:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:06.666 11:31:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:06.666 11:31:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:06.666 11:31:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:06.924 11:31:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:06.924 "name": "Existed_Raid", 00:21:06.924 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:06.924 "strip_size_kb": 0, 00:21:06.924 "state": "configuring", 00:21:06.924 "raid_level": "raid1", 00:21:06.924 "superblock": false, 00:21:06.924 "num_base_bdevs": 4, 00:21:06.924 "num_base_bdevs_discovered": 2, 00:21:06.924 "num_base_bdevs_operational": 4, 00:21:06.924 "base_bdevs_list": [ 00:21:06.924 { 00:21:06.924 "name": "BaseBdev1", 00:21:06.924 "uuid": "e7e220f1-6c42-4f6d-8266-07f4ae2fbb6e", 00:21:06.924 "is_configured": true, 00:21:06.924 "data_offset": 0, 00:21:06.924 "data_size": 65536 00:21:06.924 }, 00:21:06.924 { 00:21:06.924 "name": "BaseBdev2", 00:21:06.924 "uuid": "fec97ce1-f36e-42c7-bfcd-2d1bf00c1f3e", 00:21:06.924 "is_configured": true, 00:21:06.924 "data_offset": 0, 00:21:06.924 "data_size": 65536 00:21:06.924 }, 00:21:06.924 { 00:21:06.924 "name": "BaseBdev3", 00:21:06.924 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:06.924 "is_configured": false, 00:21:06.924 "data_offset": 0, 00:21:06.924 "data_size": 0 00:21:06.924 }, 00:21:06.924 { 00:21:06.924 "name": "BaseBdev4", 00:21:06.924 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:06.924 "is_configured": false, 00:21:06.924 "data_offset": 0, 00:21:06.924 "data_size": 0 00:21:06.924 } 00:21:06.924 ] 00:21:06.924 }' 00:21:06.924 11:31:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:06.924 11:31:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:07.489 11:31:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:21:07.748 [2024-07-15 11:31:51.092521] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:21:07.748 BaseBdev3 00:21:07.748 11:31:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev3 00:21:07.748 11:31:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:21:07.748 11:31:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:21:07.748 11:31:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:21:07.748 11:31:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:21:07.748 11:31:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:21:07.748 11:31:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:21:08.007 11:31:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:21:08.007 [ 00:21:08.007 { 00:21:08.007 "name": "BaseBdev3", 00:21:08.007 "aliases": [ 00:21:08.007 "a2bdb604-1be1-41c2-92da-3bb414937a2d" 00:21:08.007 ], 00:21:08.007 "product_name": "Malloc disk", 00:21:08.007 "block_size": 512, 00:21:08.007 "num_blocks": 65536, 00:21:08.007 "uuid": "a2bdb604-1be1-41c2-92da-3bb414937a2d", 00:21:08.007 "assigned_rate_limits": { 00:21:08.007 "rw_ios_per_sec": 0, 00:21:08.007 "rw_mbytes_per_sec": 0, 00:21:08.007 "r_mbytes_per_sec": 0, 00:21:08.007 "w_mbytes_per_sec": 0 00:21:08.007 }, 00:21:08.007 "claimed": true, 00:21:08.007 "claim_type": "exclusive_write", 00:21:08.007 "zoned": false, 00:21:08.007 "supported_io_types": { 00:21:08.007 "read": true, 00:21:08.007 "write": true, 00:21:08.007 "unmap": true, 00:21:08.007 "flush": true, 00:21:08.007 "reset": true, 00:21:08.007 "nvme_admin": false, 00:21:08.007 "nvme_io": false, 00:21:08.007 "nvme_io_md": false, 00:21:08.007 "write_zeroes": true, 00:21:08.007 "zcopy": true, 00:21:08.007 "get_zone_info": false, 00:21:08.007 "zone_management": false, 00:21:08.007 "zone_append": false, 00:21:08.007 "compare": false, 00:21:08.007 "compare_and_write": false, 00:21:08.007 "abort": true, 00:21:08.007 "seek_hole": false, 00:21:08.007 "seek_data": false, 00:21:08.007 "copy": true, 00:21:08.007 "nvme_iov_md": false 00:21:08.007 }, 00:21:08.007 "memory_domains": [ 00:21:08.007 { 00:21:08.007 "dma_device_id": "system", 00:21:08.007 "dma_device_type": 1 00:21:08.007 }, 00:21:08.007 { 00:21:08.007 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:08.007 "dma_device_type": 2 00:21:08.007 } 00:21:08.007 ], 00:21:08.007 "driver_specific": {} 00:21:08.007 } 00:21:08.007 ] 00:21:08.265 11:31:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:21:08.265 11:31:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:21:08.265 11:31:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:21:08.265 11:31:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:08.265 11:31:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:08.265 11:31:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:08.265 11:31:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:08.265 11:31:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:08.265 11:31:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:08.265 11:31:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:08.265 11:31:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:08.265 11:31:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:08.265 11:31:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:08.265 11:31:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:08.265 11:31:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:08.265 11:31:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:08.265 "name": "Existed_Raid", 00:21:08.265 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:08.265 "strip_size_kb": 0, 00:21:08.265 "state": "configuring", 00:21:08.265 "raid_level": "raid1", 00:21:08.265 "superblock": false, 00:21:08.265 "num_base_bdevs": 4, 00:21:08.265 "num_base_bdevs_discovered": 3, 00:21:08.265 "num_base_bdevs_operational": 4, 00:21:08.265 "base_bdevs_list": [ 00:21:08.265 { 00:21:08.265 "name": "BaseBdev1", 00:21:08.265 "uuid": "e7e220f1-6c42-4f6d-8266-07f4ae2fbb6e", 00:21:08.265 "is_configured": true, 00:21:08.265 "data_offset": 0, 00:21:08.265 "data_size": 65536 00:21:08.265 }, 00:21:08.265 { 00:21:08.265 "name": "BaseBdev2", 00:21:08.265 "uuid": "fec97ce1-f36e-42c7-bfcd-2d1bf00c1f3e", 00:21:08.265 "is_configured": true, 00:21:08.265 "data_offset": 0, 00:21:08.265 "data_size": 65536 00:21:08.265 }, 00:21:08.265 { 00:21:08.265 "name": "BaseBdev3", 00:21:08.265 "uuid": "a2bdb604-1be1-41c2-92da-3bb414937a2d", 00:21:08.265 "is_configured": true, 00:21:08.265 "data_offset": 0, 00:21:08.265 "data_size": 65536 00:21:08.265 }, 00:21:08.265 { 00:21:08.265 "name": "BaseBdev4", 00:21:08.265 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:08.265 "is_configured": false, 00:21:08.265 "data_offset": 0, 00:21:08.265 "data_size": 0 00:21:08.265 } 00:21:08.265 ] 00:21:08.265 }' 00:21:08.265 11:31:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:08.265 11:31:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:09.201 11:31:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4 00:21:09.201 [2024-07-15 11:31:52.668163] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:21:09.201 [2024-07-15 11:31:52.668206] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1d7c350 00:21:09.201 [2024-07-15 11:31:52.668216] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:21:09.201 [2024-07-15 11:31:52.668467] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1d7c020 00:21:09.201 [2024-07-15 11:31:52.668593] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1d7c350 00:21:09.201 [2024-07-15 11:31:52.668603] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x1d7c350 00:21:09.201 [2024-07-15 11:31:52.668769] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:21:09.201 BaseBdev4 00:21:09.201 11:31:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev4 00:21:09.201 11:31:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev4 00:21:09.201 11:31:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:21:09.201 11:31:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:21:09.201 11:31:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:21:09.201 11:31:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:21:09.201 11:31:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:21:09.459 11:31:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 -t 2000 00:21:09.717 [ 00:21:09.717 { 00:21:09.717 "name": "BaseBdev4", 00:21:09.717 "aliases": [ 00:21:09.717 "a49fb3c9-bffc-45e6-b016-5f19ecff40ec" 00:21:09.717 ], 00:21:09.717 "product_name": "Malloc disk", 00:21:09.717 "block_size": 512, 00:21:09.717 "num_blocks": 65536, 00:21:09.717 "uuid": "a49fb3c9-bffc-45e6-b016-5f19ecff40ec", 00:21:09.718 "assigned_rate_limits": { 00:21:09.718 "rw_ios_per_sec": 0, 00:21:09.718 "rw_mbytes_per_sec": 0, 00:21:09.718 "r_mbytes_per_sec": 0, 00:21:09.718 "w_mbytes_per_sec": 0 00:21:09.718 }, 00:21:09.718 "claimed": true, 00:21:09.718 "claim_type": "exclusive_write", 00:21:09.718 "zoned": false, 00:21:09.718 "supported_io_types": { 00:21:09.718 "read": true, 00:21:09.718 "write": true, 00:21:09.718 "unmap": true, 00:21:09.718 "flush": true, 00:21:09.718 "reset": true, 00:21:09.718 "nvme_admin": false, 00:21:09.718 "nvme_io": false, 00:21:09.718 "nvme_io_md": false, 00:21:09.718 "write_zeroes": true, 00:21:09.718 "zcopy": true, 00:21:09.718 "get_zone_info": false, 00:21:09.718 "zone_management": false, 00:21:09.718 "zone_append": false, 00:21:09.718 "compare": false, 00:21:09.718 "compare_and_write": false, 00:21:09.718 "abort": true, 00:21:09.718 "seek_hole": false, 00:21:09.718 "seek_data": false, 00:21:09.718 "copy": true, 00:21:09.718 "nvme_iov_md": false 00:21:09.718 }, 00:21:09.718 "memory_domains": [ 00:21:09.718 { 00:21:09.718 "dma_device_id": "system", 00:21:09.718 "dma_device_type": 1 00:21:09.718 }, 00:21:09.718 { 00:21:09.718 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:09.718 "dma_device_type": 2 00:21:09.718 } 00:21:09.718 ], 00:21:09.718 "driver_specific": {} 00:21:09.718 } 00:21:09.718 ] 00:21:09.718 11:31:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:21:09.718 11:31:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:21:09.718 11:31:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:21:09.718 11:31:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:21:09.718 11:31:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:09.718 11:31:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:21:09.718 11:31:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:09.718 11:31:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:09.718 11:31:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:09.718 11:31:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:09.718 11:31:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:09.718 11:31:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:09.718 11:31:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:09.718 11:31:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:09.718 11:31:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:09.976 11:31:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:09.976 "name": "Existed_Raid", 00:21:09.976 "uuid": "d68c6f00-6628-412e-aeb8-3e540fff3745", 00:21:09.976 "strip_size_kb": 0, 00:21:09.976 "state": "online", 00:21:09.976 "raid_level": "raid1", 00:21:09.976 "superblock": false, 00:21:09.976 "num_base_bdevs": 4, 00:21:09.976 "num_base_bdevs_discovered": 4, 00:21:09.976 "num_base_bdevs_operational": 4, 00:21:09.976 "base_bdevs_list": [ 00:21:09.976 { 00:21:09.976 "name": "BaseBdev1", 00:21:09.976 "uuid": "e7e220f1-6c42-4f6d-8266-07f4ae2fbb6e", 00:21:09.976 "is_configured": true, 00:21:09.976 "data_offset": 0, 00:21:09.976 "data_size": 65536 00:21:09.976 }, 00:21:09.976 { 00:21:09.976 "name": "BaseBdev2", 00:21:09.976 "uuid": "fec97ce1-f36e-42c7-bfcd-2d1bf00c1f3e", 00:21:09.976 "is_configured": true, 00:21:09.976 "data_offset": 0, 00:21:09.976 "data_size": 65536 00:21:09.976 }, 00:21:09.976 { 00:21:09.976 "name": "BaseBdev3", 00:21:09.976 "uuid": "a2bdb604-1be1-41c2-92da-3bb414937a2d", 00:21:09.976 "is_configured": true, 00:21:09.976 "data_offset": 0, 00:21:09.977 "data_size": 65536 00:21:09.977 }, 00:21:09.977 { 00:21:09.977 "name": "BaseBdev4", 00:21:09.977 "uuid": "a49fb3c9-bffc-45e6-b016-5f19ecff40ec", 00:21:09.977 "is_configured": true, 00:21:09.977 "data_offset": 0, 00:21:09.977 "data_size": 65536 00:21:09.977 } 00:21:09.977 ] 00:21:09.977 }' 00:21:09.977 11:31:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:09.977 11:31:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:10.543 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:21:10.543 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:21:10.543 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:21:10.543 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:21:10.543 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:21:10.543 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # local name 00:21:10.543 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:21:10.543 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:21:10.802 [2024-07-15 11:31:54.252695] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:21:10.802 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:21:10.802 "name": "Existed_Raid", 00:21:10.802 "aliases": [ 00:21:10.802 "d68c6f00-6628-412e-aeb8-3e540fff3745" 00:21:10.802 ], 00:21:10.802 "product_name": "Raid Volume", 00:21:10.802 "block_size": 512, 00:21:10.802 "num_blocks": 65536, 00:21:10.802 "uuid": "d68c6f00-6628-412e-aeb8-3e540fff3745", 00:21:10.802 "assigned_rate_limits": { 00:21:10.802 "rw_ios_per_sec": 0, 00:21:10.802 "rw_mbytes_per_sec": 0, 00:21:10.802 "r_mbytes_per_sec": 0, 00:21:10.802 "w_mbytes_per_sec": 0 00:21:10.802 }, 00:21:10.802 "claimed": false, 00:21:10.802 "zoned": false, 00:21:10.802 "supported_io_types": { 00:21:10.802 "read": true, 00:21:10.802 "write": true, 00:21:10.802 "unmap": false, 00:21:10.802 "flush": false, 00:21:10.802 "reset": true, 00:21:10.802 "nvme_admin": false, 00:21:10.802 "nvme_io": false, 00:21:10.802 "nvme_io_md": false, 00:21:10.802 "write_zeroes": true, 00:21:10.802 "zcopy": false, 00:21:10.802 "get_zone_info": false, 00:21:10.802 "zone_management": false, 00:21:10.802 "zone_append": false, 00:21:10.802 "compare": false, 00:21:10.802 "compare_and_write": false, 00:21:10.802 "abort": false, 00:21:10.802 "seek_hole": false, 00:21:10.802 "seek_data": false, 00:21:10.802 "copy": false, 00:21:10.802 "nvme_iov_md": false 00:21:10.802 }, 00:21:10.802 "memory_domains": [ 00:21:10.802 { 00:21:10.802 "dma_device_id": "system", 00:21:10.802 "dma_device_type": 1 00:21:10.802 }, 00:21:10.802 { 00:21:10.802 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:10.802 "dma_device_type": 2 00:21:10.802 }, 00:21:10.802 { 00:21:10.802 "dma_device_id": "system", 00:21:10.802 "dma_device_type": 1 00:21:10.802 }, 00:21:10.802 { 00:21:10.802 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:10.802 "dma_device_type": 2 00:21:10.802 }, 00:21:10.802 { 00:21:10.802 "dma_device_id": "system", 00:21:10.802 "dma_device_type": 1 00:21:10.802 }, 00:21:10.802 { 00:21:10.802 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:10.802 "dma_device_type": 2 00:21:10.802 }, 00:21:10.802 { 00:21:10.802 "dma_device_id": "system", 00:21:10.802 "dma_device_type": 1 00:21:10.802 }, 00:21:10.802 { 00:21:10.802 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:10.802 "dma_device_type": 2 00:21:10.802 } 00:21:10.802 ], 00:21:10.802 "driver_specific": { 00:21:10.802 "raid": { 00:21:10.802 "uuid": "d68c6f00-6628-412e-aeb8-3e540fff3745", 00:21:10.802 "strip_size_kb": 0, 00:21:10.802 "state": "online", 00:21:10.802 "raid_level": "raid1", 00:21:10.802 "superblock": false, 00:21:10.802 "num_base_bdevs": 4, 00:21:10.802 "num_base_bdevs_discovered": 4, 00:21:10.802 "num_base_bdevs_operational": 4, 00:21:10.802 "base_bdevs_list": [ 00:21:10.802 { 00:21:10.802 "name": "BaseBdev1", 00:21:10.802 "uuid": "e7e220f1-6c42-4f6d-8266-07f4ae2fbb6e", 00:21:10.802 "is_configured": true, 00:21:10.802 "data_offset": 0, 00:21:10.802 "data_size": 65536 00:21:10.802 }, 00:21:10.802 { 00:21:10.802 "name": "BaseBdev2", 00:21:10.802 "uuid": "fec97ce1-f36e-42c7-bfcd-2d1bf00c1f3e", 00:21:10.802 "is_configured": true, 00:21:10.802 "data_offset": 0, 00:21:10.802 "data_size": 65536 00:21:10.802 }, 00:21:10.802 { 00:21:10.802 "name": "BaseBdev3", 00:21:10.802 "uuid": "a2bdb604-1be1-41c2-92da-3bb414937a2d", 00:21:10.802 "is_configured": true, 00:21:10.802 "data_offset": 0, 00:21:10.802 "data_size": 65536 00:21:10.802 }, 00:21:10.802 { 00:21:10.802 "name": "BaseBdev4", 00:21:10.802 "uuid": "a49fb3c9-bffc-45e6-b016-5f19ecff40ec", 00:21:10.802 "is_configured": true, 00:21:10.802 "data_offset": 0, 00:21:10.802 "data_size": 65536 00:21:10.802 } 00:21:10.802 ] 00:21:10.802 } 00:21:10.802 } 00:21:10.802 }' 00:21:10.802 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:21:10.802 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:21:10.802 BaseBdev2 00:21:10.802 BaseBdev3 00:21:10.802 BaseBdev4' 00:21:10.802 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:21:10.802 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:21:10.802 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:21:11.061 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:21:11.061 "name": "BaseBdev1", 00:21:11.061 "aliases": [ 00:21:11.061 "e7e220f1-6c42-4f6d-8266-07f4ae2fbb6e" 00:21:11.061 ], 00:21:11.061 "product_name": "Malloc disk", 00:21:11.061 "block_size": 512, 00:21:11.061 "num_blocks": 65536, 00:21:11.061 "uuid": "e7e220f1-6c42-4f6d-8266-07f4ae2fbb6e", 00:21:11.061 "assigned_rate_limits": { 00:21:11.061 "rw_ios_per_sec": 0, 00:21:11.061 "rw_mbytes_per_sec": 0, 00:21:11.061 "r_mbytes_per_sec": 0, 00:21:11.061 "w_mbytes_per_sec": 0 00:21:11.061 }, 00:21:11.061 "claimed": true, 00:21:11.061 "claim_type": "exclusive_write", 00:21:11.061 "zoned": false, 00:21:11.061 "supported_io_types": { 00:21:11.061 "read": true, 00:21:11.061 "write": true, 00:21:11.061 "unmap": true, 00:21:11.061 "flush": true, 00:21:11.061 "reset": true, 00:21:11.061 "nvme_admin": false, 00:21:11.061 "nvme_io": false, 00:21:11.061 "nvme_io_md": false, 00:21:11.061 "write_zeroes": true, 00:21:11.061 "zcopy": true, 00:21:11.061 "get_zone_info": false, 00:21:11.061 "zone_management": false, 00:21:11.061 "zone_append": false, 00:21:11.061 "compare": false, 00:21:11.061 "compare_and_write": false, 00:21:11.061 "abort": true, 00:21:11.061 "seek_hole": false, 00:21:11.061 "seek_data": false, 00:21:11.061 "copy": true, 00:21:11.061 "nvme_iov_md": false 00:21:11.061 }, 00:21:11.061 "memory_domains": [ 00:21:11.061 { 00:21:11.061 "dma_device_id": "system", 00:21:11.061 "dma_device_type": 1 00:21:11.061 }, 00:21:11.061 { 00:21:11.061 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:11.061 "dma_device_type": 2 00:21:11.061 } 00:21:11.061 ], 00:21:11.061 "driver_specific": {} 00:21:11.061 }' 00:21:11.061 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:11.061 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:11.061 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:21:11.320 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:11.320 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:11.320 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:21:11.320 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:11.320 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:11.320 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:21:11.320 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:11.320 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:11.578 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:21:11.578 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:21:11.578 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:21:11.578 11:31:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:21:11.578 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:21:11.578 "name": "BaseBdev2", 00:21:11.578 "aliases": [ 00:21:11.578 "fec97ce1-f36e-42c7-bfcd-2d1bf00c1f3e" 00:21:11.578 ], 00:21:11.578 "product_name": "Malloc disk", 00:21:11.578 "block_size": 512, 00:21:11.578 "num_blocks": 65536, 00:21:11.578 "uuid": "fec97ce1-f36e-42c7-bfcd-2d1bf00c1f3e", 00:21:11.578 "assigned_rate_limits": { 00:21:11.578 "rw_ios_per_sec": 0, 00:21:11.578 "rw_mbytes_per_sec": 0, 00:21:11.578 "r_mbytes_per_sec": 0, 00:21:11.578 "w_mbytes_per_sec": 0 00:21:11.578 }, 00:21:11.578 "claimed": true, 00:21:11.578 "claim_type": "exclusive_write", 00:21:11.578 "zoned": false, 00:21:11.578 "supported_io_types": { 00:21:11.578 "read": true, 00:21:11.578 "write": true, 00:21:11.578 "unmap": true, 00:21:11.578 "flush": true, 00:21:11.578 "reset": true, 00:21:11.578 "nvme_admin": false, 00:21:11.578 "nvme_io": false, 00:21:11.578 "nvme_io_md": false, 00:21:11.578 "write_zeroes": true, 00:21:11.578 "zcopy": true, 00:21:11.578 "get_zone_info": false, 00:21:11.578 "zone_management": false, 00:21:11.578 "zone_append": false, 00:21:11.578 "compare": false, 00:21:11.578 "compare_and_write": false, 00:21:11.578 "abort": true, 00:21:11.578 "seek_hole": false, 00:21:11.578 "seek_data": false, 00:21:11.578 "copy": true, 00:21:11.578 "nvme_iov_md": false 00:21:11.578 }, 00:21:11.578 "memory_domains": [ 00:21:11.578 { 00:21:11.578 "dma_device_id": "system", 00:21:11.578 "dma_device_type": 1 00:21:11.578 }, 00:21:11.578 { 00:21:11.578 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:11.578 "dma_device_type": 2 00:21:11.578 } 00:21:11.578 ], 00:21:11.578 "driver_specific": {} 00:21:11.578 }' 00:21:11.578 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:11.836 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:11.836 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:21:11.836 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:11.836 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:11.836 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:21:11.836 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:11.836 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:12.094 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:21:12.094 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:12.094 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:12.094 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:21:12.094 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:21:12.094 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:21:12.094 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:21:12.352 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:21:12.352 "name": "BaseBdev3", 00:21:12.352 "aliases": [ 00:21:12.352 "a2bdb604-1be1-41c2-92da-3bb414937a2d" 00:21:12.352 ], 00:21:12.352 "product_name": "Malloc disk", 00:21:12.352 "block_size": 512, 00:21:12.352 "num_blocks": 65536, 00:21:12.352 "uuid": "a2bdb604-1be1-41c2-92da-3bb414937a2d", 00:21:12.352 "assigned_rate_limits": { 00:21:12.352 "rw_ios_per_sec": 0, 00:21:12.352 "rw_mbytes_per_sec": 0, 00:21:12.352 "r_mbytes_per_sec": 0, 00:21:12.352 "w_mbytes_per_sec": 0 00:21:12.352 }, 00:21:12.352 "claimed": true, 00:21:12.352 "claim_type": "exclusive_write", 00:21:12.352 "zoned": false, 00:21:12.352 "supported_io_types": { 00:21:12.352 "read": true, 00:21:12.352 "write": true, 00:21:12.352 "unmap": true, 00:21:12.352 "flush": true, 00:21:12.352 "reset": true, 00:21:12.352 "nvme_admin": false, 00:21:12.352 "nvme_io": false, 00:21:12.352 "nvme_io_md": false, 00:21:12.352 "write_zeroes": true, 00:21:12.352 "zcopy": true, 00:21:12.352 "get_zone_info": false, 00:21:12.352 "zone_management": false, 00:21:12.352 "zone_append": false, 00:21:12.352 "compare": false, 00:21:12.352 "compare_and_write": false, 00:21:12.352 "abort": true, 00:21:12.352 "seek_hole": false, 00:21:12.352 "seek_data": false, 00:21:12.352 "copy": true, 00:21:12.352 "nvme_iov_md": false 00:21:12.352 }, 00:21:12.352 "memory_domains": [ 00:21:12.352 { 00:21:12.352 "dma_device_id": "system", 00:21:12.352 "dma_device_type": 1 00:21:12.352 }, 00:21:12.352 { 00:21:12.352 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:12.352 "dma_device_type": 2 00:21:12.352 } 00:21:12.352 ], 00:21:12.352 "driver_specific": {} 00:21:12.352 }' 00:21:12.352 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:12.352 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:12.352 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:21:12.352 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:12.352 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:12.610 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:21:12.610 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:12.610 11:31:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:12.610 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:21:12.610 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:12.610 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:12.610 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:21:12.610 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:21:12.610 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 00:21:12.610 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:21:12.868 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:21:12.868 "name": "BaseBdev4", 00:21:12.868 "aliases": [ 00:21:12.868 "a49fb3c9-bffc-45e6-b016-5f19ecff40ec" 00:21:12.868 ], 00:21:12.868 "product_name": "Malloc disk", 00:21:12.868 "block_size": 512, 00:21:12.868 "num_blocks": 65536, 00:21:12.868 "uuid": "a49fb3c9-bffc-45e6-b016-5f19ecff40ec", 00:21:12.868 "assigned_rate_limits": { 00:21:12.869 "rw_ios_per_sec": 0, 00:21:12.869 "rw_mbytes_per_sec": 0, 00:21:12.869 "r_mbytes_per_sec": 0, 00:21:12.869 "w_mbytes_per_sec": 0 00:21:12.869 }, 00:21:12.869 "claimed": true, 00:21:12.869 "claim_type": "exclusive_write", 00:21:12.869 "zoned": false, 00:21:12.869 "supported_io_types": { 00:21:12.869 "read": true, 00:21:12.869 "write": true, 00:21:12.869 "unmap": true, 00:21:12.869 "flush": true, 00:21:12.869 "reset": true, 00:21:12.869 "nvme_admin": false, 00:21:12.869 "nvme_io": false, 00:21:12.869 "nvme_io_md": false, 00:21:12.869 "write_zeroes": true, 00:21:12.869 "zcopy": true, 00:21:12.869 "get_zone_info": false, 00:21:12.869 "zone_management": false, 00:21:12.869 "zone_append": false, 00:21:12.869 "compare": false, 00:21:12.869 "compare_and_write": false, 00:21:12.869 "abort": true, 00:21:12.869 "seek_hole": false, 00:21:12.869 "seek_data": false, 00:21:12.869 "copy": true, 00:21:12.869 "nvme_iov_md": false 00:21:12.869 }, 00:21:12.869 "memory_domains": [ 00:21:12.869 { 00:21:12.869 "dma_device_id": "system", 00:21:12.869 "dma_device_type": 1 00:21:12.869 }, 00:21:12.869 { 00:21:12.869 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:12.869 "dma_device_type": 2 00:21:12.869 } 00:21:12.869 ], 00:21:12.869 "driver_specific": {} 00:21:12.869 }' 00:21:12.869 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:12.869 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:13.127 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:21:13.127 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:13.127 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:13.127 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:21:13.127 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:13.127 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:13.127 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:21:13.127 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:13.127 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:13.127 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:21:13.127 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:21:13.385 [2024-07-15 11:31:56.931541] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:21:13.385 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@275 -- # local expected_state 00:21:13.385 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # has_redundancy raid1 00:21:13.385 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:21:13.385 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@214 -- # return 0 00:21:13.385 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # expected_state=online 00:21:13.385 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:21:13.385 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:13.385 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:21:13.385 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:13.385 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:13.385 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:21:13.385 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:13.385 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:13.385 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:13.385 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:13.385 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:13.385 11:31:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:13.643 11:31:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:13.643 "name": "Existed_Raid", 00:21:13.643 "uuid": "d68c6f00-6628-412e-aeb8-3e540fff3745", 00:21:13.643 "strip_size_kb": 0, 00:21:13.643 "state": "online", 00:21:13.643 "raid_level": "raid1", 00:21:13.643 "superblock": false, 00:21:13.643 "num_base_bdevs": 4, 00:21:13.643 "num_base_bdevs_discovered": 3, 00:21:13.643 "num_base_bdevs_operational": 3, 00:21:13.643 "base_bdevs_list": [ 00:21:13.643 { 00:21:13.643 "name": null, 00:21:13.643 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:13.643 "is_configured": false, 00:21:13.643 "data_offset": 0, 00:21:13.643 "data_size": 65536 00:21:13.643 }, 00:21:13.643 { 00:21:13.643 "name": "BaseBdev2", 00:21:13.643 "uuid": "fec97ce1-f36e-42c7-bfcd-2d1bf00c1f3e", 00:21:13.643 "is_configured": true, 00:21:13.643 "data_offset": 0, 00:21:13.643 "data_size": 65536 00:21:13.643 }, 00:21:13.643 { 00:21:13.643 "name": "BaseBdev3", 00:21:13.643 "uuid": "a2bdb604-1be1-41c2-92da-3bb414937a2d", 00:21:13.643 "is_configured": true, 00:21:13.643 "data_offset": 0, 00:21:13.643 "data_size": 65536 00:21:13.643 }, 00:21:13.643 { 00:21:13.643 "name": "BaseBdev4", 00:21:13.643 "uuid": "a49fb3c9-bffc-45e6-b016-5f19ecff40ec", 00:21:13.643 "is_configured": true, 00:21:13.643 "data_offset": 0, 00:21:13.643 "data_size": 65536 00:21:13.643 } 00:21:13.643 ] 00:21:13.643 }' 00:21:13.643 11:31:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:13.643 11:31:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:14.208 11:31:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:21:14.208 11:31:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:21:14.208 11:31:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:14.208 11:31:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:21:14.467 11:31:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:21:14.467 11:31:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:21:14.467 11:31:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:21:14.726 [2024-07-15 11:31:58.143822] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:21:14.726 11:31:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:21:14.726 11:31:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:21:14.726 11:31:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:14.726 11:31:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:21:14.984 11:31:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:21:14.984 11:31:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:21:14.984 11:31:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev3 00:21:14.984 [2024-07-15 11:31:58.511582] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:21:14.984 11:31:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:21:14.984 11:31:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:21:14.984 11:31:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:14.984 11:31:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:21:15.242 11:31:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:21:15.242 11:31:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:21:15.242 11:31:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev4 00:21:15.501 [2024-07-15 11:31:59.007541] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:21:15.501 [2024-07-15 11:31:59.007627] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:21:15.501 [2024-07-15 11:31:59.020354] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:21:15.501 [2024-07-15 11:31:59.020394] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:21:15.501 [2024-07-15 11:31:59.020407] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1d7c350 name Existed_Raid, state offline 00:21:15.501 11:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:21:15.501 11:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:21:15.501 11:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:15.501 11:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:21:15.758 11:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:21:15.758 11:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:21:15.758 11:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # '[' 4 -gt 2 ']' 00:21:15.758 11:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i = 1 )) 00:21:15.758 11:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:21:15.758 11:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:21:16.015 BaseBdev2 00:21:16.015 11:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev2 00:21:16.015 11:31:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:21:16.015 11:31:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:21:16.015 11:31:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:21:16.015 11:31:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:21:16.015 11:31:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:21:16.015 11:31:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:21:16.273 11:31:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:21:16.530 [ 00:21:16.530 { 00:21:16.530 "name": "BaseBdev2", 00:21:16.530 "aliases": [ 00:21:16.530 "b11d616e-f3ec-46fc-918b-013a4ad929f7" 00:21:16.530 ], 00:21:16.530 "product_name": "Malloc disk", 00:21:16.530 "block_size": 512, 00:21:16.530 "num_blocks": 65536, 00:21:16.530 "uuid": "b11d616e-f3ec-46fc-918b-013a4ad929f7", 00:21:16.530 "assigned_rate_limits": { 00:21:16.530 "rw_ios_per_sec": 0, 00:21:16.530 "rw_mbytes_per_sec": 0, 00:21:16.530 "r_mbytes_per_sec": 0, 00:21:16.530 "w_mbytes_per_sec": 0 00:21:16.530 }, 00:21:16.530 "claimed": false, 00:21:16.530 "zoned": false, 00:21:16.530 "supported_io_types": { 00:21:16.530 "read": true, 00:21:16.530 "write": true, 00:21:16.530 "unmap": true, 00:21:16.530 "flush": true, 00:21:16.530 "reset": true, 00:21:16.530 "nvme_admin": false, 00:21:16.530 "nvme_io": false, 00:21:16.530 "nvme_io_md": false, 00:21:16.530 "write_zeroes": true, 00:21:16.530 "zcopy": true, 00:21:16.530 "get_zone_info": false, 00:21:16.530 "zone_management": false, 00:21:16.530 "zone_append": false, 00:21:16.530 "compare": false, 00:21:16.530 "compare_and_write": false, 00:21:16.531 "abort": true, 00:21:16.531 "seek_hole": false, 00:21:16.531 "seek_data": false, 00:21:16.531 "copy": true, 00:21:16.531 "nvme_iov_md": false 00:21:16.531 }, 00:21:16.531 "memory_domains": [ 00:21:16.531 { 00:21:16.531 "dma_device_id": "system", 00:21:16.531 "dma_device_type": 1 00:21:16.531 }, 00:21:16.531 { 00:21:16.531 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:16.531 "dma_device_type": 2 00:21:16.531 } 00:21:16.531 ], 00:21:16.531 "driver_specific": {} 00:21:16.531 } 00:21:16.531 ] 00:21:16.531 11:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:21:16.531 11:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:21:16.531 11:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:21:16.531 11:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:21:16.788 BaseBdev3 00:21:16.788 11:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev3 00:21:16.788 11:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:21:16.788 11:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:21:16.788 11:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:21:16.788 11:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:21:16.788 11:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:21:16.788 11:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:21:17.046 11:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:21:17.303 [ 00:21:17.303 { 00:21:17.303 "name": "BaseBdev3", 00:21:17.303 "aliases": [ 00:21:17.303 "9dcaf87b-749b-4087-bd63-7407864f706a" 00:21:17.303 ], 00:21:17.303 "product_name": "Malloc disk", 00:21:17.303 "block_size": 512, 00:21:17.303 "num_blocks": 65536, 00:21:17.303 "uuid": "9dcaf87b-749b-4087-bd63-7407864f706a", 00:21:17.303 "assigned_rate_limits": { 00:21:17.303 "rw_ios_per_sec": 0, 00:21:17.303 "rw_mbytes_per_sec": 0, 00:21:17.303 "r_mbytes_per_sec": 0, 00:21:17.303 "w_mbytes_per_sec": 0 00:21:17.303 }, 00:21:17.303 "claimed": false, 00:21:17.303 "zoned": false, 00:21:17.303 "supported_io_types": { 00:21:17.303 "read": true, 00:21:17.303 "write": true, 00:21:17.303 "unmap": true, 00:21:17.303 "flush": true, 00:21:17.303 "reset": true, 00:21:17.303 "nvme_admin": false, 00:21:17.303 "nvme_io": false, 00:21:17.303 "nvme_io_md": false, 00:21:17.303 "write_zeroes": true, 00:21:17.303 "zcopy": true, 00:21:17.303 "get_zone_info": false, 00:21:17.303 "zone_management": false, 00:21:17.303 "zone_append": false, 00:21:17.303 "compare": false, 00:21:17.303 "compare_and_write": false, 00:21:17.303 "abort": true, 00:21:17.303 "seek_hole": false, 00:21:17.303 "seek_data": false, 00:21:17.303 "copy": true, 00:21:17.303 "nvme_iov_md": false 00:21:17.303 }, 00:21:17.303 "memory_domains": [ 00:21:17.303 { 00:21:17.303 "dma_device_id": "system", 00:21:17.303 "dma_device_type": 1 00:21:17.303 }, 00:21:17.303 { 00:21:17.303 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:17.303 "dma_device_type": 2 00:21:17.303 } 00:21:17.303 ], 00:21:17.303 "driver_specific": {} 00:21:17.303 } 00:21:17.303 ] 00:21:17.303 11:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:21:17.303 11:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:21:17.303 11:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:21:17.303 11:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4 00:21:17.582 BaseBdev4 00:21:17.582 11:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev4 00:21:17.582 11:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev4 00:21:17.582 11:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:21:17.582 11:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:21:17.582 11:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:21:17.582 11:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:21:17.582 11:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:21:17.883 11:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 -t 2000 00:21:17.883 [ 00:21:17.883 { 00:21:17.883 "name": "BaseBdev4", 00:21:17.883 "aliases": [ 00:21:17.883 "79be344d-3fcb-4007-a700-f22b518607d3" 00:21:17.883 ], 00:21:17.883 "product_name": "Malloc disk", 00:21:17.883 "block_size": 512, 00:21:17.883 "num_blocks": 65536, 00:21:17.883 "uuid": "79be344d-3fcb-4007-a700-f22b518607d3", 00:21:17.883 "assigned_rate_limits": { 00:21:17.883 "rw_ios_per_sec": 0, 00:21:17.883 "rw_mbytes_per_sec": 0, 00:21:17.883 "r_mbytes_per_sec": 0, 00:21:17.883 "w_mbytes_per_sec": 0 00:21:17.883 }, 00:21:17.883 "claimed": false, 00:21:17.883 "zoned": false, 00:21:17.883 "supported_io_types": { 00:21:17.883 "read": true, 00:21:17.883 "write": true, 00:21:17.883 "unmap": true, 00:21:17.883 "flush": true, 00:21:17.883 "reset": true, 00:21:17.883 "nvme_admin": false, 00:21:17.883 "nvme_io": false, 00:21:17.883 "nvme_io_md": false, 00:21:17.883 "write_zeroes": true, 00:21:17.883 "zcopy": true, 00:21:17.883 "get_zone_info": false, 00:21:17.883 "zone_management": false, 00:21:17.883 "zone_append": false, 00:21:17.883 "compare": false, 00:21:17.883 "compare_and_write": false, 00:21:17.883 "abort": true, 00:21:17.883 "seek_hole": false, 00:21:17.883 "seek_data": false, 00:21:17.883 "copy": true, 00:21:17.883 "nvme_iov_md": false 00:21:17.883 }, 00:21:17.883 "memory_domains": [ 00:21:17.883 { 00:21:17.883 "dma_device_id": "system", 00:21:17.883 "dma_device_type": 1 00:21:17.883 }, 00:21:17.883 { 00:21:17.883 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:17.883 "dma_device_type": 2 00:21:17.883 } 00:21:17.883 ], 00:21:17.883 "driver_specific": {} 00:21:17.883 } 00:21:17.883 ] 00:21:17.883 11:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:21:17.883 11:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:21:17.883 11:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:21:17.883 11:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@305 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:21:18.145 [2024-07-15 11:32:01.669784] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:21:18.145 [2024-07-15 11:32:01.669828] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:21:18.145 [2024-07-15 11:32:01.669847] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:21:18.145 [2024-07-15 11:32:01.671179] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:21:18.145 [2024-07-15 11:32:01.671222] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:21:18.145 11:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:18.145 11:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:18.145 11:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:18.145 11:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:18.145 11:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:18.145 11:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:18.145 11:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:18.145 11:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:18.145 11:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:18.145 11:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:18.146 11:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:18.146 11:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:18.402 11:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:18.402 "name": "Existed_Raid", 00:21:18.402 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:18.402 "strip_size_kb": 0, 00:21:18.402 "state": "configuring", 00:21:18.402 "raid_level": "raid1", 00:21:18.402 "superblock": false, 00:21:18.402 "num_base_bdevs": 4, 00:21:18.402 "num_base_bdevs_discovered": 3, 00:21:18.402 "num_base_bdevs_operational": 4, 00:21:18.402 "base_bdevs_list": [ 00:21:18.402 { 00:21:18.402 "name": "BaseBdev1", 00:21:18.402 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:18.402 "is_configured": false, 00:21:18.402 "data_offset": 0, 00:21:18.402 "data_size": 0 00:21:18.402 }, 00:21:18.402 { 00:21:18.402 "name": "BaseBdev2", 00:21:18.402 "uuid": "b11d616e-f3ec-46fc-918b-013a4ad929f7", 00:21:18.402 "is_configured": true, 00:21:18.402 "data_offset": 0, 00:21:18.402 "data_size": 65536 00:21:18.402 }, 00:21:18.402 { 00:21:18.402 "name": "BaseBdev3", 00:21:18.402 "uuid": "9dcaf87b-749b-4087-bd63-7407864f706a", 00:21:18.402 "is_configured": true, 00:21:18.402 "data_offset": 0, 00:21:18.402 "data_size": 65536 00:21:18.402 }, 00:21:18.402 { 00:21:18.402 "name": "BaseBdev4", 00:21:18.402 "uuid": "79be344d-3fcb-4007-a700-f22b518607d3", 00:21:18.402 "is_configured": true, 00:21:18.402 "data_offset": 0, 00:21:18.402 "data_size": 65536 00:21:18.402 } 00:21:18.402 ] 00:21:18.402 }' 00:21:18.402 11:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:18.402 11:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:18.967 11:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev2 00:21:19.225 [2024-07-15 11:32:02.684465] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:21:19.225 11:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@309 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:19.225 11:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:19.225 11:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:19.225 11:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:19.225 11:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:19.225 11:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:19.225 11:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:19.225 11:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:19.225 11:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:19.225 11:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:19.225 11:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:19.225 11:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:19.482 11:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:19.482 "name": "Existed_Raid", 00:21:19.482 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:19.482 "strip_size_kb": 0, 00:21:19.482 "state": "configuring", 00:21:19.482 "raid_level": "raid1", 00:21:19.482 "superblock": false, 00:21:19.482 "num_base_bdevs": 4, 00:21:19.482 "num_base_bdevs_discovered": 2, 00:21:19.482 "num_base_bdevs_operational": 4, 00:21:19.482 "base_bdevs_list": [ 00:21:19.482 { 00:21:19.482 "name": "BaseBdev1", 00:21:19.482 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:19.482 "is_configured": false, 00:21:19.482 "data_offset": 0, 00:21:19.482 "data_size": 0 00:21:19.482 }, 00:21:19.482 { 00:21:19.482 "name": null, 00:21:19.482 "uuid": "b11d616e-f3ec-46fc-918b-013a4ad929f7", 00:21:19.482 "is_configured": false, 00:21:19.482 "data_offset": 0, 00:21:19.482 "data_size": 65536 00:21:19.482 }, 00:21:19.482 { 00:21:19.482 "name": "BaseBdev3", 00:21:19.482 "uuid": "9dcaf87b-749b-4087-bd63-7407864f706a", 00:21:19.482 "is_configured": true, 00:21:19.482 "data_offset": 0, 00:21:19.482 "data_size": 65536 00:21:19.482 }, 00:21:19.482 { 00:21:19.482 "name": "BaseBdev4", 00:21:19.482 "uuid": "79be344d-3fcb-4007-a700-f22b518607d3", 00:21:19.483 "is_configured": true, 00:21:19.483 "data_offset": 0, 00:21:19.483 "data_size": 65536 00:21:19.483 } 00:21:19.483 ] 00:21:19.483 }' 00:21:19.483 11:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:19.483 11:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:20.046 11:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:20.046 11:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:21:20.046 11:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # [[ false == \f\a\l\s\e ]] 00:21:20.046 11:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:21:20.304 [2024-07-15 11:32:03.814848] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:21:20.304 BaseBdev1 00:21:20.304 11:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@313 -- # waitforbdev BaseBdev1 00:21:20.304 11:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:21:20.304 11:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:21:20.304 11:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:21:20.304 11:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:21:20.304 11:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:21:20.304 11:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:21:20.561 11:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:21:20.818 [ 00:21:20.818 { 00:21:20.818 "name": "BaseBdev1", 00:21:20.818 "aliases": [ 00:21:20.818 "b677bc61-24ba-49d1-9d8b-103a02b3841e" 00:21:20.818 ], 00:21:20.818 "product_name": "Malloc disk", 00:21:20.818 "block_size": 512, 00:21:20.818 "num_blocks": 65536, 00:21:20.818 "uuid": "b677bc61-24ba-49d1-9d8b-103a02b3841e", 00:21:20.818 "assigned_rate_limits": { 00:21:20.818 "rw_ios_per_sec": 0, 00:21:20.818 "rw_mbytes_per_sec": 0, 00:21:20.818 "r_mbytes_per_sec": 0, 00:21:20.819 "w_mbytes_per_sec": 0 00:21:20.819 }, 00:21:20.819 "claimed": true, 00:21:20.819 "claim_type": "exclusive_write", 00:21:20.819 "zoned": false, 00:21:20.819 "supported_io_types": { 00:21:20.819 "read": true, 00:21:20.819 "write": true, 00:21:20.819 "unmap": true, 00:21:20.819 "flush": true, 00:21:20.819 "reset": true, 00:21:20.819 "nvme_admin": false, 00:21:20.819 "nvme_io": false, 00:21:20.819 "nvme_io_md": false, 00:21:20.819 "write_zeroes": true, 00:21:20.819 "zcopy": true, 00:21:20.819 "get_zone_info": false, 00:21:20.819 "zone_management": false, 00:21:20.819 "zone_append": false, 00:21:20.819 "compare": false, 00:21:20.819 "compare_and_write": false, 00:21:20.819 "abort": true, 00:21:20.819 "seek_hole": false, 00:21:20.819 "seek_data": false, 00:21:20.819 "copy": true, 00:21:20.819 "nvme_iov_md": false 00:21:20.819 }, 00:21:20.819 "memory_domains": [ 00:21:20.819 { 00:21:20.819 "dma_device_id": "system", 00:21:20.819 "dma_device_type": 1 00:21:20.819 }, 00:21:20.819 { 00:21:20.819 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:20.819 "dma_device_type": 2 00:21:20.819 } 00:21:20.819 ], 00:21:20.819 "driver_specific": {} 00:21:20.819 } 00:21:20.819 ] 00:21:20.819 11:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:21:20.819 11:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:20.819 11:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:20.819 11:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:20.819 11:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:20.819 11:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:20.819 11:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:20.819 11:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:20.819 11:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:20.819 11:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:20.819 11:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:20.819 11:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:20.819 11:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:21.077 11:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:21.077 "name": "Existed_Raid", 00:21:21.077 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:21.077 "strip_size_kb": 0, 00:21:21.077 "state": "configuring", 00:21:21.077 "raid_level": "raid1", 00:21:21.077 "superblock": false, 00:21:21.077 "num_base_bdevs": 4, 00:21:21.077 "num_base_bdevs_discovered": 3, 00:21:21.077 "num_base_bdevs_operational": 4, 00:21:21.077 "base_bdevs_list": [ 00:21:21.077 { 00:21:21.077 "name": "BaseBdev1", 00:21:21.077 "uuid": "b677bc61-24ba-49d1-9d8b-103a02b3841e", 00:21:21.077 "is_configured": true, 00:21:21.077 "data_offset": 0, 00:21:21.077 "data_size": 65536 00:21:21.077 }, 00:21:21.077 { 00:21:21.077 "name": null, 00:21:21.077 "uuid": "b11d616e-f3ec-46fc-918b-013a4ad929f7", 00:21:21.077 "is_configured": false, 00:21:21.077 "data_offset": 0, 00:21:21.077 "data_size": 65536 00:21:21.077 }, 00:21:21.077 { 00:21:21.077 "name": "BaseBdev3", 00:21:21.077 "uuid": "9dcaf87b-749b-4087-bd63-7407864f706a", 00:21:21.077 "is_configured": true, 00:21:21.077 "data_offset": 0, 00:21:21.077 "data_size": 65536 00:21:21.077 }, 00:21:21.077 { 00:21:21.077 "name": "BaseBdev4", 00:21:21.077 "uuid": "79be344d-3fcb-4007-a700-f22b518607d3", 00:21:21.077 "is_configured": true, 00:21:21.077 "data_offset": 0, 00:21:21.077 "data_size": 65536 00:21:21.077 } 00:21:21.077 ] 00:21:21.077 }' 00:21:21.077 11:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:21.077 11:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:21.641 11:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:21.641 11:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:21:21.898 11:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # [[ true == \t\r\u\e ]] 00:21:21.898 11:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@317 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev3 00:21:22.155 [2024-07-15 11:32:05.579559] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:21:22.155 11:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:22.155 11:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:22.155 11:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:22.155 11:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:22.155 11:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:22.155 11:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:22.155 11:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:22.155 11:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:22.155 11:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:22.155 11:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:22.155 11:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:22.155 11:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:22.411 11:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:22.411 "name": "Existed_Raid", 00:21:22.411 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:22.411 "strip_size_kb": 0, 00:21:22.411 "state": "configuring", 00:21:22.411 "raid_level": "raid1", 00:21:22.411 "superblock": false, 00:21:22.411 "num_base_bdevs": 4, 00:21:22.411 "num_base_bdevs_discovered": 2, 00:21:22.411 "num_base_bdevs_operational": 4, 00:21:22.411 "base_bdevs_list": [ 00:21:22.411 { 00:21:22.411 "name": "BaseBdev1", 00:21:22.411 "uuid": "b677bc61-24ba-49d1-9d8b-103a02b3841e", 00:21:22.411 "is_configured": true, 00:21:22.411 "data_offset": 0, 00:21:22.411 "data_size": 65536 00:21:22.411 }, 00:21:22.411 { 00:21:22.411 "name": null, 00:21:22.411 "uuid": "b11d616e-f3ec-46fc-918b-013a4ad929f7", 00:21:22.411 "is_configured": false, 00:21:22.411 "data_offset": 0, 00:21:22.411 "data_size": 65536 00:21:22.411 }, 00:21:22.411 { 00:21:22.411 "name": null, 00:21:22.411 "uuid": "9dcaf87b-749b-4087-bd63-7407864f706a", 00:21:22.411 "is_configured": false, 00:21:22.411 "data_offset": 0, 00:21:22.411 "data_size": 65536 00:21:22.411 }, 00:21:22.411 { 00:21:22.411 "name": "BaseBdev4", 00:21:22.411 "uuid": "79be344d-3fcb-4007-a700-f22b518607d3", 00:21:22.411 "is_configured": true, 00:21:22.411 "data_offset": 0, 00:21:22.411 "data_size": 65536 00:21:22.411 } 00:21:22.411 ] 00:21:22.411 }' 00:21:22.411 11:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:22.411 11:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:22.974 11:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:21:22.974 11:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:23.232 11:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # [[ false == \f\a\l\s\e ]] 00:21:23.232 11:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:21:23.490 [2024-07-15 11:32:06.923146] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:21:23.490 11:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@322 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:23.490 11:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:23.490 11:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:23.490 11:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:23.490 11:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:23.490 11:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:23.490 11:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:23.490 11:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:23.490 11:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:23.490 11:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:23.490 11:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:23.490 11:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:23.748 11:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:23.748 "name": "Existed_Raid", 00:21:23.748 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:23.748 "strip_size_kb": 0, 00:21:23.748 "state": "configuring", 00:21:23.748 "raid_level": "raid1", 00:21:23.748 "superblock": false, 00:21:23.748 "num_base_bdevs": 4, 00:21:23.748 "num_base_bdevs_discovered": 3, 00:21:23.748 "num_base_bdevs_operational": 4, 00:21:23.748 "base_bdevs_list": [ 00:21:23.748 { 00:21:23.748 "name": "BaseBdev1", 00:21:23.748 "uuid": "b677bc61-24ba-49d1-9d8b-103a02b3841e", 00:21:23.748 "is_configured": true, 00:21:23.748 "data_offset": 0, 00:21:23.748 "data_size": 65536 00:21:23.748 }, 00:21:23.748 { 00:21:23.748 "name": null, 00:21:23.748 "uuid": "b11d616e-f3ec-46fc-918b-013a4ad929f7", 00:21:23.748 "is_configured": false, 00:21:23.748 "data_offset": 0, 00:21:23.748 "data_size": 65536 00:21:23.748 }, 00:21:23.748 { 00:21:23.748 "name": "BaseBdev3", 00:21:23.748 "uuid": "9dcaf87b-749b-4087-bd63-7407864f706a", 00:21:23.748 "is_configured": true, 00:21:23.748 "data_offset": 0, 00:21:23.748 "data_size": 65536 00:21:23.748 }, 00:21:23.748 { 00:21:23.748 "name": "BaseBdev4", 00:21:23.748 "uuid": "79be344d-3fcb-4007-a700-f22b518607d3", 00:21:23.748 "is_configured": true, 00:21:23.748 "data_offset": 0, 00:21:23.748 "data_size": 65536 00:21:23.748 } 00:21:23.749 ] 00:21:23.749 }' 00:21:23.749 11:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:23.749 11:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:24.312 11:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:21:24.312 11:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:24.570 11:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # [[ true == \t\r\u\e ]] 00:21:24.570 11:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@325 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:21:24.827 [2024-07-15 11:32:08.238783] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:21:24.827 11:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:24.827 11:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:24.827 11:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:24.827 11:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:24.827 11:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:24.827 11:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:24.827 11:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:24.827 11:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:24.827 11:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:24.827 11:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:24.827 11:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:24.827 11:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:25.085 11:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:25.085 "name": "Existed_Raid", 00:21:25.085 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:25.085 "strip_size_kb": 0, 00:21:25.085 "state": "configuring", 00:21:25.085 "raid_level": "raid1", 00:21:25.085 "superblock": false, 00:21:25.085 "num_base_bdevs": 4, 00:21:25.085 "num_base_bdevs_discovered": 2, 00:21:25.085 "num_base_bdevs_operational": 4, 00:21:25.085 "base_bdevs_list": [ 00:21:25.085 { 00:21:25.085 "name": null, 00:21:25.085 "uuid": "b677bc61-24ba-49d1-9d8b-103a02b3841e", 00:21:25.085 "is_configured": false, 00:21:25.085 "data_offset": 0, 00:21:25.085 "data_size": 65536 00:21:25.085 }, 00:21:25.085 { 00:21:25.085 "name": null, 00:21:25.085 "uuid": "b11d616e-f3ec-46fc-918b-013a4ad929f7", 00:21:25.085 "is_configured": false, 00:21:25.085 "data_offset": 0, 00:21:25.085 "data_size": 65536 00:21:25.085 }, 00:21:25.085 { 00:21:25.085 "name": "BaseBdev3", 00:21:25.085 "uuid": "9dcaf87b-749b-4087-bd63-7407864f706a", 00:21:25.085 "is_configured": true, 00:21:25.085 "data_offset": 0, 00:21:25.085 "data_size": 65536 00:21:25.085 }, 00:21:25.085 { 00:21:25.085 "name": "BaseBdev4", 00:21:25.085 "uuid": "79be344d-3fcb-4007-a700-f22b518607d3", 00:21:25.085 "is_configured": true, 00:21:25.085 "data_offset": 0, 00:21:25.085 "data_size": 65536 00:21:25.085 } 00:21:25.085 ] 00:21:25.085 }' 00:21:25.085 11:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:25.085 11:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:25.652 11:32:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@327 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:25.652 11:32:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@327 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:21:25.910 11:32:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@327 -- # [[ false == \f\a\l\s\e ]] 00:21:25.910 11:32:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@329 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:21:26.168 [2024-07-15 11:32:09.578746] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:21:26.168 11:32:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@330 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:26.168 11:32:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:26.168 11:32:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:26.168 11:32:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:26.168 11:32:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:26.168 11:32:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:26.168 11:32:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:26.168 11:32:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:26.168 11:32:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:26.168 11:32:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:26.168 11:32:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:26.168 11:32:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:26.426 11:32:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:26.426 "name": "Existed_Raid", 00:21:26.426 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:26.426 "strip_size_kb": 0, 00:21:26.426 "state": "configuring", 00:21:26.426 "raid_level": "raid1", 00:21:26.426 "superblock": false, 00:21:26.426 "num_base_bdevs": 4, 00:21:26.426 "num_base_bdevs_discovered": 3, 00:21:26.426 "num_base_bdevs_operational": 4, 00:21:26.426 "base_bdevs_list": [ 00:21:26.426 { 00:21:26.426 "name": null, 00:21:26.426 "uuid": "b677bc61-24ba-49d1-9d8b-103a02b3841e", 00:21:26.426 "is_configured": false, 00:21:26.426 "data_offset": 0, 00:21:26.426 "data_size": 65536 00:21:26.427 }, 00:21:26.427 { 00:21:26.427 "name": "BaseBdev2", 00:21:26.427 "uuid": "b11d616e-f3ec-46fc-918b-013a4ad929f7", 00:21:26.427 "is_configured": true, 00:21:26.427 "data_offset": 0, 00:21:26.427 "data_size": 65536 00:21:26.427 }, 00:21:26.427 { 00:21:26.427 "name": "BaseBdev3", 00:21:26.427 "uuid": "9dcaf87b-749b-4087-bd63-7407864f706a", 00:21:26.427 "is_configured": true, 00:21:26.427 "data_offset": 0, 00:21:26.427 "data_size": 65536 00:21:26.427 }, 00:21:26.427 { 00:21:26.427 "name": "BaseBdev4", 00:21:26.427 "uuid": "79be344d-3fcb-4007-a700-f22b518607d3", 00:21:26.427 "is_configured": true, 00:21:26.427 "data_offset": 0, 00:21:26.427 "data_size": 65536 00:21:26.427 } 00:21:26.427 ] 00:21:26.427 }' 00:21:26.427 11:32:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:26.427 11:32:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:26.993 11:32:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@331 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:26.993 11:32:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@331 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:21:27.251 11:32:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@331 -- # [[ true == \t\r\u\e ]] 00:21:27.252 11:32:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@333 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:21:27.252 11:32:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:27.570 11:32:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b NewBaseBdev -u b677bc61-24ba-49d1-9d8b-103a02b3841e 00:21:27.570 [2024-07-15 11:32:11.083332] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:21:27.570 [2024-07-15 11:32:11.083381] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1d7a610 00:21:27.570 [2024-07-15 11:32:11.083390] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:21:27.570 [2024-07-15 11:32:11.083593] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1d7ba70 00:21:27.570 [2024-07-15 11:32:11.083718] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1d7a610 00:21:27.570 [2024-07-15 11:32:11.083728] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x1d7a610 00:21:27.570 [2024-07-15 11:32:11.083905] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:21:27.570 NewBaseBdev 00:21:27.570 11:32:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@334 -- # waitforbdev NewBaseBdev 00:21:27.570 11:32:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@897 -- # local bdev_name=NewBaseBdev 00:21:27.570 11:32:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:21:27.570 11:32:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local i 00:21:27.570 11:32:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:21:27.570 11:32:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:21:27.570 11:32:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:21:27.828 11:32:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev -t 2000 00:21:28.085 [ 00:21:28.085 { 00:21:28.085 "name": "NewBaseBdev", 00:21:28.085 "aliases": [ 00:21:28.085 "b677bc61-24ba-49d1-9d8b-103a02b3841e" 00:21:28.085 ], 00:21:28.085 "product_name": "Malloc disk", 00:21:28.085 "block_size": 512, 00:21:28.085 "num_blocks": 65536, 00:21:28.085 "uuid": "b677bc61-24ba-49d1-9d8b-103a02b3841e", 00:21:28.085 "assigned_rate_limits": { 00:21:28.085 "rw_ios_per_sec": 0, 00:21:28.085 "rw_mbytes_per_sec": 0, 00:21:28.085 "r_mbytes_per_sec": 0, 00:21:28.085 "w_mbytes_per_sec": 0 00:21:28.085 }, 00:21:28.085 "claimed": true, 00:21:28.085 "claim_type": "exclusive_write", 00:21:28.085 "zoned": false, 00:21:28.085 "supported_io_types": { 00:21:28.085 "read": true, 00:21:28.085 "write": true, 00:21:28.085 "unmap": true, 00:21:28.085 "flush": true, 00:21:28.085 "reset": true, 00:21:28.085 "nvme_admin": false, 00:21:28.085 "nvme_io": false, 00:21:28.085 "nvme_io_md": false, 00:21:28.085 "write_zeroes": true, 00:21:28.085 "zcopy": true, 00:21:28.085 "get_zone_info": false, 00:21:28.085 "zone_management": false, 00:21:28.085 "zone_append": false, 00:21:28.085 "compare": false, 00:21:28.085 "compare_and_write": false, 00:21:28.085 "abort": true, 00:21:28.085 "seek_hole": false, 00:21:28.085 "seek_data": false, 00:21:28.085 "copy": true, 00:21:28.085 "nvme_iov_md": false 00:21:28.085 }, 00:21:28.085 "memory_domains": [ 00:21:28.085 { 00:21:28.085 "dma_device_id": "system", 00:21:28.085 "dma_device_type": 1 00:21:28.085 }, 00:21:28.085 { 00:21:28.085 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:28.085 "dma_device_type": 2 00:21:28.085 } 00:21:28.085 ], 00:21:28.085 "driver_specific": {} 00:21:28.085 } 00:21:28.085 ] 00:21:28.085 11:32:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # return 0 00:21:28.085 11:32:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@335 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:21:28.085 11:32:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:28.085 11:32:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:21:28.085 11:32:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:28.085 11:32:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:28.085 11:32:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:28.085 11:32:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:28.085 11:32:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:28.085 11:32:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:28.085 11:32:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:28.085 11:32:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:28.085 11:32:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:28.343 11:32:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:28.343 "name": "Existed_Raid", 00:21:28.343 "uuid": "684673ac-d98b-4600-a87d-70ae77967caa", 00:21:28.343 "strip_size_kb": 0, 00:21:28.343 "state": "online", 00:21:28.343 "raid_level": "raid1", 00:21:28.343 "superblock": false, 00:21:28.343 "num_base_bdevs": 4, 00:21:28.343 "num_base_bdevs_discovered": 4, 00:21:28.343 "num_base_bdevs_operational": 4, 00:21:28.343 "base_bdevs_list": [ 00:21:28.343 { 00:21:28.343 "name": "NewBaseBdev", 00:21:28.343 "uuid": "b677bc61-24ba-49d1-9d8b-103a02b3841e", 00:21:28.343 "is_configured": true, 00:21:28.343 "data_offset": 0, 00:21:28.343 "data_size": 65536 00:21:28.343 }, 00:21:28.343 { 00:21:28.343 "name": "BaseBdev2", 00:21:28.343 "uuid": "b11d616e-f3ec-46fc-918b-013a4ad929f7", 00:21:28.343 "is_configured": true, 00:21:28.343 "data_offset": 0, 00:21:28.343 "data_size": 65536 00:21:28.343 }, 00:21:28.343 { 00:21:28.343 "name": "BaseBdev3", 00:21:28.343 "uuid": "9dcaf87b-749b-4087-bd63-7407864f706a", 00:21:28.343 "is_configured": true, 00:21:28.343 "data_offset": 0, 00:21:28.343 "data_size": 65536 00:21:28.343 }, 00:21:28.343 { 00:21:28.343 "name": "BaseBdev4", 00:21:28.343 "uuid": "79be344d-3fcb-4007-a700-f22b518607d3", 00:21:28.343 "is_configured": true, 00:21:28.343 "data_offset": 0, 00:21:28.343 "data_size": 65536 00:21:28.343 } 00:21:28.343 ] 00:21:28.343 }' 00:21:28.343 11:32:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:28.343 11:32:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:28.908 11:32:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@336 -- # verify_raid_bdev_properties Existed_Raid 00:21:28.908 11:32:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:21:28.908 11:32:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:21:28.908 11:32:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:21:28.908 11:32:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:21:28.908 11:32:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # local name 00:21:28.908 11:32:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:21:28.908 11:32:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:21:29.167 [2024-07-15 11:32:12.571609] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:21:29.167 11:32:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:21:29.167 "name": "Existed_Raid", 00:21:29.167 "aliases": [ 00:21:29.167 "684673ac-d98b-4600-a87d-70ae77967caa" 00:21:29.167 ], 00:21:29.167 "product_name": "Raid Volume", 00:21:29.167 "block_size": 512, 00:21:29.167 "num_blocks": 65536, 00:21:29.167 "uuid": "684673ac-d98b-4600-a87d-70ae77967caa", 00:21:29.167 "assigned_rate_limits": { 00:21:29.167 "rw_ios_per_sec": 0, 00:21:29.167 "rw_mbytes_per_sec": 0, 00:21:29.167 "r_mbytes_per_sec": 0, 00:21:29.167 "w_mbytes_per_sec": 0 00:21:29.167 }, 00:21:29.168 "claimed": false, 00:21:29.168 "zoned": false, 00:21:29.168 "supported_io_types": { 00:21:29.168 "read": true, 00:21:29.168 "write": true, 00:21:29.168 "unmap": false, 00:21:29.168 "flush": false, 00:21:29.168 "reset": true, 00:21:29.168 "nvme_admin": false, 00:21:29.168 "nvme_io": false, 00:21:29.168 "nvme_io_md": false, 00:21:29.168 "write_zeroes": true, 00:21:29.168 "zcopy": false, 00:21:29.168 "get_zone_info": false, 00:21:29.168 "zone_management": false, 00:21:29.168 "zone_append": false, 00:21:29.168 "compare": false, 00:21:29.168 "compare_and_write": false, 00:21:29.168 "abort": false, 00:21:29.168 "seek_hole": false, 00:21:29.168 "seek_data": false, 00:21:29.168 "copy": false, 00:21:29.168 "nvme_iov_md": false 00:21:29.168 }, 00:21:29.168 "memory_domains": [ 00:21:29.168 { 00:21:29.168 "dma_device_id": "system", 00:21:29.168 "dma_device_type": 1 00:21:29.168 }, 00:21:29.168 { 00:21:29.168 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:29.168 "dma_device_type": 2 00:21:29.168 }, 00:21:29.168 { 00:21:29.168 "dma_device_id": "system", 00:21:29.168 "dma_device_type": 1 00:21:29.168 }, 00:21:29.168 { 00:21:29.168 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:29.168 "dma_device_type": 2 00:21:29.168 }, 00:21:29.168 { 00:21:29.168 "dma_device_id": "system", 00:21:29.168 "dma_device_type": 1 00:21:29.168 }, 00:21:29.168 { 00:21:29.168 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:29.168 "dma_device_type": 2 00:21:29.168 }, 00:21:29.168 { 00:21:29.168 "dma_device_id": "system", 00:21:29.168 "dma_device_type": 1 00:21:29.168 }, 00:21:29.168 { 00:21:29.168 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:29.168 "dma_device_type": 2 00:21:29.168 } 00:21:29.168 ], 00:21:29.168 "driver_specific": { 00:21:29.168 "raid": { 00:21:29.168 "uuid": "684673ac-d98b-4600-a87d-70ae77967caa", 00:21:29.168 "strip_size_kb": 0, 00:21:29.168 "state": "online", 00:21:29.168 "raid_level": "raid1", 00:21:29.168 "superblock": false, 00:21:29.168 "num_base_bdevs": 4, 00:21:29.168 "num_base_bdevs_discovered": 4, 00:21:29.168 "num_base_bdevs_operational": 4, 00:21:29.168 "base_bdevs_list": [ 00:21:29.168 { 00:21:29.168 "name": "NewBaseBdev", 00:21:29.168 "uuid": "b677bc61-24ba-49d1-9d8b-103a02b3841e", 00:21:29.168 "is_configured": true, 00:21:29.168 "data_offset": 0, 00:21:29.168 "data_size": 65536 00:21:29.168 }, 00:21:29.168 { 00:21:29.168 "name": "BaseBdev2", 00:21:29.168 "uuid": "b11d616e-f3ec-46fc-918b-013a4ad929f7", 00:21:29.168 "is_configured": true, 00:21:29.168 "data_offset": 0, 00:21:29.168 "data_size": 65536 00:21:29.168 }, 00:21:29.168 { 00:21:29.168 "name": "BaseBdev3", 00:21:29.168 "uuid": "9dcaf87b-749b-4087-bd63-7407864f706a", 00:21:29.168 "is_configured": true, 00:21:29.168 "data_offset": 0, 00:21:29.168 "data_size": 65536 00:21:29.168 }, 00:21:29.168 { 00:21:29.168 "name": "BaseBdev4", 00:21:29.168 "uuid": "79be344d-3fcb-4007-a700-f22b518607d3", 00:21:29.168 "is_configured": true, 00:21:29.168 "data_offset": 0, 00:21:29.168 "data_size": 65536 00:21:29.168 } 00:21:29.168 ] 00:21:29.168 } 00:21:29.168 } 00:21:29.168 }' 00:21:29.168 11:32:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:21:29.168 11:32:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='NewBaseBdev 00:21:29.168 BaseBdev2 00:21:29.168 BaseBdev3 00:21:29.168 BaseBdev4' 00:21:29.168 11:32:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:21:29.168 11:32:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev 00:21:29.168 11:32:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:21:29.425 11:32:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:21:29.425 "name": "NewBaseBdev", 00:21:29.425 "aliases": [ 00:21:29.425 "b677bc61-24ba-49d1-9d8b-103a02b3841e" 00:21:29.425 ], 00:21:29.425 "product_name": "Malloc disk", 00:21:29.425 "block_size": 512, 00:21:29.425 "num_blocks": 65536, 00:21:29.425 "uuid": "b677bc61-24ba-49d1-9d8b-103a02b3841e", 00:21:29.425 "assigned_rate_limits": { 00:21:29.425 "rw_ios_per_sec": 0, 00:21:29.425 "rw_mbytes_per_sec": 0, 00:21:29.425 "r_mbytes_per_sec": 0, 00:21:29.425 "w_mbytes_per_sec": 0 00:21:29.425 }, 00:21:29.425 "claimed": true, 00:21:29.425 "claim_type": "exclusive_write", 00:21:29.425 "zoned": false, 00:21:29.425 "supported_io_types": { 00:21:29.425 "read": true, 00:21:29.425 "write": true, 00:21:29.425 "unmap": true, 00:21:29.425 "flush": true, 00:21:29.425 "reset": true, 00:21:29.425 "nvme_admin": false, 00:21:29.425 "nvme_io": false, 00:21:29.425 "nvme_io_md": false, 00:21:29.425 "write_zeroes": true, 00:21:29.425 "zcopy": true, 00:21:29.425 "get_zone_info": false, 00:21:29.425 "zone_management": false, 00:21:29.425 "zone_append": false, 00:21:29.425 "compare": false, 00:21:29.425 "compare_and_write": false, 00:21:29.425 "abort": true, 00:21:29.425 "seek_hole": false, 00:21:29.425 "seek_data": false, 00:21:29.425 "copy": true, 00:21:29.425 "nvme_iov_md": false 00:21:29.425 }, 00:21:29.426 "memory_domains": [ 00:21:29.426 { 00:21:29.426 "dma_device_id": "system", 00:21:29.426 "dma_device_type": 1 00:21:29.426 }, 00:21:29.426 { 00:21:29.426 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:29.426 "dma_device_type": 2 00:21:29.426 } 00:21:29.426 ], 00:21:29.426 "driver_specific": {} 00:21:29.426 }' 00:21:29.426 11:32:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:29.426 11:32:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:29.426 11:32:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:21:29.426 11:32:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:29.426 11:32:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:29.683 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:21:29.683 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:29.683 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:29.683 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:21:29.683 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:29.683 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:29.683 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:21:29.683 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:21:29.683 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:21:29.683 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:21:29.941 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:21:29.941 "name": "BaseBdev2", 00:21:29.941 "aliases": [ 00:21:29.941 "b11d616e-f3ec-46fc-918b-013a4ad929f7" 00:21:29.941 ], 00:21:29.941 "product_name": "Malloc disk", 00:21:29.941 "block_size": 512, 00:21:29.941 "num_blocks": 65536, 00:21:29.941 "uuid": "b11d616e-f3ec-46fc-918b-013a4ad929f7", 00:21:29.941 "assigned_rate_limits": { 00:21:29.941 "rw_ios_per_sec": 0, 00:21:29.941 "rw_mbytes_per_sec": 0, 00:21:29.941 "r_mbytes_per_sec": 0, 00:21:29.941 "w_mbytes_per_sec": 0 00:21:29.941 }, 00:21:29.941 "claimed": true, 00:21:29.941 "claim_type": "exclusive_write", 00:21:29.941 "zoned": false, 00:21:29.941 "supported_io_types": { 00:21:29.941 "read": true, 00:21:29.941 "write": true, 00:21:29.941 "unmap": true, 00:21:29.941 "flush": true, 00:21:29.941 "reset": true, 00:21:29.941 "nvme_admin": false, 00:21:29.941 "nvme_io": false, 00:21:29.941 "nvme_io_md": false, 00:21:29.941 "write_zeroes": true, 00:21:29.941 "zcopy": true, 00:21:29.941 "get_zone_info": false, 00:21:29.941 "zone_management": false, 00:21:29.941 "zone_append": false, 00:21:29.941 "compare": false, 00:21:29.941 "compare_and_write": false, 00:21:29.941 "abort": true, 00:21:29.941 "seek_hole": false, 00:21:29.941 "seek_data": false, 00:21:29.941 "copy": true, 00:21:29.941 "nvme_iov_md": false 00:21:29.941 }, 00:21:29.941 "memory_domains": [ 00:21:29.941 { 00:21:29.941 "dma_device_id": "system", 00:21:29.941 "dma_device_type": 1 00:21:29.941 }, 00:21:29.941 { 00:21:29.941 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:29.941 "dma_device_type": 2 00:21:29.941 } 00:21:29.941 ], 00:21:29.941 "driver_specific": {} 00:21:29.941 }' 00:21:29.941 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:29.942 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:30.201 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:21:30.201 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:30.201 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:30.201 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:21:30.201 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:30.201 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:30.201 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:21:30.201 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:30.201 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:30.507 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:21:30.507 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:21:30.507 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:21:30.507 11:32:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:21:30.507 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:21:30.507 "name": "BaseBdev3", 00:21:30.507 "aliases": [ 00:21:30.507 "9dcaf87b-749b-4087-bd63-7407864f706a" 00:21:30.507 ], 00:21:30.507 "product_name": "Malloc disk", 00:21:30.507 "block_size": 512, 00:21:30.507 "num_blocks": 65536, 00:21:30.507 "uuid": "9dcaf87b-749b-4087-bd63-7407864f706a", 00:21:30.507 "assigned_rate_limits": { 00:21:30.507 "rw_ios_per_sec": 0, 00:21:30.507 "rw_mbytes_per_sec": 0, 00:21:30.507 "r_mbytes_per_sec": 0, 00:21:30.507 "w_mbytes_per_sec": 0 00:21:30.507 }, 00:21:30.507 "claimed": true, 00:21:30.507 "claim_type": "exclusive_write", 00:21:30.507 "zoned": false, 00:21:30.507 "supported_io_types": { 00:21:30.507 "read": true, 00:21:30.507 "write": true, 00:21:30.507 "unmap": true, 00:21:30.507 "flush": true, 00:21:30.507 "reset": true, 00:21:30.507 "nvme_admin": false, 00:21:30.507 "nvme_io": false, 00:21:30.507 "nvme_io_md": false, 00:21:30.507 "write_zeroes": true, 00:21:30.507 "zcopy": true, 00:21:30.507 "get_zone_info": false, 00:21:30.507 "zone_management": false, 00:21:30.507 "zone_append": false, 00:21:30.507 "compare": false, 00:21:30.507 "compare_and_write": false, 00:21:30.507 "abort": true, 00:21:30.507 "seek_hole": false, 00:21:30.507 "seek_data": false, 00:21:30.507 "copy": true, 00:21:30.507 "nvme_iov_md": false 00:21:30.507 }, 00:21:30.507 "memory_domains": [ 00:21:30.507 { 00:21:30.507 "dma_device_id": "system", 00:21:30.507 "dma_device_type": 1 00:21:30.507 }, 00:21:30.507 { 00:21:30.507 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:30.507 "dma_device_type": 2 00:21:30.507 } 00:21:30.507 ], 00:21:30.507 "driver_specific": {} 00:21:30.507 }' 00:21:30.507 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:30.777 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:30.777 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:21:30.777 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:30.777 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:30.777 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:21:30.777 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:30.777 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:30.777 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:21:30.777 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:30.777 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:30.777 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:21:30.777 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:21:30.777 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:21:30.777 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 00:21:31.035 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:21:31.035 "name": "BaseBdev4", 00:21:31.035 "aliases": [ 00:21:31.035 "79be344d-3fcb-4007-a700-f22b518607d3" 00:21:31.035 ], 00:21:31.035 "product_name": "Malloc disk", 00:21:31.035 "block_size": 512, 00:21:31.035 "num_blocks": 65536, 00:21:31.035 "uuid": "79be344d-3fcb-4007-a700-f22b518607d3", 00:21:31.035 "assigned_rate_limits": { 00:21:31.035 "rw_ios_per_sec": 0, 00:21:31.035 "rw_mbytes_per_sec": 0, 00:21:31.035 "r_mbytes_per_sec": 0, 00:21:31.035 "w_mbytes_per_sec": 0 00:21:31.036 }, 00:21:31.036 "claimed": true, 00:21:31.036 "claim_type": "exclusive_write", 00:21:31.036 "zoned": false, 00:21:31.036 "supported_io_types": { 00:21:31.036 "read": true, 00:21:31.036 "write": true, 00:21:31.036 "unmap": true, 00:21:31.036 "flush": true, 00:21:31.036 "reset": true, 00:21:31.036 "nvme_admin": false, 00:21:31.036 "nvme_io": false, 00:21:31.036 "nvme_io_md": false, 00:21:31.036 "write_zeroes": true, 00:21:31.036 "zcopy": true, 00:21:31.036 "get_zone_info": false, 00:21:31.036 "zone_management": false, 00:21:31.036 "zone_append": false, 00:21:31.036 "compare": false, 00:21:31.036 "compare_and_write": false, 00:21:31.036 "abort": true, 00:21:31.036 "seek_hole": false, 00:21:31.036 "seek_data": false, 00:21:31.036 "copy": true, 00:21:31.036 "nvme_iov_md": false 00:21:31.036 }, 00:21:31.036 "memory_domains": [ 00:21:31.036 { 00:21:31.036 "dma_device_id": "system", 00:21:31.036 "dma_device_type": 1 00:21:31.036 }, 00:21:31.036 { 00:21:31.036 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:31.036 "dma_device_type": 2 00:21:31.036 } 00:21:31.036 ], 00:21:31.036 "driver_specific": {} 00:21:31.036 }' 00:21:31.036 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:31.295 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:31.295 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:21:31.295 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:31.295 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:31.295 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:21:31.295 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:31.295 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:31.553 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:21:31.553 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:31.553 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:31.553 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:21:31.553 11:32:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@338 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:21:31.812 [2024-07-15 11:32:15.186249] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:21:31.812 [2024-07-15 11:32:15.186279] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:21:31.812 [2024-07-15 11:32:15.186334] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:21:31.812 [2024-07-15 11:32:15.186621] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:21:31.812 [2024-07-15 11:32:15.186634] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1d7a610 name Existed_Raid, state offline 00:21:31.812 11:32:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@341 -- # killprocess 952238 00:21:31.812 11:32:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@948 -- # '[' -z 952238 ']' 00:21:31.812 11:32:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@952 -- # kill -0 952238 00:21:31.812 11:32:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@953 -- # uname 00:21:31.812 11:32:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:21:31.812 11:32:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 952238 00:21:31.812 11:32:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:21:31.812 11:32:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:21:31.813 11:32:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 952238' 00:21:31.813 killing process with pid 952238 00:21:31.813 11:32:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@967 -- # kill 952238 00:21:31.813 [2024-07-15 11:32:15.254240] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:21:31.813 11:32:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # wait 952238 00:21:31.813 [2024-07-15 11:32:15.296769] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@343 -- # return 0 00:21:32.072 00:21:32.072 real 0m31.682s 00:21:32.072 user 0m58.094s 00:21:32.072 sys 0m5.702s 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:32.072 ************************************ 00:21:32.072 END TEST raid_state_function_test 00:21:32.072 ************************************ 00:21:32.072 11:32:15 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:21:32.072 11:32:15 bdev_raid -- bdev/bdev_raid.sh@868 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 4 true 00:21:32.072 11:32:15 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:21:32.072 11:32:15 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:21:32.072 11:32:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:21:32.072 ************************************ 00:21:32.072 START TEST raid_state_function_test_sb 00:21:32.072 ************************************ 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1123 -- # raid_state_function_test raid1 4 true 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@220 -- # local raid_level=raid1 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=4 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # local superblock=true 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev3 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # echo BaseBdev4 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@226 -- # local strip_size 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # '[' raid1 '!=' raid1 ']' 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@234 -- # strip_size=0 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # '[' true = true ']' 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@238 -- # superblock_create_arg=-s 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # raid_pid=957029 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 957029' 00:21:32.072 Process raid pid: 957029 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@246 -- # waitforlisten 957029 /var/tmp/spdk-raid.sock 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@829 -- # '[' -z 957029 ']' 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@834 -- # local max_retries=100 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:21:32.072 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # xtrace_disable 00:21:32.072 11:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:32.330 [2024-07-15 11:32:15.674980] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:21:32.330 [2024-07-15 11:32:15.675042] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:32.330 [2024-07-15 11:32:15.804033] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:32.330 [2024-07-15 11:32:15.901239] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:21:32.588 [2024-07-15 11:32:15.961719] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:21:32.588 [2024-07-15 11:32:15.961776] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:21:33.154 11:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:21:33.154 11:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@862 -- # return 0 00:21:33.155 11:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:21:33.413 [2024-07-15 11:32:16.834468] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:21:33.413 [2024-07-15 11:32:16.834510] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:21:33.413 [2024-07-15 11:32:16.834521] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:21:33.413 [2024-07-15 11:32:16.834533] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:21:33.413 [2024-07-15 11:32:16.834542] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:21:33.413 [2024-07-15 11:32:16.834553] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:21:33.413 [2024-07-15 11:32:16.834562] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:21:33.413 [2024-07-15 11:32:16.834573] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:21:33.413 11:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:33.413 11:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:33.413 11:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:33.413 11:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:33.413 11:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:33.413 11:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:33.413 11:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:33.413 11:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:33.413 11:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:33.413 11:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:33.413 11:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:33.413 11:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:33.671 11:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:33.671 "name": "Existed_Raid", 00:21:33.672 "uuid": "8f177643-1a9a-41ae-bd83-e21ce270bfa2", 00:21:33.672 "strip_size_kb": 0, 00:21:33.672 "state": "configuring", 00:21:33.672 "raid_level": "raid1", 00:21:33.672 "superblock": true, 00:21:33.672 "num_base_bdevs": 4, 00:21:33.672 "num_base_bdevs_discovered": 0, 00:21:33.672 "num_base_bdevs_operational": 4, 00:21:33.672 "base_bdevs_list": [ 00:21:33.672 { 00:21:33.672 "name": "BaseBdev1", 00:21:33.672 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:33.672 "is_configured": false, 00:21:33.672 "data_offset": 0, 00:21:33.672 "data_size": 0 00:21:33.672 }, 00:21:33.672 { 00:21:33.672 "name": "BaseBdev2", 00:21:33.672 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:33.672 "is_configured": false, 00:21:33.672 "data_offset": 0, 00:21:33.672 "data_size": 0 00:21:33.672 }, 00:21:33.672 { 00:21:33.672 "name": "BaseBdev3", 00:21:33.672 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:33.672 "is_configured": false, 00:21:33.672 "data_offset": 0, 00:21:33.672 "data_size": 0 00:21:33.672 }, 00:21:33.672 { 00:21:33.672 "name": "BaseBdev4", 00:21:33.672 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:33.672 "is_configured": false, 00:21:33.672 "data_offset": 0, 00:21:33.672 "data_size": 0 00:21:33.672 } 00:21:33.672 ] 00:21:33.672 }' 00:21:33.672 11:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:33.672 11:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:34.237 11:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:21:34.495 [2024-07-15 11:32:17.933265] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:21:34.495 [2024-07-15 11:32:17.933296] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x2314aa0 name Existed_Raid, state configuring 00:21:34.495 11:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:21:34.752 [2024-07-15 11:32:18.177934] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:21:34.752 [2024-07-15 11:32:18.177959] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:21:34.752 [2024-07-15 11:32:18.177969] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:21:34.752 [2024-07-15 11:32:18.177980] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:21:34.752 [2024-07-15 11:32:18.177989] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:21:34.752 [2024-07-15 11:32:18.178000] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:21:34.752 [2024-07-15 11:32:18.178009] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:21:34.752 [2024-07-15 11:32:18.178020] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:21:34.752 11:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:21:35.008 [2024-07-15 11:32:18.432404] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:21:35.008 BaseBdev1 00:21:35.008 11:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:21:35.008 11:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:21:35.008 11:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:21:35.008 11:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:21:35.008 11:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:21:35.008 11:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:21:35.008 11:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:21:35.266 11:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:21:35.524 [ 00:21:35.524 { 00:21:35.524 "name": "BaseBdev1", 00:21:35.524 "aliases": [ 00:21:35.524 "226fe381-fe45-4f1c-94cf-ee575cbee87c" 00:21:35.524 ], 00:21:35.524 "product_name": "Malloc disk", 00:21:35.524 "block_size": 512, 00:21:35.524 "num_blocks": 65536, 00:21:35.524 "uuid": "226fe381-fe45-4f1c-94cf-ee575cbee87c", 00:21:35.524 "assigned_rate_limits": { 00:21:35.524 "rw_ios_per_sec": 0, 00:21:35.524 "rw_mbytes_per_sec": 0, 00:21:35.524 "r_mbytes_per_sec": 0, 00:21:35.524 "w_mbytes_per_sec": 0 00:21:35.524 }, 00:21:35.524 "claimed": true, 00:21:35.524 "claim_type": "exclusive_write", 00:21:35.524 "zoned": false, 00:21:35.524 "supported_io_types": { 00:21:35.524 "read": true, 00:21:35.524 "write": true, 00:21:35.524 "unmap": true, 00:21:35.524 "flush": true, 00:21:35.524 "reset": true, 00:21:35.524 "nvme_admin": false, 00:21:35.524 "nvme_io": false, 00:21:35.524 "nvme_io_md": false, 00:21:35.524 "write_zeroes": true, 00:21:35.524 "zcopy": true, 00:21:35.524 "get_zone_info": false, 00:21:35.524 "zone_management": false, 00:21:35.524 "zone_append": false, 00:21:35.524 "compare": false, 00:21:35.524 "compare_and_write": false, 00:21:35.524 "abort": true, 00:21:35.524 "seek_hole": false, 00:21:35.524 "seek_data": false, 00:21:35.524 "copy": true, 00:21:35.524 "nvme_iov_md": false 00:21:35.524 }, 00:21:35.524 "memory_domains": [ 00:21:35.524 { 00:21:35.524 "dma_device_id": "system", 00:21:35.524 "dma_device_type": 1 00:21:35.524 }, 00:21:35.524 { 00:21:35.524 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:35.524 "dma_device_type": 2 00:21:35.524 } 00:21:35.524 ], 00:21:35.524 "driver_specific": {} 00:21:35.524 } 00:21:35.524 ] 00:21:35.524 11:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:21:35.524 11:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:35.524 11:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:35.524 11:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:35.524 11:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:35.524 11:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:35.524 11:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:35.524 11:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:35.524 11:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:35.524 11:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:35.524 11:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:35.524 11:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:35.524 11:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:35.783 11:32:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:35.783 "name": "Existed_Raid", 00:21:35.783 "uuid": "b1f2b8fb-eb7f-479b-b68b-ba3ec1afc8ad", 00:21:35.783 "strip_size_kb": 0, 00:21:35.783 "state": "configuring", 00:21:35.783 "raid_level": "raid1", 00:21:35.783 "superblock": true, 00:21:35.783 "num_base_bdevs": 4, 00:21:35.783 "num_base_bdevs_discovered": 1, 00:21:35.783 "num_base_bdevs_operational": 4, 00:21:35.783 "base_bdevs_list": [ 00:21:35.783 { 00:21:35.783 "name": "BaseBdev1", 00:21:35.783 "uuid": "226fe381-fe45-4f1c-94cf-ee575cbee87c", 00:21:35.783 "is_configured": true, 00:21:35.783 "data_offset": 2048, 00:21:35.783 "data_size": 63488 00:21:35.783 }, 00:21:35.783 { 00:21:35.783 "name": "BaseBdev2", 00:21:35.783 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:35.783 "is_configured": false, 00:21:35.783 "data_offset": 0, 00:21:35.783 "data_size": 0 00:21:35.783 }, 00:21:35.783 { 00:21:35.783 "name": "BaseBdev3", 00:21:35.783 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:35.783 "is_configured": false, 00:21:35.783 "data_offset": 0, 00:21:35.783 "data_size": 0 00:21:35.783 }, 00:21:35.783 { 00:21:35.783 "name": "BaseBdev4", 00:21:35.783 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:35.783 "is_configured": false, 00:21:35.783 "data_offset": 0, 00:21:35.783 "data_size": 0 00:21:35.783 } 00:21:35.783 ] 00:21:35.783 }' 00:21:35.783 11:32:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:35.783 11:32:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:36.350 11:32:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:21:36.350 [2024-07-15 11:32:19.944407] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:21:36.350 [2024-07-15 11:32:19.944448] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x2314310 name Existed_Raid, state configuring 00:21:36.608 11:32:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:21:36.608 [2024-07-15 11:32:20.197118] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:21:36.608 [2024-07-15 11:32:20.198561] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:21:36.608 [2024-07-15 11:32:20.198594] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:21:36.608 [2024-07-15 11:32:20.198605] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:21:36.608 [2024-07-15 11:32:20.198617] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:21:36.608 [2024-07-15 11:32:20.198626] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:21:36.608 [2024-07-15 11:32:20.198638] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:21:36.867 11:32:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:21:36.867 11:32:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:21:36.867 11:32:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:36.867 11:32:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:36.867 11:32:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:36.867 11:32:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:36.867 11:32:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:36.867 11:32:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:36.867 11:32:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:36.867 11:32:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:36.867 11:32:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:36.867 11:32:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:36.867 11:32:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:36.867 11:32:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:37.125 11:32:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:37.125 "name": "Existed_Raid", 00:21:37.125 "uuid": "37e412cd-385f-4a0a-b6fd-a4e8fe55db55", 00:21:37.125 "strip_size_kb": 0, 00:21:37.125 "state": "configuring", 00:21:37.125 "raid_level": "raid1", 00:21:37.125 "superblock": true, 00:21:37.125 "num_base_bdevs": 4, 00:21:37.125 "num_base_bdevs_discovered": 1, 00:21:37.125 "num_base_bdevs_operational": 4, 00:21:37.125 "base_bdevs_list": [ 00:21:37.125 { 00:21:37.125 "name": "BaseBdev1", 00:21:37.125 "uuid": "226fe381-fe45-4f1c-94cf-ee575cbee87c", 00:21:37.125 "is_configured": true, 00:21:37.125 "data_offset": 2048, 00:21:37.125 "data_size": 63488 00:21:37.125 }, 00:21:37.125 { 00:21:37.125 "name": "BaseBdev2", 00:21:37.125 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:37.125 "is_configured": false, 00:21:37.125 "data_offset": 0, 00:21:37.125 "data_size": 0 00:21:37.125 }, 00:21:37.125 { 00:21:37.125 "name": "BaseBdev3", 00:21:37.125 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:37.125 "is_configured": false, 00:21:37.125 "data_offset": 0, 00:21:37.125 "data_size": 0 00:21:37.125 }, 00:21:37.125 { 00:21:37.125 "name": "BaseBdev4", 00:21:37.126 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:37.126 "is_configured": false, 00:21:37.126 "data_offset": 0, 00:21:37.126 "data_size": 0 00:21:37.126 } 00:21:37.126 ] 00:21:37.126 }' 00:21:37.126 11:32:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:37.126 11:32:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:37.692 11:32:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:21:37.692 [2024-07-15 11:32:21.283423] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:21:37.692 BaseBdev2 00:21:37.950 11:32:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:21:37.950 11:32:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:21:37.950 11:32:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:21:37.950 11:32:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:21:37.950 11:32:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:21:37.950 11:32:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:21:37.950 11:32:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:21:37.950 11:32:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:21:38.209 [ 00:21:38.209 { 00:21:38.209 "name": "BaseBdev2", 00:21:38.209 "aliases": [ 00:21:38.209 "5fb82784-ba36-4619-b919-6fba95b13fa9" 00:21:38.209 ], 00:21:38.209 "product_name": "Malloc disk", 00:21:38.209 "block_size": 512, 00:21:38.209 "num_blocks": 65536, 00:21:38.209 "uuid": "5fb82784-ba36-4619-b919-6fba95b13fa9", 00:21:38.209 "assigned_rate_limits": { 00:21:38.209 "rw_ios_per_sec": 0, 00:21:38.209 "rw_mbytes_per_sec": 0, 00:21:38.209 "r_mbytes_per_sec": 0, 00:21:38.209 "w_mbytes_per_sec": 0 00:21:38.209 }, 00:21:38.209 "claimed": true, 00:21:38.209 "claim_type": "exclusive_write", 00:21:38.209 "zoned": false, 00:21:38.209 "supported_io_types": { 00:21:38.209 "read": true, 00:21:38.209 "write": true, 00:21:38.209 "unmap": true, 00:21:38.209 "flush": true, 00:21:38.209 "reset": true, 00:21:38.209 "nvme_admin": false, 00:21:38.209 "nvme_io": false, 00:21:38.209 "nvme_io_md": false, 00:21:38.209 "write_zeroes": true, 00:21:38.209 "zcopy": true, 00:21:38.209 "get_zone_info": false, 00:21:38.209 "zone_management": false, 00:21:38.209 "zone_append": false, 00:21:38.209 "compare": false, 00:21:38.209 "compare_and_write": false, 00:21:38.209 "abort": true, 00:21:38.209 "seek_hole": false, 00:21:38.209 "seek_data": false, 00:21:38.209 "copy": true, 00:21:38.209 "nvme_iov_md": false 00:21:38.209 }, 00:21:38.209 "memory_domains": [ 00:21:38.209 { 00:21:38.209 "dma_device_id": "system", 00:21:38.209 "dma_device_type": 1 00:21:38.209 }, 00:21:38.209 { 00:21:38.209 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:38.209 "dma_device_type": 2 00:21:38.209 } 00:21:38.209 ], 00:21:38.209 "driver_specific": {} 00:21:38.209 } 00:21:38.209 ] 00:21:38.209 11:32:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:21:38.209 11:32:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:21:38.209 11:32:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:21:38.209 11:32:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:38.209 11:32:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:38.209 11:32:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:38.209 11:32:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:38.209 11:32:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:38.209 11:32:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:38.209 11:32:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:38.209 11:32:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:38.209 11:32:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:38.209 11:32:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:38.209 11:32:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:38.209 11:32:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:38.468 11:32:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:38.468 "name": "Existed_Raid", 00:21:38.468 "uuid": "37e412cd-385f-4a0a-b6fd-a4e8fe55db55", 00:21:38.468 "strip_size_kb": 0, 00:21:38.468 "state": "configuring", 00:21:38.468 "raid_level": "raid1", 00:21:38.468 "superblock": true, 00:21:38.468 "num_base_bdevs": 4, 00:21:38.468 "num_base_bdevs_discovered": 2, 00:21:38.468 "num_base_bdevs_operational": 4, 00:21:38.468 "base_bdevs_list": [ 00:21:38.468 { 00:21:38.468 "name": "BaseBdev1", 00:21:38.468 "uuid": "226fe381-fe45-4f1c-94cf-ee575cbee87c", 00:21:38.468 "is_configured": true, 00:21:38.468 "data_offset": 2048, 00:21:38.468 "data_size": 63488 00:21:38.468 }, 00:21:38.468 { 00:21:38.468 "name": "BaseBdev2", 00:21:38.468 "uuid": "5fb82784-ba36-4619-b919-6fba95b13fa9", 00:21:38.468 "is_configured": true, 00:21:38.468 "data_offset": 2048, 00:21:38.468 "data_size": 63488 00:21:38.468 }, 00:21:38.468 { 00:21:38.468 "name": "BaseBdev3", 00:21:38.468 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:38.468 "is_configured": false, 00:21:38.468 "data_offset": 0, 00:21:38.468 "data_size": 0 00:21:38.468 }, 00:21:38.468 { 00:21:38.468 "name": "BaseBdev4", 00:21:38.468 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:38.468 "is_configured": false, 00:21:38.468 "data_offset": 0, 00:21:38.468 "data_size": 0 00:21:38.468 } 00:21:38.468 ] 00:21:38.468 }' 00:21:38.468 11:32:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:38.468 11:32:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:39.034 11:32:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:21:39.292 [2024-07-15 11:32:22.642566] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:21:39.292 BaseBdev3 00:21:39.292 11:32:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev3 00:21:39.292 11:32:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:21:39.292 11:32:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:21:39.292 11:32:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:21:39.292 11:32:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:21:39.292 11:32:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:21:39.292 11:32:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:21:39.550 11:32:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:21:39.550 [ 00:21:39.550 { 00:21:39.550 "name": "BaseBdev3", 00:21:39.550 "aliases": [ 00:21:39.550 "a6eeadab-4622-4b9b-9335-84a32af1d176" 00:21:39.550 ], 00:21:39.550 "product_name": "Malloc disk", 00:21:39.550 "block_size": 512, 00:21:39.550 "num_blocks": 65536, 00:21:39.550 "uuid": "a6eeadab-4622-4b9b-9335-84a32af1d176", 00:21:39.550 "assigned_rate_limits": { 00:21:39.550 "rw_ios_per_sec": 0, 00:21:39.550 "rw_mbytes_per_sec": 0, 00:21:39.550 "r_mbytes_per_sec": 0, 00:21:39.550 "w_mbytes_per_sec": 0 00:21:39.550 }, 00:21:39.550 "claimed": true, 00:21:39.550 "claim_type": "exclusive_write", 00:21:39.550 "zoned": false, 00:21:39.550 "supported_io_types": { 00:21:39.550 "read": true, 00:21:39.550 "write": true, 00:21:39.550 "unmap": true, 00:21:39.550 "flush": true, 00:21:39.550 "reset": true, 00:21:39.550 "nvme_admin": false, 00:21:39.550 "nvme_io": false, 00:21:39.550 "nvme_io_md": false, 00:21:39.550 "write_zeroes": true, 00:21:39.550 "zcopy": true, 00:21:39.550 "get_zone_info": false, 00:21:39.550 "zone_management": false, 00:21:39.550 "zone_append": false, 00:21:39.550 "compare": false, 00:21:39.550 "compare_and_write": false, 00:21:39.550 "abort": true, 00:21:39.550 "seek_hole": false, 00:21:39.550 "seek_data": false, 00:21:39.550 "copy": true, 00:21:39.550 "nvme_iov_md": false 00:21:39.550 }, 00:21:39.550 "memory_domains": [ 00:21:39.550 { 00:21:39.550 "dma_device_id": "system", 00:21:39.551 "dma_device_type": 1 00:21:39.551 }, 00:21:39.551 { 00:21:39.551 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:39.551 "dma_device_type": 2 00:21:39.551 } 00:21:39.551 ], 00:21:39.551 "driver_specific": {} 00:21:39.551 } 00:21:39.551 ] 00:21:39.809 11:32:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:21:39.809 11:32:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:21:39.809 11:32:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:21:39.809 11:32:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:39.809 11:32:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:39.809 11:32:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:39.809 11:32:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:39.809 11:32:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:39.809 11:32:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:39.809 11:32:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:39.809 11:32:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:39.809 11:32:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:39.809 11:32:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:39.809 11:32:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:39.809 11:32:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:40.067 11:32:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:40.067 "name": "Existed_Raid", 00:21:40.067 "uuid": "37e412cd-385f-4a0a-b6fd-a4e8fe55db55", 00:21:40.067 "strip_size_kb": 0, 00:21:40.067 "state": "configuring", 00:21:40.067 "raid_level": "raid1", 00:21:40.067 "superblock": true, 00:21:40.067 "num_base_bdevs": 4, 00:21:40.067 "num_base_bdevs_discovered": 3, 00:21:40.067 "num_base_bdevs_operational": 4, 00:21:40.067 "base_bdevs_list": [ 00:21:40.067 { 00:21:40.067 "name": "BaseBdev1", 00:21:40.067 "uuid": "226fe381-fe45-4f1c-94cf-ee575cbee87c", 00:21:40.067 "is_configured": true, 00:21:40.067 "data_offset": 2048, 00:21:40.067 "data_size": 63488 00:21:40.067 }, 00:21:40.067 { 00:21:40.067 "name": "BaseBdev2", 00:21:40.067 "uuid": "5fb82784-ba36-4619-b919-6fba95b13fa9", 00:21:40.067 "is_configured": true, 00:21:40.067 "data_offset": 2048, 00:21:40.067 "data_size": 63488 00:21:40.067 }, 00:21:40.067 { 00:21:40.067 "name": "BaseBdev3", 00:21:40.067 "uuid": "a6eeadab-4622-4b9b-9335-84a32af1d176", 00:21:40.067 "is_configured": true, 00:21:40.067 "data_offset": 2048, 00:21:40.067 "data_size": 63488 00:21:40.067 }, 00:21:40.067 { 00:21:40.067 "name": "BaseBdev4", 00:21:40.067 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:40.067 "is_configured": false, 00:21:40.067 "data_offset": 0, 00:21:40.067 "data_size": 0 00:21:40.067 } 00:21:40.067 ] 00:21:40.067 }' 00:21:40.067 11:32:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:40.067 11:32:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:40.632 11:32:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4 00:21:40.632 [2024-07-15 11:32:24.194088] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:21:40.632 [2024-07-15 11:32:24.194266] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x2315350 00:21:40.632 [2024-07-15 11:32:24.194280] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:21:40.632 [2024-07-15 11:32:24.194454] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x2315020 00:21:40.632 [2024-07-15 11:32:24.194576] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x2315350 00:21:40.632 [2024-07-15 11:32:24.194586] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x2315350 00:21:40.632 [2024-07-15 11:32:24.194678] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:21:40.632 BaseBdev4 00:21:40.632 11:32:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev4 00:21:40.632 11:32:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev4 00:21:40.632 11:32:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:21:40.632 11:32:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:21:40.632 11:32:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:21:40.632 11:32:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:21:40.632 11:32:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:21:40.889 11:32:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 -t 2000 00:21:41.147 [ 00:21:41.147 { 00:21:41.147 "name": "BaseBdev4", 00:21:41.147 "aliases": [ 00:21:41.147 "e5fbc537-16d3-4794-92fb-ede5cbd55338" 00:21:41.147 ], 00:21:41.147 "product_name": "Malloc disk", 00:21:41.147 "block_size": 512, 00:21:41.147 "num_blocks": 65536, 00:21:41.147 "uuid": "e5fbc537-16d3-4794-92fb-ede5cbd55338", 00:21:41.147 "assigned_rate_limits": { 00:21:41.147 "rw_ios_per_sec": 0, 00:21:41.147 "rw_mbytes_per_sec": 0, 00:21:41.147 "r_mbytes_per_sec": 0, 00:21:41.147 "w_mbytes_per_sec": 0 00:21:41.147 }, 00:21:41.147 "claimed": true, 00:21:41.147 "claim_type": "exclusive_write", 00:21:41.147 "zoned": false, 00:21:41.147 "supported_io_types": { 00:21:41.147 "read": true, 00:21:41.147 "write": true, 00:21:41.147 "unmap": true, 00:21:41.147 "flush": true, 00:21:41.147 "reset": true, 00:21:41.147 "nvme_admin": false, 00:21:41.147 "nvme_io": false, 00:21:41.147 "nvme_io_md": false, 00:21:41.147 "write_zeroes": true, 00:21:41.147 "zcopy": true, 00:21:41.147 "get_zone_info": false, 00:21:41.147 "zone_management": false, 00:21:41.147 "zone_append": false, 00:21:41.147 "compare": false, 00:21:41.147 "compare_and_write": false, 00:21:41.147 "abort": true, 00:21:41.147 "seek_hole": false, 00:21:41.147 "seek_data": false, 00:21:41.147 "copy": true, 00:21:41.147 "nvme_iov_md": false 00:21:41.147 }, 00:21:41.147 "memory_domains": [ 00:21:41.147 { 00:21:41.147 "dma_device_id": "system", 00:21:41.147 "dma_device_type": 1 00:21:41.147 }, 00:21:41.147 { 00:21:41.147 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:41.147 "dma_device_type": 2 00:21:41.147 } 00:21:41.147 ], 00:21:41.147 "driver_specific": {} 00:21:41.147 } 00:21:41.147 ] 00:21:41.147 11:32:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:21:41.147 11:32:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:21:41.147 11:32:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:21:41.147 11:32:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:21:41.147 11:32:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:41.147 11:32:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:21:41.147 11:32:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:41.147 11:32:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:41.147 11:32:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:41.147 11:32:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:41.147 11:32:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:41.147 11:32:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:41.147 11:32:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:41.147 11:32:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:41.147 11:32:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:41.405 11:32:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:41.405 "name": "Existed_Raid", 00:21:41.405 "uuid": "37e412cd-385f-4a0a-b6fd-a4e8fe55db55", 00:21:41.405 "strip_size_kb": 0, 00:21:41.405 "state": "online", 00:21:41.405 "raid_level": "raid1", 00:21:41.405 "superblock": true, 00:21:41.405 "num_base_bdevs": 4, 00:21:41.405 "num_base_bdevs_discovered": 4, 00:21:41.405 "num_base_bdevs_operational": 4, 00:21:41.405 "base_bdevs_list": [ 00:21:41.405 { 00:21:41.405 "name": "BaseBdev1", 00:21:41.405 "uuid": "226fe381-fe45-4f1c-94cf-ee575cbee87c", 00:21:41.405 "is_configured": true, 00:21:41.405 "data_offset": 2048, 00:21:41.405 "data_size": 63488 00:21:41.405 }, 00:21:41.405 { 00:21:41.405 "name": "BaseBdev2", 00:21:41.405 "uuid": "5fb82784-ba36-4619-b919-6fba95b13fa9", 00:21:41.405 "is_configured": true, 00:21:41.405 "data_offset": 2048, 00:21:41.405 "data_size": 63488 00:21:41.405 }, 00:21:41.405 { 00:21:41.405 "name": "BaseBdev3", 00:21:41.405 "uuid": "a6eeadab-4622-4b9b-9335-84a32af1d176", 00:21:41.405 "is_configured": true, 00:21:41.405 "data_offset": 2048, 00:21:41.405 "data_size": 63488 00:21:41.405 }, 00:21:41.405 { 00:21:41.405 "name": "BaseBdev4", 00:21:41.405 "uuid": "e5fbc537-16d3-4794-92fb-ede5cbd55338", 00:21:41.405 "is_configured": true, 00:21:41.405 "data_offset": 2048, 00:21:41.405 "data_size": 63488 00:21:41.405 } 00:21:41.405 ] 00:21:41.405 }' 00:21:41.405 11:32:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:41.406 11:32:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:41.971 11:32:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:21:41.971 11:32:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:21:41.971 11:32:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:21:41.971 11:32:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:21:41.971 11:32:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:21:41.971 11:32:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # local name 00:21:41.971 11:32:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:21:41.971 11:32:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:21:42.229 [2024-07-15 11:32:25.750535] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:21:42.229 11:32:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:21:42.229 "name": "Existed_Raid", 00:21:42.229 "aliases": [ 00:21:42.229 "37e412cd-385f-4a0a-b6fd-a4e8fe55db55" 00:21:42.229 ], 00:21:42.229 "product_name": "Raid Volume", 00:21:42.229 "block_size": 512, 00:21:42.229 "num_blocks": 63488, 00:21:42.230 "uuid": "37e412cd-385f-4a0a-b6fd-a4e8fe55db55", 00:21:42.230 "assigned_rate_limits": { 00:21:42.230 "rw_ios_per_sec": 0, 00:21:42.230 "rw_mbytes_per_sec": 0, 00:21:42.230 "r_mbytes_per_sec": 0, 00:21:42.230 "w_mbytes_per_sec": 0 00:21:42.230 }, 00:21:42.230 "claimed": false, 00:21:42.230 "zoned": false, 00:21:42.230 "supported_io_types": { 00:21:42.230 "read": true, 00:21:42.230 "write": true, 00:21:42.230 "unmap": false, 00:21:42.230 "flush": false, 00:21:42.230 "reset": true, 00:21:42.230 "nvme_admin": false, 00:21:42.230 "nvme_io": false, 00:21:42.230 "nvme_io_md": false, 00:21:42.230 "write_zeroes": true, 00:21:42.230 "zcopy": false, 00:21:42.230 "get_zone_info": false, 00:21:42.230 "zone_management": false, 00:21:42.230 "zone_append": false, 00:21:42.230 "compare": false, 00:21:42.230 "compare_and_write": false, 00:21:42.230 "abort": false, 00:21:42.230 "seek_hole": false, 00:21:42.230 "seek_data": false, 00:21:42.230 "copy": false, 00:21:42.230 "nvme_iov_md": false 00:21:42.230 }, 00:21:42.230 "memory_domains": [ 00:21:42.230 { 00:21:42.230 "dma_device_id": "system", 00:21:42.230 "dma_device_type": 1 00:21:42.230 }, 00:21:42.230 { 00:21:42.230 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:42.230 "dma_device_type": 2 00:21:42.230 }, 00:21:42.230 { 00:21:42.230 "dma_device_id": "system", 00:21:42.230 "dma_device_type": 1 00:21:42.230 }, 00:21:42.230 { 00:21:42.230 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:42.230 "dma_device_type": 2 00:21:42.230 }, 00:21:42.230 { 00:21:42.230 "dma_device_id": "system", 00:21:42.230 "dma_device_type": 1 00:21:42.230 }, 00:21:42.230 { 00:21:42.230 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:42.230 "dma_device_type": 2 00:21:42.230 }, 00:21:42.230 { 00:21:42.230 "dma_device_id": "system", 00:21:42.230 "dma_device_type": 1 00:21:42.230 }, 00:21:42.230 { 00:21:42.230 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:42.230 "dma_device_type": 2 00:21:42.230 } 00:21:42.230 ], 00:21:42.230 "driver_specific": { 00:21:42.230 "raid": { 00:21:42.230 "uuid": "37e412cd-385f-4a0a-b6fd-a4e8fe55db55", 00:21:42.230 "strip_size_kb": 0, 00:21:42.230 "state": "online", 00:21:42.230 "raid_level": "raid1", 00:21:42.230 "superblock": true, 00:21:42.230 "num_base_bdevs": 4, 00:21:42.230 "num_base_bdevs_discovered": 4, 00:21:42.230 "num_base_bdevs_operational": 4, 00:21:42.230 "base_bdevs_list": [ 00:21:42.230 { 00:21:42.230 "name": "BaseBdev1", 00:21:42.230 "uuid": "226fe381-fe45-4f1c-94cf-ee575cbee87c", 00:21:42.230 "is_configured": true, 00:21:42.230 "data_offset": 2048, 00:21:42.230 "data_size": 63488 00:21:42.230 }, 00:21:42.230 { 00:21:42.230 "name": "BaseBdev2", 00:21:42.230 "uuid": "5fb82784-ba36-4619-b919-6fba95b13fa9", 00:21:42.230 "is_configured": true, 00:21:42.230 "data_offset": 2048, 00:21:42.230 "data_size": 63488 00:21:42.230 }, 00:21:42.230 { 00:21:42.230 "name": "BaseBdev3", 00:21:42.230 "uuid": "a6eeadab-4622-4b9b-9335-84a32af1d176", 00:21:42.230 "is_configured": true, 00:21:42.230 "data_offset": 2048, 00:21:42.230 "data_size": 63488 00:21:42.230 }, 00:21:42.230 { 00:21:42.230 "name": "BaseBdev4", 00:21:42.230 "uuid": "e5fbc537-16d3-4794-92fb-ede5cbd55338", 00:21:42.230 "is_configured": true, 00:21:42.230 "data_offset": 2048, 00:21:42.230 "data_size": 63488 00:21:42.230 } 00:21:42.230 ] 00:21:42.230 } 00:21:42.230 } 00:21:42.230 }' 00:21:42.230 11:32:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:21:42.230 11:32:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:21:42.230 BaseBdev2 00:21:42.230 BaseBdev3 00:21:42.230 BaseBdev4' 00:21:42.230 11:32:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:21:42.230 11:32:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:21:42.230 11:32:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:21:42.488 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:21:42.488 "name": "BaseBdev1", 00:21:42.488 "aliases": [ 00:21:42.488 "226fe381-fe45-4f1c-94cf-ee575cbee87c" 00:21:42.488 ], 00:21:42.488 "product_name": "Malloc disk", 00:21:42.488 "block_size": 512, 00:21:42.488 "num_blocks": 65536, 00:21:42.488 "uuid": "226fe381-fe45-4f1c-94cf-ee575cbee87c", 00:21:42.488 "assigned_rate_limits": { 00:21:42.488 "rw_ios_per_sec": 0, 00:21:42.488 "rw_mbytes_per_sec": 0, 00:21:42.488 "r_mbytes_per_sec": 0, 00:21:42.488 "w_mbytes_per_sec": 0 00:21:42.488 }, 00:21:42.488 "claimed": true, 00:21:42.488 "claim_type": "exclusive_write", 00:21:42.488 "zoned": false, 00:21:42.488 "supported_io_types": { 00:21:42.488 "read": true, 00:21:42.488 "write": true, 00:21:42.488 "unmap": true, 00:21:42.488 "flush": true, 00:21:42.488 "reset": true, 00:21:42.488 "nvme_admin": false, 00:21:42.488 "nvme_io": false, 00:21:42.488 "nvme_io_md": false, 00:21:42.488 "write_zeroes": true, 00:21:42.488 "zcopy": true, 00:21:42.488 "get_zone_info": false, 00:21:42.488 "zone_management": false, 00:21:42.488 "zone_append": false, 00:21:42.488 "compare": false, 00:21:42.488 "compare_and_write": false, 00:21:42.488 "abort": true, 00:21:42.488 "seek_hole": false, 00:21:42.488 "seek_data": false, 00:21:42.488 "copy": true, 00:21:42.488 "nvme_iov_md": false 00:21:42.488 }, 00:21:42.488 "memory_domains": [ 00:21:42.488 { 00:21:42.488 "dma_device_id": "system", 00:21:42.488 "dma_device_type": 1 00:21:42.488 }, 00:21:42.488 { 00:21:42.488 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:42.488 "dma_device_type": 2 00:21:42.488 } 00:21:42.488 ], 00:21:42.488 "driver_specific": {} 00:21:42.488 }' 00:21:42.488 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:42.745 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:42.745 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:21:42.745 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:42.745 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:42.745 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:21:42.745 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:42.745 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:42.745 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:21:42.745 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:43.003 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:43.003 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:21:43.003 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:21:43.003 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:21:43.003 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:21:43.260 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:21:43.260 "name": "BaseBdev2", 00:21:43.260 "aliases": [ 00:21:43.260 "5fb82784-ba36-4619-b919-6fba95b13fa9" 00:21:43.260 ], 00:21:43.260 "product_name": "Malloc disk", 00:21:43.260 "block_size": 512, 00:21:43.260 "num_blocks": 65536, 00:21:43.260 "uuid": "5fb82784-ba36-4619-b919-6fba95b13fa9", 00:21:43.260 "assigned_rate_limits": { 00:21:43.260 "rw_ios_per_sec": 0, 00:21:43.260 "rw_mbytes_per_sec": 0, 00:21:43.260 "r_mbytes_per_sec": 0, 00:21:43.260 "w_mbytes_per_sec": 0 00:21:43.260 }, 00:21:43.260 "claimed": true, 00:21:43.260 "claim_type": "exclusive_write", 00:21:43.260 "zoned": false, 00:21:43.260 "supported_io_types": { 00:21:43.260 "read": true, 00:21:43.260 "write": true, 00:21:43.260 "unmap": true, 00:21:43.260 "flush": true, 00:21:43.260 "reset": true, 00:21:43.260 "nvme_admin": false, 00:21:43.260 "nvme_io": false, 00:21:43.260 "nvme_io_md": false, 00:21:43.260 "write_zeroes": true, 00:21:43.260 "zcopy": true, 00:21:43.260 "get_zone_info": false, 00:21:43.261 "zone_management": false, 00:21:43.261 "zone_append": false, 00:21:43.261 "compare": false, 00:21:43.261 "compare_and_write": false, 00:21:43.261 "abort": true, 00:21:43.261 "seek_hole": false, 00:21:43.261 "seek_data": false, 00:21:43.261 "copy": true, 00:21:43.261 "nvme_iov_md": false 00:21:43.261 }, 00:21:43.261 "memory_domains": [ 00:21:43.261 { 00:21:43.261 "dma_device_id": "system", 00:21:43.261 "dma_device_type": 1 00:21:43.261 }, 00:21:43.261 { 00:21:43.261 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:43.261 "dma_device_type": 2 00:21:43.261 } 00:21:43.261 ], 00:21:43.261 "driver_specific": {} 00:21:43.261 }' 00:21:43.261 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:43.261 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:43.261 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:21:43.261 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:43.261 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:43.261 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:21:43.261 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:43.518 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:43.518 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:21:43.518 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:43.518 11:32:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:43.518 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:21:43.518 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:21:43.518 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:21:43.518 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:21:43.787 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:21:43.787 "name": "BaseBdev3", 00:21:43.787 "aliases": [ 00:21:43.787 "a6eeadab-4622-4b9b-9335-84a32af1d176" 00:21:43.787 ], 00:21:43.787 "product_name": "Malloc disk", 00:21:43.787 "block_size": 512, 00:21:43.787 "num_blocks": 65536, 00:21:43.787 "uuid": "a6eeadab-4622-4b9b-9335-84a32af1d176", 00:21:43.787 "assigned_rate_limits": { 00:21:43.787 "rw_ios_per_sec": 0, 00:21:43.787 "rw_mbytes_per_sec": 0, 00:21:43.787 "r_mbytes_per_sec": 0, 00:21:43.787 "w_mbytes_per_sec": 0 00:21:43.787 }, 00:21:43.787 "claimed": true, 00:21:43.787 "claim_type": "exclusive_write", 00:21:43.787 "zoned": false, 00:21:43.787 "supported_io_types": { 00:21:43.787 "read": true, 00:21:43.787 "write": true, 00:21:43.787 "unmap": true, 00:21:43.787 "flush": true, 00:21:43.787 "reset": true, 00:21:43.787 "nvme_admin": false, 00:21:43.787 "nvme_io": false, 00:21:43.787 "nvme_io_md": false, 00:21:43.787 "write_zeroes": true, 00:21:43.787 "zcopy": true, 00:21:43.787 "get_zone_info": false, 00:21:43.787 "zone_management": false, 00:21:43.787 "zone_append": false, 00:21:43.787 "compare": false, 00:21:43.787 "compare_and_write": false, 00:21:43.787 "abort": true, 00:21:43.787 "seek_hole": false, 00:21:43.787 "seek_data": false, 00:21:43.787 "copy": true, 00:21:43.787 "nvme_iov_md": false 00:21:43.787 }, 00:21:43.787 "memory_domains": [ 00:21:43.787 { 00:21:43.787 "dma_device_id": "system", 00:21:43.787 "dma_device_type": 1 00:21:43.787 }, 00:21:43.787 { 00:21:43.787 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:43.787 "dma_device_type": 2 00:21:43.787 } 00:21:43.787 ], 00:21:43.787 "driver_specific": {} 00:21:43.787 }' 00:21:43.787 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:43.787 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:43.787 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:21:43.787 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:44.052 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:44.052 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:21:44.052 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:44.052 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:44.052 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:21:44.052 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:44.052 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:44.052 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:21:44.052 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:21:44.053 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:21:44.053 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 00:21:44.309 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:21:44.309 "name": "BaseBdev4", 00:21:44.309 "aliases": [ 00:21:44.309 "e5fbc537-16d3-4794-92fb-ede5cbd55338" 00:21:44.309 ], 00:21:44.309 "product_name": "Malloc disk", 00:21:44.309 "block_size": 512, 00:21:44.309 "num_blocks": 65536, 00:21:44.309 "uuid": "e5fbc537-16d3-4794-92fb-ede5cbd55338", 00:21:44.309 "assigned_rate_limits": { 00:21:44.309 "rw_ios_per_sec": 0, 00:21:44.309 "rw_mbytes_per_sec": 0, 00:21:44.309 "r_mbytes_per_sec": 0, 00:21:44.309 "w_mbytes_per_sec": 0 00:21:44.309 }, 00:21:44.309 "claimed": true, 00:21:44.309 "claim_type": "exclusive_write", 00:21:44.309 "zoned": false, 00:21:44.309 "supported_io_types": { 00:21:44.309 "read": true, 00:21:44.309 "write": true, 00:21:44.309 "unmap": true, 00:21:44.309 "flush": true, 00:21:44.309 "reset": true, 00:21:44.309 "nvme_admin": false, 00:21:44.309 "nvme_io": false, 00:21:44.309 "nvme_io_md": false, 00:21:44.309 "write_zeroes": true, 00:21:44.309 "zcopy": true, 00:21:44.309 "get_zone_info": false, 00:21:44.309 "zone_management": false, 00:21:44.309 "zone_append": false, 00:21:44.309 "compare": false, 00:21:44.309 "compare_and_write": false, 00:21:44.309 "abort": true, 00:21:44.309 "seek_hole": false, 00:21:44.309 "seek_data": false, 00:21:44.309 "copy": true, 00:21:44.309 "nvme_iov_md": false 00:21:44.309 }, 00:21:44.309 "memory_domains": [ 00:21:44.309 { 00:21:44.309 "dma_device_id": "system", 00:21:44.309 "dma_device_type": 1 00:21:44.309 }, 00:21:44.309 { 00:21:44.309 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:44.309 "dma_device_type": 2 00:21:44.309 } 00:21:44.309 ], 00:21:44.309 "driver_specific": {} 00:21:44.309 }' 00:21:44.309 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:44.309 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:21:44.565 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:21:44.565 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:44.565 11:32:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:21:44.565 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:21:44.565 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:44.565 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:21:44.565 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:21:44.565 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:44.565 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:21:44.822 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:21:44.822 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:21:44.822 [2024-07-15 11:32:28.365218] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:21:44.822 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@275 -- # local expected_state 00:21:44.822 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # has_redundancy raid1 00:21:44.822 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # case $1 in 00:21:44.822 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@214 -- # return 0 00:21:44.822 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # expected_state=online 00:21:44.822 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:21:44.822 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:44.822 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:21:44.822 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:44.822 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:44.822 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:21:44.822 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:44.822 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:44.822 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:44.822 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:44.822 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:44.822 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:45.079 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:45.079 "name": "Existed_Raid", 00:21:45.079 "uuid": "37e412cd-385f-4a0a-b6fd-a4e8fe55db55", 00:21:45.079 "strip_size_kb": 0, 00:21:45.079 "state": "online", 00:21:45.079 "raid_level": "raid1", 00:21:45.079 "superblock": true, 00:21:45.079 "num_base_bdevs": 4, 00:21:45.079 "num_base_bdevs_discovered": 3, 00:21:45.079 "num_base_bdevs_operational": 3, 00:21:45.079 "base_bdevs_list": [ 00:21:45.079 { 00:21:45.079 "name": null, 00:21:45.079 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:45.079 "is_configured": false, 00:21:45.079 "data_offset": 2048, 00:21:45.079 "data_size": 63488 00:21:45.079 }, 00:21:45.079 { 00:21:45.079 "name": "BaseBdev2", 00:21:45.079 "uuid": "5fb82784-ba36-4619-b919-6fba95b13fa9", 00:21:45.079 "is_configured": true, 00:21:45.079 "data_offset": 2048, 00:21:45.079 "data_size": 63488 00:21:45.079 }, 00:21:45.079 { 00:21:45.079 "name": "BaseBdev3", 00:21:45.079 "uuid": "a6eeadab-4622-4b9b-9335-84a32af1d176", 00:21:45.079 "is_configured": true, 00:21:45.079 "data_offset": 2048, 00:21:45.079 "data_size": 63488 00:21:45.079 }, 00:21:45.079 { 00:21:45.079 "name": "BaseBdev4", 00:21:45.079 "uuid": "e5fbc537-16d3-4794-92fb-ede5cbd55338", 00:21:45.079 "is_configured": true, 00:21:45.079 "data_offset": 2048, 00:21:45.079 "data_size": 63488 00:21:45.079 } 00:21:45.079 ] 00:21:45.079 }' 00:21:45.079 11:32:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:45.079 11:32:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:45.642 11:32:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:21:45.642 11:32:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:21:45.642 11:32:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:21:45.642 11:32:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:45.899 11:32:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:21:45.899 11:32:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:21:45.899 11:32:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:21:46.154 [2024-07-15 11:32:29.697842] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:21:46.154 11:32:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:21:46.154 11:32:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:21:46.154 11:32:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:46.154 11:32:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:21:46.411 11:32:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:21:46.411 11:32:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:21:46.411 11:32:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev3 00:21:46.667 [2024-07-15 11:32:30.205875] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:21:46.667 11:32:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:21:46.667 11:32:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:21:46.667 11:32:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:46.667 11:32:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:21:46.924 11:32:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:21:46.924 11:32:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:21:46.924 11:32:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev4 00:21:47.182 [2024-07-15 11:32:30.693758] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:21:47.182 [2024-07-15 11:32:30.693839] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:21:47.182 [2024-07-15 11:32:30.706434] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:21:47.182 [2024-07-15 11:32:30.706471] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:21:47.182 [2024-07-15 11:32:30.706483] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x2315350 name Existed_Raid, state offline 00:21:47.182 11:32:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:21:47.182 11:32:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:21:47.182 11:32:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:47.182 11:32:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:21:47.440 11:32:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:21:47.440 11:32:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:21:47.440 11:32:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # '[' 4 -gt 2 ']' 00:21:47.440 11:32:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i = 1 )) 00:21:47.440 11:32:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:21:47.440 11:32:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2 00:21:47.699 BaseBdev2 00:21:47.699 11:32:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev2 00:21:47.699 11:32:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:21:47.699 11:32:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:21:47.699 11:32:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:21:47.699 11:32:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:21:47.699 11:32:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:21:47.699 11:32:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:21:48.055 11:32:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:21:48.313 [ 00:21:48.313 { 00:21:48.313 "name": "BaseBdev2", 00:21:48.313 "aliases": [ 00:21:48.313 "f1b1369f-f272-4167-aa26-c9e5953c23cb" 00:21:48.313 ], 00:21:48.313 "product_name": "Malloc disk", 00:21:48.313 "block_size": 512, 00:21:48.313 "num_blocks": 65536, 00:21:48.313 "uuid": "f1b1369f-f272-4167-aa26-c9e5953c23cb", 00:21:48.313 "assigned_rate_limits": { 00:21:48.313 "rw_ios_per_sec": 0, 00:21:48.313 "rw_mbytes_per_sec": 0, 00:21:48.313 "r_mbytes_per_sec": 0, 00:21:48.313 "w_mbytes_per_sec": 0 00:21:48.313 }, 00:21:48.313 "claimed": false, 00:21:48.313 "zoned": false, 00:21:48.313 "supported_io_types": { 00:21:48.313 "read": true, 00:21:48.313 "write": true, 00:21:48.313 "unmap": true, 00:21:48.313 "flush": true, 00:21:48.313 "reset": true, 00:21:48.313 "nvme_admin": false, 00:21:48.313 "nvme_io": false, 00:21:48.313 "nvme_io_md": false, 00:21:48.313 "write_zeroes": true, 00:21:48.313 "zcopy": true, 00:21:48.313 "get_zone_info": false, 00:21:48.313 "zone_management": false, 00:21:48.313 "zone_append": false, 00:21:48.313 "compare": false, 00:21:48.313 "compare_and_write": false, 00:21:48.313 "abort": true, 00:21:48.313 "seek_hole": false, 00:21:48.313 "seek_data": false, 00:21:48.313 "copy": true, 00:21:48.313 "nvme_iov_md": false 00:21:48.313 }, 00:21:48.313 "memory_domains": [ 00:21:48.313 { 00:21:48.313 "dma_device_id": "system", 00:21:48.313 "dma_device_type": 1 00:21:48.313 }, 00:21:48.313 { 00:21:48.313 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:48.313 "dma_device_type": 2 00:21:48.313 } 00:21:48.313 ], 00:21:48.313 "driver_specific": {} 00:21:48.313 } 00:21:48.313 ] 00:21:48.313 11:32:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:21:48.313 11:32:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:21:48.313 11:32:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:21:48.313 11:32:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3 00:21:48.571 BaseBdev3 00:21:48.571 11:32:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev3 00:21:48.571 11:32:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev3 00:21:48.571 11:32:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:21:48.571 11:32:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:21:48.571 11:32:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:21:48.571 11:32:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:21:48.571 11:32:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:21:48.828 11:32:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 -t 2000 00:21:48.828 [ 00:21:48.828 { 00:21:48.828 "name": "BaseBdev3", 00:21:48.828 "aliases": [ 00:21:48.828 "5bf03455-4402-4317-bb2f-d6fe7b2bb3bf" 00:21:48.828 ], 00:21:48.828 "product_name": "Malloc disk", 00:21:48.828 "block_size": 512, 00:21:48.828 "num_blocks": 65536, 00:21:48.828 "uuid": "5bf03455-4402-4317-bb2f-d6fe7b2bb3bf", 00:21:48.828 "assigned_rate_limits": { 00:21:48.828 "rw_ios_per_sec": 0, 00:21:48.828 "rw_mbytes_per_sec": 0, 00:21:48.828 "r_mbytes_per_sec": 0, 00:21:48.828 "w_mbytes_per_sec": 0 00:21:48.828 }, 00:21:48.828 "claimed": false, 00:21:48.828 "zoned": false, 00:21:48.828 "supported_io_types": { 00:21:48.828 "read": true, 00:21:48.828 "write": true, 00:21:48.828 "unmap": true, 00:21:48.828 "flush": true, 00:21:48.828 "reset": true, 00:21:48.828 "nvme_admin": false, 00:21:48.828 "nvme_io": false, 00:21:48.828 "nvme_io_md": false, 00:21:48.828 "write_zeroes": true, 00:21:48.828 "zcopy": true, 00:21:48.829 "get_zone_info": false, 00:21:48.829 "zone_management": false, 00:21:48.829 "zone_append": false, 00:21:48.829 "compare": false, 00:21:48.829 "compare_and_write": false, 00:21:48.829 "abort": true, 00:21:48.829 "seek_hole": false, 00:21:48.829 "seek_data": false, 00:21:48.829 "copy": true, 00:21:48.829 "nvme_iov_md": false 00:21:48.829 }, 00:21:48.829 "memory_domains": [ 00:21:48.829 { 00:21:48.829 "dma_device_id": "system", 00:21:48.829 "dma_device_type": 1 00:21:48.829 }, 00:21:48.829 { 00:21:48.829 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:48.829 "dma_device_type": 2 00:21:48.829 } 00:21:48.829 ], 00:21:48.829 "driver_specific": {} 00:21:48.829 } 00:21:48.829 ] 00:21:49.087 11:32:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:21:49.087 11:32:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:21:49.087 11:32:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:21:49.087 11:32:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4 00:21:49.087 BaseBdev4 00:21:49.087 11:32:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # waitforbdev BaseBdev4 00:21:49.087 11:32:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev4 00:21:49.087 11:32:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:21:49.345 11:32:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:21:49.345 11:32:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:21:49.345 11:32:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:21:49.345 11:32:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:21:49.345 11:32:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 -t 2000 00:21:49.603 [ 00:21:49.603 { 00:21:49.603 "name": "BaseBdev4", 00:21:49.603 "aliases": [ 00:21:49.603 "fbc0fced-b415-4286-a3b5-75d75bd77d33" 00:21:49.603 ], 00:21:49.603 "product_name": "Malloc disk", 00:21:49.603 "block_size": 512, 00:21:49.603 "num_blocks": 65536, 00:21:49.603 "uuid": "fbc0fced-b415-4286-a3b5-75d75bd77d33", 00:21:49.603 "assigned_rate_limits": { 00:21:49.603 "rw_ios_per_sec": 0, 00:21:49.603 "rw_mbytes_per_sec": 0, 00:21:49.603 "r_mbytes_per_sec": 0, 00:21:49.603 "w_mbytes_per_sec": 0 00:21:49.603 }, 00:21:49.603 "claimed": false, 00:21:49.603 "zoned": false, 00:21:49.603 "supported_io_types": { 00:21:49.603 "read": true, 00:21:49.603 "write": true, 00:21:49.603 "unmap": true, 00:21:49.603 "flush": true, 00:21:49.603 "reset": true, 00:21:49.603 "nvme_admin": false, 00:21:49.603 "nvme_io": false, 00:21:49.603 "nvme_io_md": false, 00:21:49.603 "write_zeroes": true, 00:21:49.603 "zcopy": true, 00:21:49.603 "get_zone_info": false, 00:21:49.603 "zone_management": false, 00:21:49.603 "zone_append": false, 00:21:49.603 "compare": false, 00:21:49.603 "compare_and_write": false, 00:21:49.603 "abort": true, 00:21:49.603 "seek_hole": false, 00:21:49.603 "seek_data": false, 00:21:49.603 "copy": true, 00:21:49.603 "nvme_iov_md": false 00:21:49.603 }, 00:21:49.603 "memory_domains": [ 00:21:49.603 { 00:21:49.603 "dma_device_id": "system", 00:21:49.603 "dma_device_type": 1 00:21:49.603 }, 00:21:49.603 { 00:21:49.603 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:49.603 "dma_device_type": 2 00:21:49.603 } 00:21:49.603 ], 00:21:49.603 "driver_specific": {} 00:21:49.603 } 00:21:49.603 ] 00:21:49.603 11:32:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:21:49.603 11:32:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i++ )) 00:21:49.603 11:32:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@301 -- # (( i < num_base_bdevs )) 00:21:49.603 11:32:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@305 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n Existed_Raid 00:21:49.861 [2024-07-15 11:32:33.391876] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:21:49.861 [2024-07-15 11:32:33.391915] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:21:49.861 [2024-07-15 11:32:33.391939] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:21:49.861 [2024-07-15 11:32:33.393301] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:21:49.861 [2024-07-15 11:32:33.393344] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:21:49.861 11:32:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:49.861 11:32:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:49.861 11:32:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:49.861 11:32:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:49.861 11:32:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:49.861 11:32:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:49.861 11:32:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:49.861 11:32:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:49.861 11:32:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:49.861 11:32:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:49.861 11:32:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:49.861 11:32:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:50.118 11:32:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:50.118 "name": "Existed_Raid", 00:21:50.118 "uuid": "49536c8f-bcdc-4c20-9501-5cb45bacfc7d", 00:21:50.118 "strip_size_kb": 0, 00:21:50.118 "state": "configuring", 00:21:50.118 "raid_level": "raid1", 00:21:50.118 "superblock": true, 00:21:50.118 "num_base_bdevs": 4, 00:21:50.118 "num_base_bdevs_discovered": 3, 00:21:50.118 "num_base_bdevs_operational": 4, 00:21:50.118 "base_bdevs_list": [ 00:21:50.118 { 00:21:50.118 "name": "BaseBdev1", 00:21:50.118 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:50.118 "is_configured": false, 00:21:50.118 "data_offset": 0, 00:21:50.118 "data_size": 0 00:21:50.118 }, 00:21:50.118 { 00:21:50.118 "name": "BaseBdev2", 00:21:50.118 "uuid": "f1b1369f-f272-4167-aa26-c9e5953c23cb", 00:21:50.118 "is_configured": true, 00:21:50.118 "data_offset": 2048, 00:21:50.118 "data_size": 63488 00:21:50.118 }, 00:21:50.118 { 00:21:50.118 "name": "BaseBdev3", 00:21:50.118 "uuid": "5bf03455-4402-4317-bb2f-d6fe7b2bb3bf", 00:21:50.118 "is_configured": true, 00:21:50.118 "data_offset": 2048, 00:21:50.118 "data_size": 63488 00:21:50.118 }, 00:21:50.118 { 00:21:50.118 "name": "BaseBdev4", 00:21:50.118 "uuid": "fbc0fced-b415-4286-a3b5-75d75bd77d33", 00:21:50.119 "is_configured": true, 00:21:50.119 "data_offset": 2048, 00:21:50.119 "data_size": 63488 00:21:50.119 } 00:21:50.119 ] 00:21:50.119 }' 00:21:50.119 11:32:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:50.119 11:32:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:50.683 11:32:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev2 00:21:50.941 [2024-07-15 11:32:34.378495] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:21:50.941 11:32:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@309 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:50.941 11:32:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:50.941 11:32:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:50.941 11:32:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:50.941 11:32:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:50.941 11:32:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:50.941 11:32:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:50.941 11:32:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:50.941 11:32:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:50.941 11:32:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:50.941 11:32:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:50.941 11:32:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:51.199 11:32:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:51.199 "name": "Existed_Raid", 00:21:51.199 "uuid": "49536c8f-bcdc-4c20-9501-5cb45bacfc7d", 00:21:51.199 "strip_size_kb": 0, 00:21:51.199 "state": "configuring", 00:21:51.199 "raid_level": "raid1", 00:21:51.199 "superblock": true, 00:21:51.199 "num_base_bdevs": 4, 00:21:51.199 "num_base_bdevs_discovered": 2, 00:21:51.199 "num_base_bdevs_operational": 4, 00:21:51.199 "base_bdevs_list": [ 00:21:51.199 { 00:21:51.199 "name": "BaseBdev1", 00:21:51.199 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:51.199 "is_configured": false, 00:21:51.199 "data_offset": 0, 00:21:51.199 "data_size": 0 00:21:51.199 }, 00:21:51.199 { 00:21:51.199 "name": null, 00:21:51.199 "uuid": "f1b1369f-f272-4167-aa26-c9e5953c23cb", 00:21:51.199 "is_configured": false, 00:21:51.199 "data_offset": 2048, 00:21:51.199 "data_size": 63488 00:21:51.199 }, 00:21:51.199 { 00:21:51.199 "name": "BaseBdev3", 00:21:51.199 "uuid": "5bf03455-4402-4317-bb2f-d6fe7b2bb3bf", 00:21:51.199 "is_configured": true, 00:21:51.199 "data_offset": 2048, 00:21:51.199 "data_size": 63488 00:21:51.199 }, 00:21:51.199 { 00:21:51.199 "name": "BaseBdev4", 00:21:51.199 "uuid": "fbc0fced-b415-4286-a3b5-75d75bd77d33", 00:21:51.199 "is_configured": true, 00:21:51.199 "data_offset": 2048, 00:21:51.199 "data_size": 63488 00:21:51.199 } 00:21:51.199 ] 00:21:51.199 }' 00:21:51.199 11:32:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:51.199 11:32:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:51.790 11:32:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:51.790 11:32:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:21:52.354 11:32:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # [[ false == \f\a\l\s\e ]] 00:21:52.354 11:32:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1 00:21:52.612 [2024-07-15 11:32:35.950018] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:21:52.612 BaseBdev1 00:21:52.612 11:32:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@313 -- # waitforbdev BaseBdev1 00:21:52.612 11:32:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:21:52.612 11:32:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:21:52.612 11:32:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:21:52.612 11:32:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:21:52.612 11:32:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:21:52.612 11:32:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:21:52.870 11:32:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:21:52.870 [ 00:21:52.870 { 00:21:52.870 "name": "BaseBdev1", 00:21:52.870 "aliases": [ 00:21:52.870 "d3e056f8-5454-4c48-8163-2af4875df6ae" 00:21:52.870 ], 00:21:52.870 "product_name": "Malloc disk", 00:21:52.870 "block_size": 512, 00:21:52.870 "num_blocks": 65536, 00:21:52.870 "uuid": "d3e056f8-5454-4c48-8163-2af4875df6ae", 00:21:52.870 "assigned_rate_limits": { 00:21:52.870 "rw_ios_per_sec": 0, 00:21:52.870 "rw_mbytes_per_sec": 0, 00:21:52.870 "r_mbytes_per_sec": 0, 00:21:52.870 "w_mbytes_per_sec": 0 00:21:52.870 }, 00:21:52.870 "claimed": true, 00:21:52.870 "claim_type": "exclusive_write", 00:21:52.870 "zoned": false, 00:21:52.870 "supported_io_types": { 00:21:52.870 "read": true, 00:21:52.870 "write": true, 00:21:52.870 "unmap": true, 00:21:52.870 "flush": true, 00:21:52.870 "reset": true, 00:21:52.870 "nvme_admin": false, 00:21:52.870 "nvme_io": false, 00:21:52.870 "nvme_io_md": false, 00:21:52.870 "write_zeroes": true, 00:21:52.870 "zcopy": true, 00:21:52.870 "get_zone_info": false, 00:21:52.870 "zone_management": false, 00:21:52.870 "zone_append": false, 00:21:52.870 "compare": false, 00:21:52.870 "compare_and_write": false, 00:21:52.870 "abort": true, 00:21:52.870 "seek_hole": false, 00:21:52.870 "seek_data": false, 00:21:52.870 "copy": true, 00:21:52.870 "nvme_iov_md": false 00:21:52.870 }, 00:21:52.870 "memory_domains": [ 00:21:52.870 { 00:21:52.870 "dma_device_id": "system", 00:21:52.870 "dma_device_type": 1 00:21:52.870 }, 00:21:52.870 { 00:21:52.870 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:52.870 "dma_device_type": 2 00:21:52.870 } 00:21:52.870 ], 00:21:52.870 "driver_specific": {} 00:21:52.870 } 00:21:52.870 ] 00:21:52.870 11:32:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:21:52.870 11:32:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:52.870 11:32:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:52.870 11:32:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:52.870 11:32:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:52.870 11:32:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:52.870 11:32:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:52.870 11:32:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:52.870 11:32:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:52.870 11:32:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:52.870 11:32:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:52.870 11:32:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:52.870 11:32:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:53.128 11:32:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:53.128 "name": "Existed_Raid", 00:21:53.128 "uuid": "49536c8f-bcdc-4c20-9501-5cb45bacfc7d", 00:21:53.128 "strip_size_kb": 0, 00:21:53.128 "state": "configuring", 00:21:53.128 "raid_level": "raid1", 00:21:53.128 "superblock": true, 00:21:53.128 "num_base_bdevs": 4, 00:21:53.128 "num_base_bdevs_discovered": 3, 00:21:53.128 "num_base_bdevs_operational": 4, 00:21:53.128 "base_bdevs_list": [ 00:21:53.128 { 00:21:53.128 "name": "BaseBdev1", 00:21:53.128 "uuid": "d3e056f8-5454-4c48-8163-2af4875df6ae", 00:21:53.128 "is_configured": true, 00:21:53.128 "data_offset": 2048, 00:21:53.128 "data_size": 63488 00:21:53.128 }, 00:21:53.128 { 00:21:53.128 "name": null, 00:21:53.128 "uuid": "f1b1369f-f272-4167-aa26-c9e5953c23cb", 00:21:53.128 "is_configured": false, 00:21:53.128 "data_offset": 2048, 00:21:53.128 "data_size": 63488 00:21:53.128 }, 00:21:53.128 { 00:21:53.128 "name": "BaseBdev3", 00:21:53.128 "uuid": "5bf03455-4402-4317-bb2f-d6fe7b2bb3bf", 00:21:53.128 "is_configured": true, 00:21:53.128 "data_offset": 2048, 00:21:53.128 "data_size": 63488 00:21:53.128 }, 00:21:53.128 { 00:21:53.128 "name": "BaseBdev4", 00:21:53.128 "uuid": "fbc0fced-b415-4286-a3b5-75d75bd77d33", 00:21:53.128 "is_configured": true, 00:21:53.128 "data_offset": 2048, 00:21:53.128 "data_size": 63488 00:21:53.128 } 00:21:53.128 ] 00:21:53.128 }' 00:21:53.128 11:32:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:53.128 11:32:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:53.693 11:32:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:53.693 11:32:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:21:53.951 11:32:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # [[ true == \t\r\u\e ]] 00:21:53.951 11:32:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@317 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev3 00:21:54.209 [2024-07-15 11:32:37.698671] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:21:54.209 11:32:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:54.209 11:32:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:54.209 11:32:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:54.209 11:32:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:54.209 11:32:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:54.209 11:32:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:54.209 11:32:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:54.209 11:32:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:54.209 11:32:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:54.209 11:32:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:54.209 11:32:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:54.209 11:32:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:54.466 11:32:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:54.466 "name": "Existed_Raid", 00:21:54.466 "uuid": "49536c8f-bcdc-4c20-9501-5cb45bacfc7d", 00:21:54.466 "strip_size_kb": 0, 00:21:54.466 "state": "configuring", 00:21:54.466 "raid_level": "raid1", 00:21:54.466 "superblock": true, 00:21:54.466 "num_base_bdevs": 4, 00:21:54.466 "num_base_bdevs_discovered": 2, 00:21:54.466 "num_base_bdevs_operational": 4, 00:21:54.466 "base_bdevs_list": [ 00:21:54.466 { 00:21:54.466 "name": "BaseBdev1", 00:21:54.466 "uuid": "d3e056f8-5454-4c48-8163-2af4875df6ae", 00:21:54.466 "is_configured": true, 00:21:54.466 "data_offset": 2048, 00:21:54.466 "data_size": 63488 00:21:54.466 }, 00:21:54.466 { 00:21:54.466 "name": null, 00:21:54.466 "uuid": "f1b1369f-f272-4167-aa26-c9e5953c23cb", 00:21:54.466 "is_configured": false, 00:21:54.466 "data_offset": 2048, 00:21:54.466 "data_size": 63488 00:21:54.466 }, 00:21:54.466 { 00:21:54.466 "name": null, 00:21:54.466 "uuid": "5bf03455-4402-4317-bb2f-d6fe7b2bb3bf", 00:21:54.466 "is_configured": false, 00:21:54.466 "data_offset": 2048, 00:21:54.466 "data_size": 63488 00:21:54.466 }, 00:21:54.466 { 00:21:54.466 "name": "BaseBdev4", 00:21:54.466 "uuid": "fbc0fced-b415-4286-a3b5-75d75bd77d33", 00:21:54.466 "is_configured": true, 00:21:54.466 "data_offset": 2048, 00:21:54.466 "data_size": 63488 00:21:54.466 } 00:21:54.466 ] 00:21:54.466 }' 00:21:54.466 11:32:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:54.466 11:32:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:55.031 11:32:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:21:55.031 11:32:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:55.289 11:32:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # [[ false == \f\a\l\s\e ]] 00:21:55.289 11:32:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:21:55.547 [2024-07-15 11:32:39.030225] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:21:55.547 11:32:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@322 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:55.547 11:32:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:55.547 11:32:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:55.547 11:32:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:55.547 11:32:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:55.547 11:32:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:55.547 11:32:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:55.547 11:32:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:55.547 11:32:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:55.547 11:32:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:55.547 11:32:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:55.547 11:32:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:55.805 11:32:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:55.805 "name": "Existed_Raid", 00:21:55.805 "uuid": "49536c8f-bcdc-4c20-9501-5cb45bacfc7d", 00:21:55.805 "strip_size_kb": 0, 00:21:55.805 "state": "configuring", 00:21:55.805 "raid_level": "raid1", 00:21:55.805 "superblock": true, 00:21:55.805 "num_base_bdevs": 4, 00:21:55.805 "num_base_bdevs_discovered": 3, 00:21:55.805 "num_base_bdevs_operational": 4, 00:21:55.805 "base_bdevs_list": [ 00:21:55.805 { 00:21:55.805 "name": "BaseBdev1", 00:21:55.805 "uuid": "d3e056f8-5454-4c48-8163-2af4875df6ae", 00:21:55.805 "is_configured": true, 00:21:55.805 "data_offset": 2048, 00:21:55.805 "data_size": 63488 00:21:55.805 }, 00:21:55.805 { 00:21:55.805 "name": null, 00:21:55.805 "uuid": "f1b1369f-f272-4167-aa26-c9e5953c23cb", 00:21:55.805 "is_configured": false, 00:21:55.805 "data_offset": 2048, 00:21:55.805 "data_size": 63488 00:21:55.805 }, 00:21:55.805 { 00:21:55.805 "name": "BaseBdev3", 00:21:55.805 "uuid": "5bf03455-4402-4317-bb2f-d6fe7b2bb3bf", 00:21:55.805 "is_configured": true, 00:21:55.805 "data_offset": 2048, 00:21:55.805 "data_size": 63488 00:21:55.805 }, 00:21:55.805 { 00:21:55.805 "name": "BaseBdev4", 00:21:55.805 "uuid": "fbc0fced-b415-4286-a3b5-75d75bd77d33", 00:21:55.805 "is_configured": true, 00:21:55.805 "data_offset": 2048, 00:21:55.805 "data_size": 63488 00:21:55.805 } 00:21:55.805 ] 00:21:55.805 }' 00:21:55.805 11:32:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:55.805 11:32:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:56.370 11:32:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:56.370 11:32:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:21:56.628 11:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # [[ true == \t\r\u\e ]] 00:21:56.628 11:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@325 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:21:56.901 [2024-07-15 11:32:40.361778] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:21:56.901 11:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:56.901 11:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:56.901 11:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:56.901 11:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:56.901 11:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:56.901 11:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:56.901 11:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:56.901 11:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:56.901 11:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:56.901 11:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:56.901 11:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:56.901 11:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:57.159 11:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:57.159 "name": "Existed_Raid", 00:21:57.159 "uuid": "49536c8f-bcdc-4c20-9501-5cb45bacfc7d", 00:21:57.159 "strip_size_kb": 0, 00:21:57.159 "state": "configuring", 00:21:57.159 "raid_level": "raid1", 00:21:57.159 "superblock": true, 00:21:57.159 "num_base_bdevs": 4, 00:21:57.159 "num_base_bdevs_discovered": 2, 00:21:57.159 "num_base_bdevs_operational": 4, 00:21:57.159 "base_bdevs_list": [ 00:21:57.159 { 00:21:57.159 "name": null, 00:21:57.159 "uuid": "d3e056f8-5454-4c48-8163-2af4875df6ae", 00:21:57.159 "is_configured": false, 00:21:57.159 "data_offset": 2048, 00:21:57.159 "data_size": 63488 00:21:57.159 }, 00:21:57.159 { 00:21:57.159 "name": null, 00:21:57.159 "uuid": "f1b1369f-f272-4167-aa26-c9e5953c23cb", 00:21:57.159 "is_configured": false, 00:21:57.159 "data_offset": 2048, 00:21:57.159 "data_size": 63488 00:21:57.159 }, 00:21:57.159 { 00:21:57.159 "name": "BaseBdev3", 00:21:57.159 "uuid": "5bf03455-4402-4317-bb2f-d6fe7b2bb3bf", 00:21:57.159 "is_configured": true, 00:21:57.159 "data_offset": 2048, 00:21:57.159 "data_size": 63488 00:21:57.159 }, 00:21:57.159 { 00:21:57.159 "name": "BaseBdev4", 00:21:57.159 "uuid": "fbc0fced-b415-4286-a3b5-75d75bd77d33", 00:21:57.159 "is_configured": true, 00:21:57.159 "data_offset": 2048, 00:21:57.159 "data_size": 63488 00:21:57.159 } 00:21:57.159 ] 00:21:57.159 }' 00:21:57.159 11:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:57.159 11:32:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:57.724 11:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@327 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:57.724 11:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@327 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:21:57.982 11:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@327 -- # [[ false == \f\a\l\s\e ]] 00:21:57.982 11:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@329 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:21:58.239 [2024-07-15 11:32:41.727969] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:21:58.239 11:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@330 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:21:58.239 11:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:21:58.239 11:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:21:58.239 11:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:21:58.239 11:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:21:58.239 11:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:21:58.239 11:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:21:58.239 11:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:21:58.239 11:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:21:58.239 11:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:21:58.239 11:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:58.239 11:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:58.497 11:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:21:58.497 "name": "Existed_Raid", 00:21:58.497 "uuid": "49536c8f-bcdc-4c20-9501-5cb45bacfc7d", 00:21:58.497 "strip_size_kb": 0, 00:21:58.497 "state": "configuring", 00:21:58.497 "raid_level": "raid1", 00:21:58.497 "superblock": true, 00:21:58.497 "num_base_bdevs": 4, 00:21:58.497 "num_base_bdevs_discovered": 3, 00:21:58.497 "num_base_bdevs_operational": 4, 00:21:58.497 "base_bdevs_list": [ 00:21:58.497 { 00:21:58.497 "name": null, 00:21:58.497 "uuid": "d3e056f8-5454-4c48-8163-2af4875df6ae", 00:21:58.497 "is_configured": false, 00:21:58.497 "data_offset": 2048, 00:21:58.497 "data_size": 63488 00:21:58.497 }, 00:21:58.497 { 00:21:58.497 "name": "BaseBdev2", 00:21:58.497 "uuid": "f1b1369f-f272-4167-aa26-c9e5953c23cb", 00:21:58.497 "is_configured": true, 00:21:58.497 "data_offset": 2048, 00:21:58.497 "data_size": 63488 00:21:58.497 }, 00:21:58.497 { 00:21:58.497 "name": "BaseBdev3", 00:21:58.497 "uuid": "5bf03455-4402-4317-bb2f-d6fe7b2bb3bf", 00:21:58.497 "is_configured": true, 00:21:58.497 "data_offset": 2048, 00:21:58.497 "data_size": 63488 00:21:58.497 }, 00:21:58.497 { 00:21:58.497 "name": "BaseBdev4", 00:21:58.497 "uuid": "fbc0fced-b415-4286-a3b5-75d75bd77d33", 00:21:58.497 "is_configured": true, 00:21:58.497 "data_offset": 2048, 00:21:58.497 "data_size": 63488 00:21:58.497 } 00:21:58.497 ] 00:21:58.497 }' 00:21:58.497 11:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:21:58.497 11:32:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:59.063 11:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@331 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:21:59.063 11:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@331 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:59.321 11:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@331 -- # [[ true == \t\r\u\e ]] 00:21:59.321 11:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:21:59.321 11:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@333 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:21:59.579 11:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@333 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b NewBaseBdev -u d3e056f8-5454-4c48-8163-2af4875df6ae 00:21:59.835 [2024-07-15 11:32:43.252526] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:21:59.835 [2024-07-15 11:32:43.252687] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x2317180 00:21:59.835 [2024-07-15 11:32:43.252700] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:21:59.835 [2024-07-15 11:32:43.252874] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x2317c20 00:21:59.835 [2024-07-15 11:32:43.253019] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x2317180 00:21:59.835 [2024-07-15 11:32:43.253030] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x2317180 00:21:59.835 [2024-07-15 11:32:43.253127] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:21:59.835 NewBaseBdev 00:21:59.835 11:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@334 -- # waitforbdev NewBaseBdev 00:21:59.835 11:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@897 -- # local bdev_name=NewBaseBdev 00:21:59.835 11:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:21:59.835 11:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local i 00:21:59.835 11:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:21:59.835 11:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:21:59.835 11:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:22:00.091 11:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev -t 2000 00:22:00.348 [ 00:22:00.348 { 00:22:00.348 "name": "NewBaseBdev", 00:22:00.348 "aliases": [ 00:22:00.348 "d3e056f8-5454-4c48-8163-2af4875df6ae" 00:22:00.348 ], 00:22:00.348 "product_name": "Malloc disk", 00:22:00.348 "block_size": 512, 00:22:00.348 "num_blocks": 65536, 00:22:00.348 "uuid": "d3e056f8-5454-4c48-8163-2af4875df6ae", 00:22:00.348 "assigned_rate_limits": { 00:22:00.348 "rw_ios_per_sec": 0, 00:22:00.348 "rw_mbytes_per_sec": 0, 00:22:00.348 "r_mbytes_per_sec": 0, 00:22:00.348 "w_mbytes_per_sec": 0 00:22:00.348 }, 00:22:00.348 "claimed": true, 00:22:00.348 "claim_type": "exclusive_write", 00:22:00.348 "zoned": false, 00:22:00.348 "supported_io_types": { 00:22:00.348 "read": true, 00:22:00.348 "write": true, 00:22:00.348 "unmap": true, 00:22:00.348 "flush": true, 00:22:00.348 "reset": true, 00:22:00.348 "nvme_admin": false, 00:22:00.348 "nvme_io": false, 00:22:00.348 "nvme_io_md": false, 00:22:00.348 "write_zeroes": true, 00:22:00.348 "zcopy": true, 00:22:00.348 "get_zone_info": false, 00:22:00.348 "zone_management": false, 00:22:00.348 "zone_append": false, 00:22:00.348 "compare": false, 00:22:00.348 "compare_and_write": false, 00:22:00.348 "abort": true, 00:22:00.348 "seek_hole": false, 00:22:00.348 "seek_data": false, 00:22:00.348 "copy": true, 00:22:00.348 "nvme_iov_md": false 00:22:00.348 }, 00:22:00.348 "memory_domains": [ 00:22:00.348 { 00:22:00.348 "dma_device_id": "system", 00:22:00.348 "dma_device_type": 1 00:22:00.348 }, 00:22:00.348 { 00:22:00.348 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:00.348 "dma_device_type": 2 00:22:00.348 } 00:22:00.348 ], 00:22:00.348 "driver_specific": {} 00:22:00.348 } 00:22:00.348 ] 00:22:00.348 11:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # return 0 00:22:00.348 11:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@335 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:22:00.348 11:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:22:00.348 11:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:22:00.348 11:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:22:00.348 11:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:22:00.348 11:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:22:00.348 11:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:22:00.348 11:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:22:00.348 11:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:22:00.348 11:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:22:00.348 11:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:00.348 11:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:00.606 11:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:22:00.606 "name": "Existed_Raid", 00:22:00.606 "uuid": "49536c8f-bcdc-4c20-9501-5cb45bacfc7d", 00:22:00.606 "strip_size_kb": 0, 00:22:00.606 "state": "online", 00:22:00.606 "raid_level": "raid1", 00:22:00.606 "superblock": true, 00:22:00.606 "num_base_bdevs": 4, 00:22:00.606 "num_base_bdevs_discovered": 4, 00:22:00.606 "num_base_bdevs_operational": 4, 00:22:00.606 "base_bdevs_list": [ 00:22:00.606 { 00:22:00.606 "name": "NewBaseBdev", 00:22:00.606 "uuid": "d3e056f8-5454-4c48-8163-2af4875df6ae", 00:22:00.606 "is_configured": true, 00:22:00.606 "data_offset": 2048, 00:22:00.606 "data_size": 63488 00:22:00.606 }, 00:22:00.606 { 00:22:00.606 "name": "BaseBdev2", 00:22:00.606 "uuid": "f1b1369f-f272-4167-aa26-c9e5953c23cb", 00:22:00.606 "is_configured": true, 00:22:00.606 "data_offset": 2048, 00:22:00.606 "data_size": 63488 00:22:00.606 }, 00:22:00.606 { 00:22:00.606 "name": "BaseBdev3", 00:22:00.606 "uuid": "5bf03455-4402-4317-bb2f-d6fe7b2bb3bf", 00:22:00.606 "is_configured": true, 00:22:00.606 "data_offset": 2048, 00:22:00.606 "data_size": 63488 00:22:00.606 }, 00:22:00.606 { 00:22:00.606 "name": "BaseBdev4", 00:22:00.606 "uuid": "fbc0fced-b415-4286-a3b5-75d75bd77d33", 00:22:00.606 "is_configured": true, 00:22:00.606 "data_offset": 2048, 00:22:00.606 "data_size": 63488 00:22:00.606 } 00:22:00.606 ] 00:22:00.606 }' 00:22:00.606 11:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:22:00.606 11:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:01.171 11:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@336 -- # verify_raid_bdev_properties Existed_Raid 00:22:01.171 11:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:22:01.171 11:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:22:01.171 11:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:22:01.171 11:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:22:01.171 11:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # local name 00:22:01.171 11:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:22:01.171 11:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:22:01.429 [2024-07-15 11:32:44.817006] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:22:01.429 11:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:22:01.429 "name": "Existed_Raid", 00:22:01.429 "aliases": [ 00:22:01.429 "49536c8f-bcdc-4c20-9501-5cb45bacfc7d" 00:22:01.429 ], 00:22:01.429 "product_name": "Raid Volume", 00:22:01.429 "block_size": 512, 00:22:01.429 "num_blocks": 63488, 00:22:01.429 "uuid": "49536c8f-bcdc-4c20-9501-5cb45bacfc7d", 00:22:01.429 "assigned_rate_limits": { 00:22:01.429 "rw_ios_per_sec": 0, 00:22:01.429 "rw_mbytes_per_sec": 0, 00:22:01.429 "r_mbytes_per_sec": 0, 00:22:01.429 "w_mbytes_per_sec": 0 00:22:01.429 }, 00:22:01.429 "claimed": false, 00:22:01.429 "zoned": false, 00:22:01.429 "supported_io_types": { 00:22:01.429 "read": true, 00:22:01.429 "write": true, 00:22:01.429 "unmap": false, 00:22:01.429 "flush": false, 00:22:01.429 "reset": true, 00:22:01.429 "nvme_admin": false, 00:22:01.429 "nvme_io": false, 00:22:01.429 "nvme_io_md": false, 00:22:01.429 "write_zeroes": true, 00:22:01.429 "zcopy": false, 00:22:01.429 "get_zone_info": false, 00:22:01.429 "zone_management": false, 00:22:01.429 "zone_append": false, 00:22:01.429 "compare": false, 00:22:01.429 "compare_and_write": false, 00:22:01.429 "abort": false, 00:22:01.429 "seek_hole": false, 00:22:01.429 "seek_data": false, 00:22:01.429 "copy": false, 00:22:01.429 "nvme_iov_md": false 00:22:01.429 }, 00:22:01.429 "memory_domains": [ 00:22:01.429 { 00:22:01.429 "dma_device_id": "system", 00:22:01.429 "dma_device_type": 1 00:22:01.429 }, 00:22:01.429 { 00:22:01.429 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:01.429 "dma_device_type": 2 00:22:01.429 }, 00:22:01.429 { 00:22:01.429 "dma_device_id": "system", 00:22:01.429 "dma_device_type": 1 00:22:01.429 }, 00:22:01.429 { 00:22:01.429 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:01.429 "dma_device_type": 2 00:22:01.429 }, 00:22:01.429 { 00:22:01.429 "dma_device_id": "system", 00:22:01.429 "dma_device_type": 1 00:22:01.429 }, 00:22:01.429 { 00:22:01.429 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:01.429 "dma_device_type": 2 00:22:01.429 }, 00:22:01.429 { 00:22:01.429 "dma_device_id": "system", 00:22:01.429 "dma_device_type": 1 00:22:01.429 }, 00:22:01.429 { 00:22:01.429 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:01.429 "dma_device_type": 2 00:22:01.429 } 00:22:01.429 ], 00:22:01.429 "driver_specific": { 00:22:01.429 "raid": { 00:22:01.429 "uuid": "49536c8f-bcdc-4c20-9501-5cb45bacfc7d", 00:22:01.429 "strip_size_kb": 0, 00:22:01.429 "state": "online", 00:22:01.429 "raid_level": "raid1", 00:22:01.429 "superblock": true, 00:22:01.429 "num_base_bdevs": 4, 00:22:01.429 "num_base_bdevs_discovered": 4, 00:22:01.429 "num_base_bdevs_operational": 4, 00:22:01.429 "base_bdevs_list": [ 00:22:01.429 { 00:22:01.429 "name": "NewBaseBdev", 00:22:01.429 "uuid": "d3e056f8-5454-4c48-8163-2af4875df6ae", 00:22:01.429 "is_configured": true, 00:22:01.429 "data_offset": 2048, 00:22:01.429 "data_size": 63488 00:22:01.429 }, 00:22:01.429 { 00:22:01.429 "name": "BaseBdev2", 00:22:01.429 "uuid": "f1b1369f-f272-4167-aa26-c9e5953c23cb", 00:22:01.429 "is_configured": true, 00:22:01.429 "data_offset": 2048, 00:22:01.429 "data_size": 63488 00:22:01.429 }, 00:22:01.429 { 00:22:01.429 "name": "BaseBdev3", 00:22:01.429 "uuid": "5bf03455-4402-4317-bb2f-d6fe7b2bb3bf", 00:22:01.429 "is_configured": true, 00:22:01.429 "data_offset": 2048, 00:22:01.429 "data_size": 63488 00:22:01.429 }, 00:22:01.429 { 00:22:01.429 "name": "BaseBdev4", 00:22:01.429 "uuid": "fbc0fced-b415-4286-a3b5-75d75bd77d33", 00:22:01.429 "is_configured": true, 00:22:01.429 "data_offset": 2048, 00:22:01.429 "data_size": 63488 00:22:01.429 } 00:22:01.429 ] 00:22:01.429 } 00:22:01.429 } 00:22:01.429 }' 00:22:01.429 11:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:22:01.429 11:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@201 -- # base_bdev_names='NewBaseBdev 00:22:01.429 BaseBdev2 00:22:01.429 BaseBdev3 00:22:01.429 BaseBdev4' 00:22:01.429 11:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:22:01.429 11:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b NewBaseBdev 00:22:01.429 11:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:22:01.687 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:22:01.687 "name": "NewBaseBdev", 00:22:01.687 "aliases": [ 00:22:01.687 "d3e056f8-5454-4c48-8163-2af4875df6ae" 00:22:01.687 ], 00:22:01.687 "product_name": "Malloc disk", 00:22:01.687 "block_size": 512, 00:22:01.687 "num_blocks": 65536, 00:22:01.687 "uuid": "d3e056f8-5454-4c48-8163-2af4875df6ae", 00:22:01.687 "assigned_rate_limits": { 00:22:01.687 "rw_ios_per_sec": 0, 00:22:01.687 "rw_mbytes_per_sec": 0, 00:22:01.687 "r_mbytes_per_sec": 0, 00:22:01.687 "w_mbytes_per_sec": 0 00:22:01.687 }, 00:22:01.687 "claimed": true, 00:22:01.687 "claim_type": "exclusive_write", 00:22:01.687 "zoned": false, 00:22:01.687 "supported_io_types": { 00:22:01.687 "read": true, 00:22:01.687 "write": true, 00:22:01.687 "unmap": true, 00:22:01.687 "flush": true, 00:22:01.687 "reset": true, 00:22:01.687 "nvme_admin": false, 00:22:01.687 "nvme_io": false, 00:22:01.687 "nvme_io_md": false, 00:22:01.687 "write_zeroes": true, 00:22:01.687 "zcopy": true, 00:22:01.687 "get_zone_info": false, 00:22:01.687 "zone_management": false, 00:22:01.687 "zone_append": false, 00:22:01.687 "compare": false, 00:22:01.687 "compare_and_write": false, 00:22:01.687 "abort": true, 00:22:01.687 "seek_hole": false, 00:22:01.687 "seek_data": false, 00:22:01.687 "copy": true, 00:22:01.687 "nvme_iov_md": false 00:22:01.687 }, 00:22:01.687 "memory_domains": [ 00:22:01.687 { 00:22:01.687 "dma_device_id": "system", 00:22:01.687 "dma_device_type": 1 00:22:01.687 }, 00:22:01.687 { 00:22:01.687 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:01.687 "dma_device_type": 2 00:22:01.687 } 00:22:01.687 ], 00:22:01.687 "driver_specific": {} 00:22:01.687 }' 00:22:01.687 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:01.687 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:01.687 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:22:01.687 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:01.687 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:01.945 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:22:01.945 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:01.945 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:01.945 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:22:01.945 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:01.945 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:01.945 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:22:01.945 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:22:01.945 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:22:01.945 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:22:02.203 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:22:02.203 "name": "BaseBdev2", 00:22:02.203 "aliases": [ 00:22:02.203 "f1b1369f-f272-4167-aa26-c9e5953c23cb" 00:22:02.203 ], 00:22:02.203 "product_name": "Malloc disk", 00:22:02.203 "block_size": 512, 00:22:02.203 "num_blocks": 65536, 00:22:02.203 "uuid": "f1b1369f-f272-4167-aa26-c9e5953c23cb", 00:22:02.203 "assigned_rate_limits": { 00:22:02.203 "rw_ios_per_sec": 0, 00:22:02.203 "rw_mbytes_per_sec": 0, 00:22:02.203 "r_mbytes_per_sec": 0, 00:22:02.203 "w_mbytes_per_sec": 0 00:22:02.203 }, 00:22:02.203 "claimed": true, 00:22:02.203 "claim_type": "exclusive_write", 00:22:02.203 "zoned": false, 00:22:02.203 "supported_io_types": { 00:22:02.203 "read": true, 00:22:02.203 "write": true, 00:22:02.203 "unmap": true, 00:22:02.203 "flush": true, 00:22:02.203 "reset": true, 00:22:02.203 "nvme_admin": false, 00:22:02.203 "nvme_io": false, 00:22:02.203 "nvme_io_md": false, 00:22:02.203 "write_zeroes": true, 00:22:02.203 "zcopy": true, 00:22:02.203 "get_zone_info": false, 00:22:02.203 "zone_management": false, 00:22:02.203 "zone_append": false, 00:22:02.203 "compare": false, 00:22:02.203 "compare_and_write": false, 00:22:02.203 "abort": true, 00:22:02.203 "seek_hole": false, 00:22:02.203 "seek_data": false, 00:22:02.203 "copy": true, 00:22:02.203 "nvme_iov_md": false 00:22:02.203 }, 00:22:02.203 "memory_domains": [ 00:22:02.203 { 00:22:02.203 "dma_device_id": "system", 00:22:02.203 "dma_device_type": 1 00:22:02.203 }, 00:22:02.203 { 00:22:02.203 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:02.203 "dma_device_type": 2 00:22:02.203 } 00:22:02.203 ], 00:22:02.203 "driver_specific": {} 00:22:02.203 }' 00:22:02.203 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:02.203 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:02.460 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:22:02.460 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:02.460 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:02.460 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:22:02.460 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:02.460 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:02.460 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:22:02.460 11:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:02.460 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:02.718 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:22:02.718 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:22:02.718 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev3 00:22:02.718 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:22:02.975 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:22:02.975 "name": "BaseBdev3", 00:22:02.975 "aliases": [ 00:22:02.975 "5bf03455-4402-4317-bb2f-d6fe7b2bb3bf" 00:22:02.975 ], 00:22:02.975 "product_name": "Malloc disk", 00:22:02.975 "block_size": 512, 00:22:02.975 "num_blocks": 65536, 00:22:02.975 "uuid": "5bf03455-4402-4317-bb2f-d6fe7b2bb3bf", 00:22:02.975 "assigned_rate_limits": { 00:22:02.975 "rw_ios_per_sec": 0, 00:22:02.975 "rw_mbytes_per_sec": 0, 00:22:02.975 "r_mbytes_per_sec": 0, 00:22:02.975 "w_mbytes_per_sec": 0 00:22:02.975 }, 00:22:02.975 "claimed": true, 00:22:02.975 "claim_type": "exclusive_write", 00:22:02.975 "zoned": false, 00:22:02.975 "supported_io_types": { 00:22:02.975 "read": true, 00:22:02.975 "write": true, 00:22:02.975 "unmap": true, 00:22:02.975 "flush": true, 00:22:02.975 "reset": true, 00:22:02.975 "nvme_admin": false, 00:22:02.975 "nvme_io": false, 00:22:02.975 "nvme_io_md": false, 00:22:02.975 "write_zeroes": true, 00:22:02.975 "zcopy": true, 00:22:02.975 "get_zone_info": false, 00:22:02.976 "zone_management": false, 00:22:02.976 "zone_append": false, 00:22:02.976 "compare": false, 00:22:02.976 "compare_and_write": false, 00:22:02.976 "abort": true, 00:22:02.976 "seek_hole": false, 00:22:02.976 "seek_data": false, 00:22:02.976 "copy": true, 00:22:02.976 "nvme_iov_md": false 00:22:02.976 }, 00:22:02.976 "memory_domains": [ 00:22:02.976 { 00:22:02.976 "dma_device_id": "system", 00:22:02.976 "dma_device_type": 1 00:22:02.976 }, 00:22:02.976 { 00:22:02.976 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:02.976 "dma_device_type": 2 00:22:02.976 } 00:22:02.976 ], 00:22:02.976 "driver_specific": {} 00:22:02.976 }' 00:22:02.976 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:02.976 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:02.976 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:22:02.976 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:02.976 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:02.976 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:22:02.976 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:02.976 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:03.233 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:22:03.233 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:03.233 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:03.233 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:22:03.233 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:22:03.233 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev4 00:22:03.233 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:22:03.491 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:22:03.491 "name": "BaseBdev4", 00:22:03.491 "aliases": [ 00:22:03.491 "fbc0fced-b415-4286-a3b5-75d75bd77d33" 00:22:03.491 ], 00:22:03.491 "product_name": "Malloc disk", 00:22:03.491 "block_size": 512, 00:22:03.491 "num_blocks": 65536, 00:22:03.491 "uuid": "fbc0fced-b415-4286-a3b5-75d75bd77d33", 00:22:03.491 "assigned_rate_limits": { 00:22:03.491 "rw_ios_per_sec": 0, 00:22:03.491 "rw_mbytes_per_sec": 0, 00:22:03.491 "r_mbytes_per_sec": 0, 00:22:03.491 "w_mbytes_per_sec": 0 00:22:03.491 }, 00:22:03.491 "claimed": true, 00:22:03.491 "claim_type": "exclusive_write", 00:22:03.491 "zoned": false, 00:22:03.491 "supported_io_types": { 00:22:03.491 "read": true, 00:22:03.491 "write": true, 00:22:03.491 "unmap": true, 00:22:03.491 "flush": true, 00:22:03.491 "reset": true, 00:22:03.491 "nvme_admin": false, 00:22:03.491 "nvme_io": false, 00:22:03.491 "nvme_io_md": false, 00:22:03.491 "write_zeroes": true, 00:22:03.491 "zcopy": true, 00:22:03.491 "get_zone_info": false, 00:22:03.491 "zone_management": false, 00:22:03.491 "zone_append": false, 00:22:03.491 "compare": false, 00:22:03.491 "compare_and_write": false, 00:22:03.491 "abort": true, 00:22:03.491 "seek_hole": false, 00:22:03.491 "seek_data": false, 00:22:03.491 "copy": true, 00:22:03.491 "nvme_iov_md": false 00:22:03.491 }, 00:22:03.491 "memory_domains": [ 00:22:03.491 { 00:22:03.491 "dma_device_id": "system", 00:22:03.491 "dma_device_type": 1 00:22:03.491 }, 00:22:03.491 { 00:22:03.491 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:03.491 "dma_device_type": 2 00:22:03.491 } 00:22:03.491 ], 00:22:03.491 "driver_specific": {} 00:22:03.491 }' 00:22:03.491 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:03.491 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:03.491 11:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:22:03.491 11:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:03.491 11:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:03.749 11:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:22:03.749 11:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:03.749 11:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:03.749 11:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:22:03.749 11:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:03.749 11:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:03.749 11:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:22:03.749 11:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@338 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:22:04.052 [2024-07-15 11:32:47.495830] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:22:04.052 [2024-07-15 11:32:47.495857] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:22:04.052 [2024-07-15 11:32:47.495906] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:22:04.052 [2024-07-15 11:32:47.496202] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:22:04.052 [2024-07-15 11:32:47.496216] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x2317180 name Existed_Raid, state offline 00:22:04.052 11:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@341 -- # killprocess 957029 00:22:04.052 11:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@948 -- # '[' -z 957029 ']' 00:22:04.052 11:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@952 -- # kill -0 957029 00:22:04.052 11:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@953 -- # uname 00:22:04.052 11:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:22:04.052 11:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 957029 00:22:04.052 11:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:22:04.052 11:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:22:04.052 11:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@966 -- # echo 'killing process with pid 957029' 00:22:04.052 killing process with pid 957029 00:22:04.052 11:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@967 -- # kill 957029 00:22:04.052 [2024-07-15 11:32:47.560896] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:22:04.052 11:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # wait 957029 00:22:04.052 [2024-07-15 11:32:47.598597] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:22:04.324 11:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@343 -- # return 0 00:22:04.324 00:22:04.324 real 0m32.220s 00:22:04.324 user 0m58.987s 00:22:04.324 sys 0m5.895s 00:22:04.324 11:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1124 -- # xtrace_disable 00:22:04.324 11:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:04.324 ************************************ 00:22:04.324 END TEST raid_state_function_test_sb 00:22:04.324 ************************************ 00:22:04.324 11:32:47 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:22:04.324 11:32:47 bdev_raid -- bdev/bdev_raid.sh@869 -- # run_test raid_superblock_test raid_superblock_test raid1 4 00:22:04.324 11:32:47 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:22:04.324 11:32:47 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:22:04.324 11:32:47 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:22:04.324 ************************************ 00:22:04.324 START TEST raid_superblock_test 00:22:04.324 ************************************ 00:22:04.324 11:32:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1123 -- # raid_superblock_test raid1 4 00:22:04.324 11:32:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@392 -- # local raid_level=raid1 00:22:04.324 11:32:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local num_base_bdevs=4 00:22:04.324 11:32:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # base_bdevs_malloc=() 00:22:04.324 11:32:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local base_bdevs_malloc 00:22:04.324 11:32:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_pt=() 00:22:04.324 11:32:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_pt 00:22:04.324 11:32:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt_uuid=() 00:22:04.324 11:32:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt_uuid 00:22:04.324 11:32:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local raid_bdev_name=raid_bdev1 00:22:04.324 11:32:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local strip_size 00:22:04.324 11:32:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size_create_arg 00:22:04.324 11:32:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local raid_bdev_uuid 00:22:04.324 11:32:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev 00:22:04.324 11:32:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@403 -- # '[' raid1 '!=' raid1 ']' 00:22:04.324 11:32:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@407 -- # strip_size=0 00:22:04.324 11:32:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # raid_pid=961900 00:22:04.324 11:32:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # waitforlisten 961900 /var/tmp/spdk-raid.sock 00:22:04.324 11:32:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@410 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -L bdev_raid 00:22:04.324 11:32:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@829 -- # '[' -z 961900 ']' 00:22:04.583 11:32:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:22:04.583 11:32:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:22:04.583 11:32:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:22:04.583 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:22:04.583 11:32:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:22:04.583 11:32:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:04.583 [2024-07-15 11:32:47.975471] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:22:04.583 [2024-07-15 11:32:47.975539] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid961900 ] 00:22:04.583 [2024-07-15 11:32:48.107650] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:22:04.842 [2024-07-15 11:32:48.205977] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:22:04.842 [2024-07-15 11:32:48.263005] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:22:04.842 [2024-07-15 11:32:48.263037] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:22:05.408 11:32:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:22:05.408 11:32:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@862 -- # return 0 00:22:05.408 11:32:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i = 1 )) 00:22:05.408 11:32:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:22:05.408 11:32:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc1 00:22:05.408 11:32:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt1 00:22:05.408 11:32:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:22:05.408 11:32:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:22:05.408 11:32:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:22:05.408 11:32:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:22:05.408 11:32:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc1 00:22:05.666 malloc1 00:22:05.666 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:22:05.924 [2024-07-15 11:32:49.326524] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:22:05.924 [2024-07-15 11:32:49.326575] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:05.924 [2024-07-15 11:32:49.326597] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x21c7570 00:22:05.924 [2024-07-15 11:32:49.326610] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:05.924 [2024-07-15 11:32:49.328250] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:05.924 [2024-07-15 11:32:49.328280] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:22:05.924 pt1 00:22:05.924 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:22:05.924 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:22:05.924 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc2 00:22:05.924 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt2 00:22:05.924 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:22:05.924 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:22:05.924 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:22:05.924 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:22:05.924 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc2 00:22:05.924 malloc2 00:22:06.182 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:22:06.182 [2024-07-15 11:32:49.672332] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:22:06.182 [2024-07-15 11:32:49.672380] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:06.182 [2024-07-15 11:32:49.672398] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x21c8970 00:22:06.182 [2024-07-15 11:32:49.672411] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:06.182 [2024-07-15 11:32:49.673940] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:06.182 [2024-07-15 11:32:49.673970] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:22:06.182 pt2 00:22:06.182 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:22:06.182 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:22:06.182 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc3 00:22:06.182 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt3 00:22:06.182 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:22:06.182 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:22:06.182 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:22:06.182 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:22:06.182 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc3 00:22:06.440 malloc3 00:22:06.440 11:32:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:22:06.440 [2024-07-15 11:32:50.033950] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:22:06.440 [2024-07-15 11:32:50.033996] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:06.440 [2024-07-15 11:32:50.034016] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x235f340 00:22:06.440 [2024-07-15 11:32:50.034030] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:06.698 [2024-07-15 11:32:50.035494] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:06.698 [2024-07-15 11:32:50.035522] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:22:06.698 pt3 00:22:06.698 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:22:06.698 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:22:06.698 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc4 00:22:06.698 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt4 00:22:06.698 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:22:06.698 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:22:06.698 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:22:06.698 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:22:06.698 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b malloc4 00:22:06.698 malloc4 00:22:06.956 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:22:06.956 [2024-07-15 11:32:50.455762] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:22:06.956 [2024-07-15 11:32:50.455808] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:06.956 [2024-07-15 11:32:50.455828] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2361c60 00:22:06.956 [2024-07-15 11:32:50.455841] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:06.956 [2024-07-15 11:32:50.457226] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:06.956 [2024-07-15 11:32:50.457253] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:22:06.956 pt4 00:22:06.956 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:22:06.956 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:22:06.956 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@429 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'pt1 pt2 pt3 pt4' -n raid_bdev1 -s 00:22:07.214 [2024-07-15 11:32:50.632263] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:22:07.214 [2024-07-15 11:32:50.633434] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:22:07.214 [2024-07-15 11:32:50.633488] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:22:07.214 [2024-07-15 11:32:50.633531] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:22:07.214 [2024-07-15 11:32:50.633704] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x21bf530 00:22:07.214 [2024-07-15 11:32:50.633715] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:22:07.214 [2024-07-15 11:32:50.633894] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x21bd770 00:22:07.214 [2024-07-15 11:32:50.634048] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x21bf530 00:22:07.214 [2024-07-15 11:32:50.634059] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x21bf530 00:22:07.214 [2024-07-15 11:32:50.634149] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:07.214 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:22:07.214 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:22:07.214 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:22:07.214 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:22:07.214 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:22:07.214 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:22:07.214 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:22:07.214 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:22:07.214 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:22:07.214 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:22:07.214 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:07.214 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:07.472 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:22:07.472 "name": "raid_bdev1", 00:22:07.472 "uuid": "e56cdd97-02a3-4569-80cb-249866628e6f", 00:22:07.472 "strip_size_kb": 0, 00:22:07.472 "state": "online", 00:22:07.472 "raid_level": "raid1", 00:22:07.472 "superblock": true, 00:22:07.472 "num_base_bdevs": 4, 00:22:07.472 "num_base_bdevs_discovered": 4, 00:22:07.472 "num_base_bdevs_operational": 4, 00:22:07.472 "base_bdevs_list": [ 00:22:07.472 { 00:22:07.472 "name": "pt1", 00:22:07.472 "uuid": "00000000-0000-0000-0000-000000000001", 00:22:07.472 "is_configured": true, 00:22:07.472 "data_offset": 2048, 00:22:07.472 "data_size": 63488 00:22:07.472 }, 00:22:07.472 { 00:22:07.472 "name": "pt2", 00:22:07.472 "uuid": "00000000-0000-0000-0000-000000000002", 00:22:07.472 "is_configured": true, 00:22:07.472 "data_offset": 2048, 00:22:07.472 "data_size": 63488 00:22:07.472 }, 00:22:07.472 { 00:22:07.472 "name": "pt3", 00:22:07.472 "uuid": "00000000-0000-0000-0000-000000000003", 00:22:07.472 "is_configured": true, 00:22:07.472 "data_offset": 2048, 00:22:07.472 "data_size": 63488 00:22:07.472 }, 00:22:07.472 { 00:22:07.472 "name": "pt4", 00:22:07.472 "uuid": "00000000-0000-0000-0000-000000000004", 00:22:07.472 "is_configured": true, 00:22:07.472 "data_offset": 2048, 00:22:07.472 "data_size": 63488 00:22:07.472 } 00:22:07.472 ] 00:22:07.472 }' 00:22:07.472 11:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:22:07.472 11:32:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:08.037 11:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_properties raid_bdev1 00:22:08.037 11:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:22:08.037 11:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:22:08.037 11:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:22:08.037 11:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:22:08.037 11:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # local name 00:22:08.037 11:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:22:08.037 11:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:22:08.294 [2024-07-15 11:32:51.659282] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:22:08.294 11:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:22:08.294 "name": "raid_bdev1", 00:22:08.294 "aliases": [ 00:22:08.294 "e56cdd97-02a3-4569-80cb-249866628e6f" 00:22:08.294 ], 00:22:08.294 "product_name": "Raid Volume", 00:22:08.294 "block_size": 512, 00:22:08.294 "num_blocks": 63488, 00:22:08.294 "uuid": "e56cdd97-02a3-4569-80cb-249866628e6f", 00:22:08.294 "assigned_rate_limits": { 00:22:08.294 "rw_ios_per_sec": 0, 00:22:08.294 "rw_mbytes_per_sec": 0, 00:22:08.294 "r_mbytes_per_sec": 0, 00:22:08.294 "w_mbytes_per_sec": 0 00:22:08.294 }, 00:22:08.294 "claimed": false, 00:22:08.294 "zoned": false, 00:22:08.294 "supported_io_types": { 00:22:08.294 "read": true, 00:22:08.294 "write": true, 00:22:08.294 "unmap": false, 00:22:08.294 "flush": false, 00:22:08.294 "reset": true, 00:22:08.294 "nvme_admin": false, 00:22:08.294 "nvme_io": false, 00:22:08.294 "nvme_io_md": false, 00:22:08.294 "write_zeroes": true, 00:22:08.294 "zcopy": false, 00:22:08.294 "get_zone_info": false, 00:22:08.294 "zone_management": false, 00:22:08.294 "zone_append": false, 00:22:08.294 "compare": false, 00:22:08.294 "compare_and_write": false, 00:22:08.294 "abort": false, 00:22:08.294 "seek_hole": false, 00:22:08.294 "seek_data": false, 00:22:08.294 "copy": false, 00:22:08.294 "nvme_iov_md": false 00:22:08.294 }, 00:22:08.294 "memory_domains": [ 00:22:08.294 { 00:22:08.294 "dma_device_id": "system", 00:22:08.294 "dma_device_type": 1 00:22:08.294 }, 00:22:08.294 { 00:22:08.294 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:08.294 "dma_device_type": 2 00:22:08.294 }, 00:22:08.294 { 00:22:08.294 "dma_device_id": "system", 00:22:08.294 "dma_device_type": 1 00:22:08.294 }, 00:22:08.294 { 00:22:08.294 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:08.294 "dma_device_type": 2 00:22:08.294 }, 00:22:08.294 { 00:22:08.294 "dma_device_id": "system", 00:22:08.294 "dma_device_type": 1 00:22:08.294 }, 00:22:08.294 { 00:22:08.294 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:08.294 "dma_device_type": 2 00:22:08.294 }, 00:22:08.294 { 00:22:08.294 "dma_device_id": "system", 00:22:08.294 "dma_device_type": 1 00:22:08.294 }, 00:22:08.294 { 00:22:08.294 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:08.295 "dma_device_type": 2 00:22:08.295 } 00:22:08.295 ], 00:22:08.295 "driver_specific": { 00:22:08.295 "raid": { 00:22:08.295 "uuid": "e56cdd97-02a3-4569-80cb-249866628e6f", 00:22:08.295 "strip_size_kb": 0, 00:22:08.295 "state": "online", 00:22:08.295 "raid_level": "raid1", 00:22:08.295 "superblock": true, 00:22:08.295 "num_base_bdevs": 4, 00:22:08.295 "num_base_bdevs_discovered": 4, 00:22:08.295 "num_base_bdevs_operational": 4, 00:22:08.295 "base_bdevs_list": [ 00:22:08.295 { 00:22:08.295 "name": "pt1", 00:22:08.295 "uuid": "00000000-0000-0000-0000-000000000001", 00:22:08.295 "is_configured": true, 00:22:08.295 "data_offset": 2048, 00:22:08.295 "data_size": 63488 00:22:08.295 }, 00:22:08.295 { 00:22:08.295 "name": "pt2", 00:22:08.295 "uuid": "00000000-0000-0000-0000-000000000002", 00:22:08.295 "is_configured": true, 00:22:08.295 "data_offset": 2048, 00:22:08.295 "data_size": 63488 00:22:08.295 }, 00:22:08.295 { 00:22:08.295 "name": "pt3", 00:22:08.295 "uuid": "00000000-0000-0000-0000-000000000003", 00:22:08.295 "is_configured": true, 00:22:08.295 "data_offset": 2048, 00:22:08.295 "data_size": 63488 00:22:08.295 }, 00:22:08.295 { 00:22:08.295 "name": "pt4", 00:22:08.295 "uuid": "00000000-0000-0000-0000-000000000004", 00:22:08.295 "is_configured": true, 00:22:08.295 "data_offset": 2048, 00:22:08.295 "data_size": 63488 00:22:08.295 } 00:22:08.295 ] 00:22:08.295 } 00:22:08.295 } 00:22:08.295 }' 00:22:08.295 11:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:22:08.295 11:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:22:08.295 pt2 00:22:08.295 pt3 00:22:08.295 pt4' 00:22:08.295 11:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:22:08.295 11:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:22:08.295 11:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:22:08.552 11:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:22:08.552 "name": "pt1", 00:22:08.552 "aliases": [ 00:22:08.552 "00000000-0000-0000-0000-000000000001" 00:22:08.552 ], 00:22:08.552 "product_name": "passthru", 00:22:08.552 "block_size": 512, 00:22:08.552 "num_blocks": 65536, 00:22:08.552 "uuid": "00000000-0000-0000-0000-000000000001", 00:22:08.552 "assigned_rate_limits": { 00:22:08.552 "rw_ios_per_sec": 0, 00:22:08.552 "rw_mbytes_per_sec": 0, 00:22:08.552 "r_mbytes_per_sec": 0, 00:22:08.552 "w_mbytes_per_sec": 0 00:22:08.552 }, 00:22:08.552 "claimed": true, 00:22:08.552 "claim_type": "exclusive_write", 00:22:08.552 "zoned": false, 00:22:08.552 "supported_io_types": { 00:22:08.552 "read": true, 00:22:08.552 "write": true, 00:22:08.552 "unmap": true, 00:22:08.552 "flush": true, 00:22:08.552 "reset": true, 00:22:08.552 "nvme_admin": false, 00:22:08.552 "nvme_io": false, 00:22:08.552 "nvme_io_md": false, 00:22:08.552 "write_zeroes": true, 00:22:08.552 "zcopy": true, 00:22:08.552 "get_zone_info": false, 00:22:08.552 "zone_management": false, 00:22:08.552 "zone_append": false, 00:22:08.552 "compare": false, 00:22:08.552 "compare_and_write": false, 00:22:08.552 "abort": true, 00:22:08.552 "seek_hole": false, 00:22:08.552 "seek_data": false, 00:22:08.552 "copy": true, 00:22:08.552 "nvme_iov_md": false 00:22:08.552 }, 00:22:08.552 "memory_domains": [ 00:22:08.552 { 00:22:08.552 "dma_device_id": "system", 00:22:08.552 "dma_device_type": 1 00:22:08.552 }, 00:22:08.552 { 00:22:08.552 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:08.552 "dma_device_type": 2 00:22:08.552 } 00:22:08.552 ], 00:22:08.552 "driver_specific": { 00:22:08.552 "passthru": { 00:22:08.552 "name": "pt1", 00:22:08.552 "base_bdev_name": "malloc1" 00:22:08.552 } 00:22:08.552 } 00:22:08.552 }' 00:22:08.552 11:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:08.552 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:08.552 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:22:08.552 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:08.552 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:08.552 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:22:08.552 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:08.809 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:08.809 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:22:08.809 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:08.809 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:08.809 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:22:08.809 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:22:08.809 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:22:08.809 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:22:09.067 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:22:09.067 "name": "pt2", 00:22:09.067 "aliases": [ 00:22:09.067 "00000000-0000-0000-0000-000000000002" 00:22:09.067 ], 00:22:09.067 "product_name": "passthru", 00:22:09.067 "block_size": 512, 00:22:09.067 "num_blocks": 65536, 00:22:09.067 "uuid": "00000000-0000-0000-0000-000000000002", 00:22:09.067 "assigned_rate_limits": { 00:22:09.067 "rw_ios_per_sec": 0, 00:22:09.067 "rw_mbytes_per_sec": 0, 00:22:09.067 "r_mbytes_per_sec": 0, 00:22:09.067 "w_mbytes_per_sec": 0 00:22:09.067 }, 00:22:09.067 "claimed": true, 00:22:09.067 "claim_type": "exclusive_write", 00:22:09.067 "zoned": false, 00:22:09.067 "supported_io_types": { 00:22:09.067 "read": true, 00:22:09.067 "write": true, 00:22:09.067 "unmap": true, 00:22:09.067 "flush": true, 00:22:09.067 "reset": true, 00:22:09.067 "nvme_admin": false, 00:22:09.067 "nvme_io": false, 00:22:09.067 "nvme_io_md": false, 00:22:09.067 "write_zeroes": true, 00:22:09.067 "zcopy": true, 00:22:09.067 "get_zone_info": false, 00:22:09.067 "zone_management": false, 00:22:09.067 "zone_append": false, 00:22:09.067 "compare": false, 00:22:09.067 "compare_and_write": false, 00:22:09.067 "abort": true, 00:22:09.067 "seek_hole": false, 00:22:09.067 "seek_data": false, 00:22:09.067 "copy": true, 00:22:09.067 "nvme_iov_md": false 00:22:09.067 }, 00:22:09.067 "memory_domains": [ 00:22:09.067 { 00:22:09.067 "dma_device_id": "system", 00:22:09.067 "dma_device_type": 1 00:22:09.067 }, 00:22:09.067 { 00:22:09.067 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:09.067 "dma_device_type": 2 00:22:09.067 } 00:22:09.067 ], 00:22:09.067 "driver_specific": { 00:22:09.067 "passthru": { 00:22:09.067 "name": "pt2", 00:22:09.067 "base_bdev_name": "malloc2" 00:22:09.067 } 00:22:09.067 } 00:22:09.067 }' 00:22:09.067 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:09.067 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:09.067 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:22:09.067 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:09.067 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:09.067 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:22:09.325 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:09.325 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:09.325 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:22:09.325 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:09.325 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:09.325 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:22:09.325 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:22:09.325 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt3 00:22:09.325 11:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:22:09.890 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:22:09.890 "name": "pt3", 00:22:09.890 "aliases": [ 00:22:09.890 "00000000-0000-0000-0000-000000000003" 00:22:09.890 ], 00:22:09.890 "product_name": "passthru", 00:22:09.890 "block_size": 512, 00:22:09.890 "num_blocks": 65536, 00:22:09.890 "uuid": "00000000-0000-0000-0000-000000000003", 00:22:09.890 "assigned_rate_limits": { 00:22:09.890 "rw_ios_per_sec": 0, 00:22:09.890 "rw_mbytes_per_sec": 0, 00:22:09.890 "r_mbytes_per_sec": 0, 00:22:09.890 "w_mbytes_per_sec": 0 00:22:09.890 }, 00:22:09.890 "claimed": true, 00:22:09.890 "claim_type": "exclusive_write", 00:22:09.890 "zoned": false, 00:22:09.890 "supported_io_types": { 00:22:09.890 "read": true, 00:22:09.890 "write": true, 00:22:09.890 "unmap": true, 00:22:09.890 "flush": true, 00:22:09.890 "reset": true, 00:22:09.890 "nvme_admin": false, 00:22:09.890 "nvme_io": false, 00:22:09.890 "nvme_io_md": false, 00:22:09.890 "write_zeroes": true, 00:22:09.890 "zcopy": true, 00:22:09.890 "get_zone_info": false, 00:22:09.890 "zone_management": false, 00:22:09.890 "zone_append": false, 00:22:09.890 "compare": false, 00:22:09.890 "compare_and_write": false, 00:22:09.890 "abort": true, 00:22:09.890 "seek_hole": false, 00:22:09.890 "seek_data": false, 00:22:09.890 "copy": true, 00:22:09.890 "nvme_iov_md": false 00:22:09.890 }, 00:22:09.890 "memory_domains": [ 00:22:09.890 { 00:22:09.890 "dma_device_id": "system", 00:22:09.890 "dma_device_type": 1 00:22:09.890 }, 00:22:09.890 { 00:22:09.890 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:09.890 "dma_device_type": 2 00:22:09.890 } 00:22:09.890 ], 00:22:09.890 "driver_specific": { 00:22:09.890 "passthru": { 00:22:09.890 "name": "pt3", 00:22:09.890 "base_bdev_name": "malloc3" 00:22:09.890 } 00:22:09.890 } 00:22:09.890 }' 00:22:09.890 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:09.890 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:09.890 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:22:09.890 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:09.890 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:09.890 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:22:09.890 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:10.147 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:10.147 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:22:10.147 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:10.147 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:10.147 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:22:10.147 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:22:10.147 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt4 00:22:10.147 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:22:10.404 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:22:10.404 "name": "pt4", 00:22:10.404 "aliases": [ 00:22:10.404 "00000000-0000-0000-0000-000000000004" 00:22:10.404 ], 00:22:10.404 "product_name": "passthru", 00:22:10.404 "block_size": 512, 00:22:10.404 "num_blocks": 65536, 00:22:10.404 "uuid": "00000000-0000-0000-0000-000000000004", 00:22:10.404 "assigned_rate_limits": { 00:22:10.404 "rw_ios_per_sec": 0, 00:22:10.404 "rw_mbytes_per_sec": 0, 00:22:10.404 "r_mbytes_per_sec": 0, 00:22:10.404 "w_mbytes_per_sec": 0 00:22:10.404 }, 00:22:10.404 "claimed": true, 00:22:10.404 "claim_type": "exclusive_write", 00:22:10.404 "zoned": false, 00:22:10.404 "supported_io_types": { 00:22:10.404 "read": true, 00:22:10.404 "write": true, 00:22:10.404 "unmap": true, 00:22:10.404 "flush": true, 00:22:10.404 "reset": true, 00:22:10.404 "nvme_admin": false, 00:22:10.404 "nvme_io": false, 00:22:10.404 "nvme_io_md": false, 00:22:10.404 "write_zeroes": true, 00:22:10.404 "zcopy": true, 00:22:10.404 "get_zone_info": false, 00:22:10.404 "zone_management": false, 00:22:10.404 "zone_append": false, 00:22:10.404 "compare": false, 00:22:10.404 "compare_and_write": false, 00:22:10.404 "abort": true, 00:22:10.404 "seek_hole": false, 00:22:10.404 "seek_data": false, 00:22:10.404 "copy": true, 00:22:10.404 "nvme_iov_md": false 00:22:10.404 }, 00:22:10.404 "memory_domains": [ 00:22:10.404 { 00:22:10.404 "dma_device_id": "system", 00:22:10.404 "dma_device_type": 1 00:22:10.404 }, 00:22:10.404 { 00:22:10.404 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:10.404 "dma_device_type": 2 00:22:10.404 } 00:22:10.404 ], 00:22:10.404 "driver_specific": { 00:22:10.404 "passthru": { 00:22:10.404 "name": "pt4", 00:22:10.404 "base_bdev_name": "malloc4" 00:22:10.404 } 00:22:10.404 } 00:22:10.404 }' 00:22:10.404 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:10.404 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:10.404 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:22:10.404 11:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:10.662 11:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:10.662 11:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:22:10.662 11:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:10.662 11:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:10.662 11:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:22:10.662 11:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:10.662 11:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:10.920 11:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:22:10.920 11:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:22:10.920 11:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # jq -r '.[] | .uuid' 00:22:10.920 [2024-07-15 11:32:54.494768] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:22:11.178 11:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@434 -- # raid_bdev_uuid=e56cdd97-02a3-4569-80cb-249866628e6f 00:22:11.178 11:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # '[' -z e56cdd97-02a3-4569-80cb-249866628e6f ']' 00:22:11.178 11:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@440 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:22:11.178 [2024-07-15 11:32:54.727080] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:22:11.178 [2024-07-15 11:32:54.727100] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:22:11.178 [2024-07-15 11:32:54.727146] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:22:11.178 [2024-07-15 11:32:54.727229] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:22:11.178 [2024-07-15 11:32:54.727241] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x21bf530 name raid_bdev1, state offline 00:22:11.178 11:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:11.178 11:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # jq -r '.[]' 00:22:11.436 11:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # raid_bdev= 00:22:11.436 11:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # '[' -n '' ']' 00:22:11.436 11:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:22:11.436 11:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt1 00:22:11.695 11:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:22:11.695 11:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:22:11.952 11:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:22:11.952 11:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt3 00:22:12.210 11:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:22:12.210 11:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt4 00:22:12.210 11:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs 00:22:12.210 11:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:22:12.467 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@450 -- # '[' false == true ']' 00:22:12.467 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@456 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'malloc1 malloc2 malloc3 malloc4' -n raid_bdev1 00:22:12.467 11:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@648 -- # local es=0 00:22:12.467 11:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'malloc1 malloc2 malloc3 malloc4' -n raid_bdev1 00:22:12.467 11:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:22:12.467 11:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:22:12.467 11:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:22:12.468 11:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:22:12.468 11:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:22:12.468 11:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:22:12.468 11:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:22:12.468 11:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py ]] 00:22:12.468 11:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'malloc1 malloc2 malloc3 malloc4' -n raid_bdev1 00:22:12.725 [2024-07-15 11:32:56.251043] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:22:12.725 [2024-07-15 11:32:56.252445] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:22:12.725 [2024-07-15 11:32:56.252491] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:22:12.725 [2024-07-15 11:32:56.252526] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:22:12.725 [2024-07-15 11:32:56.252572] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:22:12.725 [2024-07-15 11:32:56.252612] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:22:12.725 [2024-07-15 11:32:56.252634] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:22:12.725 [2024-07-15 11:32:56.252657] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:22:12.725 [2024-07-15 11:32:56.252675] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:22:12.725 [2024-07-15 11:32:56.252686] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x236aff0 name raid_bdev1, state configuring 00:22:12.725 request: 00:22:12.725 { 00:22:12.725 "name": "raid_bdev1", 00:22:12.725 "raid_level": "raid1", 00:22:12.725 "base_bdevs": [ 00:22:12.725 "malloc1", 00:22:12.725 "malloc2", 00:22:12.725 "malloc3", 00:22:12.725 "malloc4" 00:22:12.725 ], 00:22:12.725 "superblock": false, 00:22:12.725 "method": "bdev_raid_create", 00:22:12.725 "req_id": 1 00:22:12.725 } 00:22:12.725 Got JSON-RPC error response 00:22:12.725 response: 00:22:12.725 { 00:22:12.725 "code": -17, 00:22:12.725 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:22:12.725 } 00:22:12.725 11:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@651 -- # es=1 00:22:12.725 11:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:22:12.725 11:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:22:12.725 11:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:22:12.725 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:12.725 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # jq -r '.[]' 00:22:12.982 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@458 -- # raid_bdev= 00:22:12.982 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # '[' -n '' ']' 00:22:12.982 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@464 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:22:13.248 [2024-07-15 11:32:56.656061] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:22:13.248 [2024-07-15 11:32:56.656102] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:13.248 [2024-07-15 11:32:56.656122] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x21c77a0 00:22:13.248 [2024-07-15 11:32:56.656136] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:13.248 [2024-07-15 11:32:56.657730] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:13.248 [2024-07-15 11:32:56.657759] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:22:13.248 [2024-07-15 11:32:56.657823] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:22:13.248 [2024-07-15 11:32:56.657851] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:22:13.248 pt1 00:22:13.248 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@467 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:22:13.248 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:22:13.248 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:22:13.248 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:22:13.248 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:22:13.248 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:22:13.248 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:22:13.248 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:22:13.248 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:22:13.248 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:22:13.248 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:13.248 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:13.506 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:22:13.506 "name": "raid_bdev1", 00:22:13.506 "uuid": "e56cdd97-02a3-4569-80cb-249866628e6f", 00:22:13.506 "strip_size_kb": 0, 00:22:13.506 "state": "configuring", 00:22:13.506 "raid_level": "raid1", 00:22:13.506 "superblock": true, 00:22:13.506 "num_base_bdevs": 4, 00:22:13.506 "num_base_bdevs_discovered": 1, 00:22:13.506 "num_base_bdevs_operational": 4, 00:22:13.506 "base_bdevs_list": [ 00:22:13.506 { 00:22:13.506 "name": "pt1", 00:22:13.506 "uuid": "00000000-0000-0000-0000-000000000001", 00:22:13.506 "is_configured": true, 00:22:13.506 "data_offset": 2048, 00:22:13.506 "data_size": 63488 00:22:13.506 }, 00:22:13.506 { 00:22:13.506 "name": null, 00:22:13.506 "uuid": "00000000-0000-0000-0000-000000000002", 00:22:13.506 "is_configured": false, 00:22:13.506 "data_offset": 2048, 00:22:13.506 "data_size": 63488 00:22:13.506 }, 00:22:13.506 { 00:22:13.506 "name": null, 00:22:13.506 "uuid": "00000000-0000-0000-0000-000000000003", 00:22:13.506 "is_configured": false, 00:22:13.506 "data_offset": 2048, 00:22:13.506 "data_size": 63488 00:22:13.506 }, 00:22:13.506 { 00:22:13.506 "name": null, 00:22:13.506 "uuid": "00000000-0000-0000-0000-000000000004", 00:22:13.506 "is_configured": false, 00:22:13.506 "data_offset": 2048, 00:22:13.506 "data_size": 63488 00:22:13.506 } 00:22:13.506 ] 00:22:13.506 }' 00:22:13.506 11:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:22:13.506 11:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:14.072 11:32:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@469 -- # '[' 4 -gt 2 ']' 00:22:14.072 11:32:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@471 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:22:14.329 [2024-07-15 11:32:57.734933] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:22:14.329 [2024-07-15 11:32:57.734986] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:14.329 [2024-07-15 11:32:57.735006] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2360940 00:22:14.329 [2024-07-15 11:32:57.735019] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:14.329 [2024-07-15 11:32:57.735362] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:14.329 [2024-07-15 11:32:57.735380] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:22:14.329 [2024-07-15 11:32:57.735441] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:22:14.329 [2024-07-15 11:32:57.735460] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:22:14.329 pt2 00:22:14.329 11:32:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:22:14.587 [2024-07-15 11:32:57.983595] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:22:14.587 11:32:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:22:14.587 11:32:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:22:14.587 11:32:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:22:14.587 11:32:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:22:14.587 11:32:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:22:14.587 11:32:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:22:14.587 11:32:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:22:14.587 11:32:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:22:14.587 11:32:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:22:14.587 11:32:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:22:14.587 11:32:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:14.587 11:32:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:14.845 11:32:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:22:14.845 "name": "raid_bdev1", 00:22:14.845 "uuid": "e56cdd97-02a3-4569-80cb-249866628e6f", 00:22:14.845 "strip_size_kb": 0, 00:22:14.845 "state": "configuring", 00:22:14.845 "raid_level": "raid1", 00:22:14.845 "superblock": true, 00:22:14.845 "num_base_bdevs": 4, 00:22:14.845 "num_base_bdevs_discovered": 1, 00:22:14.845 "num_base_bdevs_operational": 4, 00:22:14.845 "base_bdevs_list": [ 00:22:14.845 { 00:22:14.845 "name": "pt1", 00:22:14.845 "uuid": "00000000-0000-0000-0000-000000000001", 00:22:14.845 "is_configured": true, 00:22:14.845 "data_offset": 2048, 00:22:14.845 "data_size": 63488 00:22:14.845 }, 00:22:14.845 { 00:22:14.845 "name": null, 00:22:14.845 "uuid": "00000000-0000-0000-0000-000000000002", 00:22:14.845 "is_configured": false, 00:22:14.845 "data_offset": 2048, 00:22:14.845 "data_size": 63488 00:22:14.845 }, 00:22:14.845 { 00:22:14.845 "name": null, 00:22:14.845 "uuid": "00000000-0000-0000-0000-000000000003", 00:22:14.845 "is_configured": false, 00:22:14.845 "data_offset": 2048, 00:22:14.845 "data_size": 63488 00:22:14.845 }, 00:22:14.845 { 00:22:14.845 "name": null, 00:22:14.845 "uuid": "00000000-0000-0000-0000-000000000004", 00:22:14.845 "is_configured": false, 00:22:14.845 "data_offset": 2048, 00:22:14.845 "data_size": 63488 00:22:14.845 } 00:22:14.845 ] 00:22:14.845 }' 00:22:14.845 11:32:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:22:14.845 11:32:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:15.409 11:32:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i = 1 )) 00:22:15.409 11:32:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:22:15.409 11:32:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:22:15.667 [2024-07-15 11:32:59.050435] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:22:15.667 [2024-07-15 11:32:59.050490] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:15.667 [2024-07-15 11:32:59.050510] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x21be060 00:22:15.667 [2024-07-15 11:32:59.050522] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:15.667 [2024-07-15 11:32:59.050869] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:15.667 [2024-07-15 11:32:59.050887] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:22:15.667 [2024-07-15 11:32:59.050971] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:22:15.667 [2024-07-15 11:32:59.050999] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:22:15.667 pt2 00:22:15.667 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:22:15.667 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:22:15.667 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:22:15.924 [2024-07-15 11:32:59.283063] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:22:15.924 [2024-07-15 11:32:59.283106] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:15.924 [2024-07-15 11:32:59.283128] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x21c08d0 00:22:15.924 [2024-07-15 11:32:59.283141] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:15.924 [2024-07-15 11:32:59.283466] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:15.924 [2024-07-15 11:32:59.283484] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:22:15.924 [2024-07-15 11:32:59.283544] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:22:15.924 [2024-07-15 11:32:59.283562] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:22:15.924 pt3 00:22:15.924 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:22:15.924 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:22:15.924 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:22:15.924 [2024-07-15 11:32:59.515691] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:22:15.924 [2024-07-15 11:32:59.515739] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:15.924 [2024-07-15 11:32:59.515760] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x21c1b80 00:22:15.924 [2024-07-15 11:32:59.515773] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:15.924 [2024-07-15 11:32:59.516134] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:15.924 [2024-07-15 11:32:59.516153] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:22:15.924 [2024-07-15 11:32:59.516216] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:22:15.924 [2024-07-15 11:32:59.516235] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:22:15.924 [2024-07-15 11:32:59.516364] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x21be780 00:22:15.924 [2024-07-15 11:32:59.516374] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:22:15.924 [2024-07-15 11:32:59.516551] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x21c3fa0 00:22:15.924 [2024-07-15 11:32:59.516685] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x21be780 00:22:15.924 [2024-07-15 11:32:59.516695] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x21be780 00:22:15.924 [2024-07-15 11:32:59.516792] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:16.182 pt4 00:22:16.182 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:22:16.182 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:22:16.182 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@482 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:22:16.182 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:22:16.182 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:22:16.182 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:22:16.182 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:22:16.182 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:22:16.182 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:22:16.182 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:22:16.182 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:22:16.182 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:22:16.182 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:16.182 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:16.440 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:22:16.440 "name": "raid_bdev1", 00:22:16.440 "uuid": "e56cdd97-02a3-4569-80cb-249866628e6f", 00:22:16.440 "strip_size_kb": 0, 00:22:16.440 "state": "online", 00:22:16.440 "raid_level": "raid1", 00:22:16.440 "superblock": true, 00:22:16.440 "num_base_bdevs": 4, 00:22:16.440 "num_base_bdevs_discovered": 4, 00:22:16.440 "num_base_bdevs_operational": 4, 00:22:16.440 "base_bdevs_list": [ 00:22:16.440 { 00:22:16.440 "name": "pt1", 00:22:16.440 "uuid": "00000000-0000-0000-0000-000000000001", 00:22:16.440 "is_configured": true, 00:22:16.440 "data_offset": 2048, 00:22:16.440 "data_size": 63488 00:22:16.440 }, 00:22:16.440 { 00:22:16.440 "name": "pt2", 00:22:16.440 "uuid": "00000000-0000-0000-0000-000000000002", 00:22:16.440 "is_configured": true, 00:22:16.440 "data_offset": 2048, 00:22:16.440 "data_size": 63488 00:22:16.440 }, 00:22:16.440 { 00:22:16.440 "name": "pt3", 00:22:16.440 "uuid": "00000000-0000-0000-0000-000000000003", 00:22:16.440 "is_configured": true, 00:22:16.440 "data_offset": 2048, 00:22:16.440 "data_size": 63488 00:22:16.440 }, 00:22:16.440 { 00:22:16.440 "name": "pt4", 00:22:16.440 "uuid": "00000000-0000-0000-0000-000000000004", 00:22:16.440 "is_configured": true, 00:22:16.440 "data_offset": 2048, 00:22:16.440 "data_size": 63488 00:22:16.440 } 00:22:16.440 ] 00:22:16.440 }' 00:22:16.440 11:32:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:22:16.440 11:32:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:17.371 11:33:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_properties raid_bdev1 00:22:17.371 11:33:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:22:17.371 11:33:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:22:17.371 11:33:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:22:17.371 11:33:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:22:17.371 11:33:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # local name 00:22:17.371 11:33:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:22:17.371 11:33:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:22:17.371 [2024-07-15 11:33:00.871605] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:22:17.371 11:33:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:22:17.371 "name": "raid_bdev1", 00:22:17.371 "aliases": [ 00:22:17.371 "e56cdd97-02a3-4569-80cb-249866628e6f" 00:22:17.371 ], 00:22:17.371 "product_name": "Raid Volume", 00:22:17.371 "block_size": 512, 00:22:17.371 "num_blocks": 63488, 00:22:17.371 "uuid": "e56cdd97-02a3-4569-80cb-249866628e6f", 00:22:17.371 "assigned_rate_limits": { 00:22:17.371 "rw_ios_per_sec": 0, 00:22:17.371 "rw_mbytes_per_sec": 0, 00:22:17.371 "r_mbytes_per_sec": 0, 00:22:17.371 "w_mbytes_per_sec": 0 00:22:17.371 }, 00:22:17.371 "claimed": false, 00:22:17.371 "zoned": false, 00:22:17.371 "supported_io_types": { 00:22:17.371 "read": true, 00:22:17.371 "write": true, 00:22:17.371 "unmap": false, 00:22:17.371 "flush": false, 00:22:17.371 "reset": true, 00:22:17.371 "nvme_admin": false, 00:22:17.371 "nvme_io": false, 00:22:17.371 "nvme_io_md": false, 00:22:17.371 "write_zeroes": true, 00:22:17.371 "zcopy": false, 00:22:17.371 "get_zone_info": false, 00:22:17.371 "zone_management": false, 00:22:17.371 "zone_append": false, 00:22:17.371 "compare": false, 00:22:17.371 "compare_and_write": false, 00:22:17.371 "abort": false, 00:22:17.371 "seek_hole": false, 00:22:17.371 "seek_data": false, 00:22:17.371 "copy": false, 00:22:17.371 "nvme_iov_md": false 00:22:17.371 }, 00:22:17.371 "memory_domains": [ 00:22:17.371 { 00:22:17.371 "dma_device_id": "system", 00:22:17.371 "dma_device_type": 1 00:22:17.371 }, 00:22:17.371 { 00:22:17.371 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:17.371 "dma_device_type": 2 00:22:17.371 }, 00:22:17.371 { 00:22:17.371 "dma_device_id": "system", 00:22:17.371 "dma_device_type": 1 00:22:17.371 }, 00:22:17.371 { 00:22:17.371 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:17.371 "dma_device_type": 2 00:22:17.371 }, 00:22:17.371 { 00:22:17.371 "dma_device_id": "system", 00:22:17.371 "dma_device_type": 1 00:22:17.371 }, 00:22:17.371 { 00:22:17.371 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:17.371 "dma_device_type": 2 00:22:17.371 }, 00:22:17.371 { 00:22:17.371 "dma_device_id": "system", 00:22:17.371 "dma_device_type": 1 00:22:17.372 }, 00:22:17.372 { 00:22:17.372 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:17.372 "dma_device_type": 2 00:22:17.372 } 00:22:17.372 ], 00:22:17.372 "driver_specific": { 00:22:17.372 "raid": { 00:22:17.372 "uuid": "e56cdd97-02a3-4569-80cb-249866628e6f", 00:22:17.372 "strip_size_kb": 0, 00:22:17.372 "state": "online", 00:22:17.372 "raid_level": "raid1", 00:22:17.372 "superblock": true, 00:22:17.372 "num_base_bdevs": 4, 00:22:17.372 "num_base_bdevs_discovered": 4, 00:22:17.372 "num_base_bdevs_operational": 4, 00:22:17.372 "base_bdevs_list": [ 00:22:17.372 { 00:22:17.372 "name": "pt1", 00:22:17.372 "uuid": "00000000-0000-0000-0000-000000000001", 00:22:17.372 "is_configured": true, 00:22:17.372 "data_offset": 2048, 00:22:17.372 "data_size": 63488 00:22:17.372 }, 00:22:17.372 { 00:22:17.372 "name": "pt2", 00:22:17.372 "uuid": "00000000-0000-0000-0000-000000000002", 00:22:17.372 "is_configured": true, 00:22:17.372 "data_offset": 2048, 00:22:17.372 "data_size": 63488 00:22:17.372 }, 00:22:17.372 { 00:22:17.372 "name": "pt3", 00:22:17.372 "uuid": "00000000-0000-0000-0000-000000000003", 00:22:17.372 "is_configured": true, 00:22:17.372 "data_offset": 2048, 00:22:17.372 "data_size": 63488 00:22:17.372 }, 00:22:17.372 { 00:22:17.372 "name": "pt4", 00:22:17.372 "uuid": "00000000-0000-0000-0000-000000000004", 00:22:17.372 "is_configured": true, 00:22:17.372 "data_offset": 2048, 00:22:17.372 "data_size": 63488 00:22:17.372 } 00:22:17.372 ] 00:22:17.372 } 00:22:17.372 } 00:22:17.372 }' 00:22:17.372 11:33:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:22:17.372 11:33:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:22:17.372 pt2 00:22:17.372 pt3 00:22:17.372 pt4' 00:22:17.372 11:33:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:22:17.372 11:33:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:22:17.372 11:33:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:22:17.628 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:22:17.628 "name": "pt1", 00:22:17.628 "aliases": [ 00:22:17.628 "00000000-0000-0000-0000-000000000001" 00:22:17.628 ], 00:22:17.628 "product_name": "passthru", 00:22:17.628 "block_size": 512, 00:22:17.628 "num_blocks": 65536, 00:22:17.628 "uuid": "00000000-0000-0000-0000-000000000001", 00:22:17.629 "assigned_rate_limits": { 00:22:17.629 "rw_ios_per_sec": 0, 00:22:17.629 "rw_mbytes_per_sec": 0, 00:22:17.629 "r_mbytes_per_sec": 0, 00:22:17.629 "w_mbytes_per_sec": 0 00:22:17.629 }, 00:22:17.629 "claimed": true, 00:22:17.629 "claim_type": "exclusive_write", 00:22:17.629 "zoned": false, 00:22:17.629 "supported_io_types": { 00:22:17.629 "read": true, 00:22:17.629 "write": true, 00:22:17.629 "unmap": true, 00:22:17.629 "flush": true, 00:22:17.629 "reset": true, 00:22:17.629 "nvme_admin": false, 00:22:17.629 "nvme_io": false, 00:22:17.629 "nvme_io_md": false, 00:22:17.629 "write_zeroes": true, 00:22:17.629 "zcopy": true, 00:22:17.629 "get_zone_info": false, 00:22:17.629 "zone_management": false, 00:22:17.629 "zone_append": false, 00:22:17.629 "compare": false, 00:22:17.629 "compare_and_write": false, 00:22:17.629 "abort": true, 00:22:17.629 "seek_hole": false, 00:22:17.629 "seek_data": false, 00:22:17.629 "copy": true, 00:22:17.629 "nvme_iov_md": false 00:22:17.629 }, 00:22:17.629 "memory_domains": [ 00:22:17.629 { 00:22:17.629 "dma_device_id": "system", 00:22:17.629 "dma_device_type": 1 00:22:17.629 }, 00:22:17.629 { 00:22:17.629 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:17.629 "dma_device_type": 2 00:22:17.629 } 00:22:17.629 ], 00:22:17.629 "driver_specific": { 00:22:17.629 "passthru": { 00:22:17.629 "name": "pt1", 00:22:17.629 "base_bdev_name": "malloc1" 00:22:17.629 } 00:22:17.629 } 00:22:17.629 }' 00:22:17.629 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:17.629 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:17.885 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:22:17.885 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:17.885 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:17.885 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:22:17.885 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:17.885 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:17.885 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:22:17.885 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:18.142 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:18.143 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:22:18.143 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:22:18.143 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:22:18.143 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:22:18.400 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:22:18.400 "name": "pt2", 00:22:18.400 "aliases": [ 00:22:18.400 "00000000-0000-0000-0000-000000000002" 00:22:18.400 ], 00:22:18.400 "product_name": "passthru", 00:22:18.400 "block_size": 512, 00:22:18.400 "num_blocks": 65536, 00:22:18.400 "uuid": "00000000-0000-0000-0000-000000000002", 00:22:18.400 "assigned_rate_limits": { 00:22:18.400 "rw_ios_per_sec": 0, 00:22:18.400 "rw_mbytes_per_sec": 0, 00:22:18.400 "r_mbytes_per_sec": 0, 00:22:18.400 "w_mbytes_per_sec": 0 00:22:18.400 }, 00:22:18.400 "claimed": true, 00:22:18.400 "claim_type": "exclusive_write", 00:22:18.400 "zoned": false, 00:22:18.400 "supported_io_types": { 00:22:18.400 "read": true, 00:22:18.400 "write": true, 00:22:18.400 "unmap": true, 00:22:18.400 "flush": true, 00:22:18.400 "reset": true, 00:22:18.400 "nvme_admin": false, 00:22:18.400 "nvme_io": false, 00:22:18.400 "nvme_io_md": false, 00:22:18.400 "write_zeroes": true, 00:22:18.400 "zcopy": true, 00:22:18.400 "get_zone_info": false, 00:22:18.400 "zone_management": false, 00:22:18.400 "zone_append": false, 00:22:18.400 "compare": false, 00:22:18.400 "compare_and_write": false, 00:22:18.400 "abort": true, 00:22:18.400 "seek_hole": false, 00:22:18.400 "seek_data": false, 00:22:18.400 "copy": true, 00:22:18.400 "nvme_iov_md": false 00:22:18.400 }, 00:22:18.400 "memory_domains": [ 00:22:18.400 { 00:22:18.400 "dma_device_id": "system", 00:22:18.400 "dma_device_type": 1 00:22:18.400 }, 00:22:18.400 { 00:22:18.400 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:18.400 "dma_device_type": 2 00:22:18.400 } 00:22:18.400 ], 00:22:18.400 "driver_specific": { 00:22:18.400 "passthru": { 00:22:18.400 "name": "pt2", 00:22:18.400 "base_bdev_name": "malloc2" 00:22:18.400 } 00:22:18.400 } 00:22:18.400 }' 00:22:18.400 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:18.400 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:18.400 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:22:18.400 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:18.400 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:18.400 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:22:18.400 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:18.657 11:33:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:18.657 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:22:18.657 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:18.657 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:18.657 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:22:18.657 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:22:18.657 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt3 00:22:18.657 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:22:18.915 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:22:18.915 "name": "pt3", 00:22:18.915 "aliases": [ 00:22:18.915 "00000000-0000-0000-0000-000000000003" 00:22:18.915 ], 00:22:18.915 "product_name": "passthru", 00:22:18.915 "block_size": 512, 00:22:18.915 "num_blocks": 65536, 00:22:18.915 "uuid": "00000000-0000-0000-0000-000000000003", 00:22:18.915 "assigned_rate_limits": { 00:22:18.915 "rw_ios_per_sec": 0, 00:22:18.915 "rw_mbytes_per_sec": 0, 00:22:18.915 "r_mbytes_per_sec": 0, 00:22:18.915 "w_mbytes_per_sec": 0 00:22:18.915 }, 00:22:18.915 "claimed": true, 00:22:18.915 "claim_type": "exclusive_write", 00:22:18.915 "zoned": false, 00:22:18.915 "supported_io_types": { 00:22:18.915 "read": true, 00:22:18.915 "write": true, 00:22:18.915 "unmap": true, 00:22:18.915 "flush": true, 00:22:18.915 "reset": true, 00:22:18.915 "nvme_admin": false, 00:22:18.915 "nvme_io": false, 00:22:18.915 "nvme_io_md": false, 00:22:18.915 "write_zeroes": true, 00:22:18.915 "zcopy": true, 00:22:18.915 "get_zone_info": false, 00:22:18.915 "zone_management": false, 00:22:18.915 "zone_append": false, 00:22:18.915 "compare": false, 00:22:18.915 "compare_and_write": false, 00:22:18.915 "abort": true, 00:22:18.915 "seek_hole": false, 00:22:18.915 "seek_data": false, 00:22:18.915 "copy": true, 00:22:18.915 "nvme_iov_md": false 00:22:18.915 }, 00:22:18.915 "memory_domains": [ 00:22:18.915 { 00:22:18.915 "dma_device_id": "system", 00:22:18.915 "dma_device_type": 1 00:22:18.915 }, 00:22:18.915 { 00:22:18.915 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:18.915 "dma_device_type": 2 00:22:18.915 } 00:22:18.915 ], 00:22:18.915 "driver_specific": { 00:22:18.915 "passthru": { 00:22:18.915 "name": "pt3", 00:22:18.915 "base_bdev_name": "malloc3" 00:22:18.915 } 00:22:18.915 } 00:22:18.915 }' 00:22:18.915 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:18.915 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:18.915 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:22:18.915 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:18.915 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:18.915 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:22:18.915 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:19.172 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:19.172 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:22:19.172 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:19.172 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:19.172 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:22:19.172 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:22:19.172 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt4 00:22:19.172 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:22:19.429 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:22:19.429 "name": "pt4", 00:22:19.429 "aliases": [ 00:22:19.429 "00000000-0000-0000-0000-000000000004" 00:22:19.429 ], 00:22:19.429 "product_name": "passthru", 00:22:19.429 "block_size": 512, 00:22:19.429 "num_blocks": 65536, 00:22:19.429 "uuid": "00000000-0000-0000-0000-000000000004", 00:22:19.429 "assigned_rate_limits": { 00:22:19.429 "rw_ios_per_sec": 0, 00:22:19.429 "rw_mbytes_per_sec": 0, 00:22:19.429 "r_mbytes_per_sec": 0, 00:22:19.429 "w_mbytes_per_sec": 0 00:22:19.429 }, 00:22:19.429 "claimed": true, 00:22:19.429 "claim_type": "exclusive_write", 00:22:19.429 "zoned": false, 00:22:19.429 "supported_io_types": { 00:22:19.429 "read": true, 00:22:19.429 "write": true, 00:22:19.429 "unmap": true, 00:22:19.429 "flush": true, 00:22:19.429 "reset": true, 00:22:19.429 "nvme_admin": false, 00:22:19.429 "nvme_io": false, 00:22:19.429 "nvme_io_md": false, 00:22:19.429 "write_zeroes": true, 00:22:19.429 "zcopy": true, 00:22:19.429 "get_zone_info": false, 00:22:19.429 "zone_management": false, 00:22:19.429 "zone_append": false, 00:22:19.429 "compare": false, 00:22:19.429 "compare_and_write": false, 00:22:19.429 "abort": true, 00:22:19.429 "seek_hole": false, 00:22:19.429 "seek_data": false, 00:22:19.429 "copy": true, 00:22:19.429 "nvme_iov_md": false 00:22:19.429 }, 00:22:19.429 "memory_domains": [ 00:22:19.429 { 00:22:19.429 "dma_device_id": "system", 00:22:19.429 "dma_device_type": 1 00:22:19.429 }, 00:22:19.429 { 00:22:19.429 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:19.429 "dma_device_type": 2 00:22:19.429 } 00:22:19.429 ], 00:22:19.429 "driver_specific": { 00:22:19.429 "passthru": { 00:22:19.429 "name": "pt4", 00:22:19.429 "base_bdev_name": "malloc4" 00:22:19.429 } 00:22:19.429 } 00:22:19.429 }' 00:22:19.429 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:19.429 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:22:19.429 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@205 -- # [[ 512 == 512 ]] 00:22:19.429 11:33:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:19.429 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:22:19.686 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:22:19.686 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:19.686 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:22:19.686 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:22:19.686 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:19.686 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:22:19.686 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:22:19.686 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:22:19.686 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # jq -r '.[] | .uuid' 00:22:19.943 [2024-07-15 11:33:03.454463] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:22:19.943 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@486 -- # '[' e56cdd97-02a3-4569-80cb-249866628e6f '!=' e56cdd97-02a3-4569-80cb-249866628e6f ']' 00:22:19.943 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@490 -- # has_redundancy raid1 00:22:19.943 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:22:19.943 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@214 -- # return 0 00:22:19.943 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@492 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt1 00:22:20.201 [2024-07-15 11:33:03.690805] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:22:20.201 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@495 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:22:20.201 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:22:20.201 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:22:20.201 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:22:20.201 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:22:20.201 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:22:20.201 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:22:20.201 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:22:20.201 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:22:20.201 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:22:20.201 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:20.201 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:20.501 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:22:20.501 "name": "raid_bdev1", 00:22:20.501 "uuid": "e56cdd97-02a3-4569-80cb-249866628e6f", 00:22:20.501 "strip_size_kb": 0, 00:22:20.501 "state": "online", 00:22:20.501 "raid_level": "raid1", 00:22:20.501 "superblock": true, 00:22:20.501 "num_base_bdevs": 4, 00:22:20.501 "num_base_bdevs_discovered": 3, 00:22:20.501 "num_base_bdevs_operational": 3, 00:22:20.501 "base_bdevs_list": [ 00:22:20.501 { 00:22:20.501 "name": null, 00:22:20.501 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:20.501 "is_configured": false, 00:22:20.501 "data_offset": 2048, 00:22:20.501 "data_size": 63488 00:22:20.501 }, 00:22:20.501 { 00:22:20.501 "name": "pt2", 00:22:20.501 "uuid": "00000000-0000-0000-0000-000000000002", 00:22:20.501 "is_configured": true, 00:22:20.501 "data_offset": 2048, 00:22:20.501 "data_size": 63488 00:22:20.501 }, 00:22:20.501 { 00:22:20.501 "name": "pt3", 00:22:20.501 "uuid": "00000000-0000-0000-0000-000000000003", 00:22:20.501 "is_configured": true, 00:22:20.501 "data_offset": 2048, 00:22:20.501 "data_size": 63488 00:22:20.501 }, 00:22:20.501 { 00:22:20.501 "name": "pt4", 00:22:20.501 "uuid": "00000000-0000-0000-0000-000000000004", 00:22:20.501 "is_configured": true, 00:22:20.501 "data_offset": 2048, 00:22:20.501 "data_size": 63488 00:22:20.501 } 00:22:20.501 ] 00:22:20.501 }' 00:22:20.501 11:33:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:22:20.501 11:33:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:21.064 11:33:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@498 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:22:21.321 [2024-07-15 11:33:04.773663] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:22:21.321 [2024-07-15 11:33:04.773694] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:22:21.321 [2024-07-15 11:33:04.773750] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:22:21.321 [2024-07-15 11:33:04.773819] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:22:21.321 [2024-07-15 11:33:04.773831] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x21be780 name raid_bdev1, state offline 00:22:21.321 11:33:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:21.321 11:33:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # jq -r '.[]' 00:22:21.578 11:33:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # raid_bdev= 00:22:21.578 11:33:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # '[' -n '' ']' 00:22:21.578 11:33:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@505 -- # (( i = 1 )) 00:22:21.578 11:33:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@505 -- # (( i < num_base_bdevs )) 00:22:21.578 11:33:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:22:21.578 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@505 -- # (( i++ )) 00:22:21.579 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@505 -- # (( i < num_base_bdevs )) 00:22:21.579 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt3 00:22:21.836 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@505 -- # (( i++ )) 00:22:21.836 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@505 -- # (( i < num_base_bdevs )) 00:22:21.836 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt4 00:22:22.092 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@505 -- # (( i++ )) 00:22:22.092 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@505 -- # (( i < num_base_bdevs )) 00:22:22.092 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@510 -- # (( i = 1 )) 00:22:22.092 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@510 -- # (( i < num_base_bdevs - 1 )) 00:22:22.092 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:22:22.349 [2024-07-15 11:33:05.864483] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:22:22.349 [2024-07-15 11:33:05.864535] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:22.349 [2024-07-15 11:33:05.864557] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2361700 00:22:22.349 [2024-07-15 11:33:05.864570] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:22.349 [2024-07-15 11:33:05.866210] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:22.349 [2024-07-15 11:33:05.866241] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:22:22.349 [2024-07-15 11:33:05.866311] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:22:22.349 [2024-07-15 11:33:05.866340] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:22:22.349 pt2 00:22:22.349 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@514 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:22:22.349 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:22:22.349 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:22:22.349 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:22:22.349 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:22:22.349 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:22:22.349 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:22:22.349 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:22:22.349 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:22:22.349 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:22:22.349 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:22.349 11:33:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:22.606 11:33:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:22:22.606 "name": "raid_bdev1", 00:22:22.606 "uuid": "e56cdd97-02a3-4569-80cb-249866628e6f", 00:22:22.606 "strip_size_kb": 0, 00:22:22.606 "state": "configuring", 00:22:22.606 "raid_level": "raid1", 00:22:22.606 "superblock": true, 00:22:22.606 "num_base_bdevs": 4, 00:22:22.606 "num_base_bdevs_discovered": 1, 00:22:22.606 "num_base_bdevs_operational": 3, 00:22:22.606 "base_bdevs_list": [ 00:22:22.606 { 00:22:22.606 "name": null, 00:22:22.606 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:22.606 "is_configured": false, 00:22:22.606 "data_offset": 2048, 00:22:22.606 "data_size": 63488 00:22:22.606 }, 00:22:22.606 { 00:22:22.606 "name": "pt2", 00:22:22.606 "uuid": "00000000-0000-0000-0000-000000000002", 00:22:22.606 "is_configured": true, 00:22:22.606 "data_offset": 2048, 00:22:22.606 "data_size": 63488 00:22:22.606 }, 00:22:22.606 { 00:22:22.606 "name": null, 00:22:22.606 "uuid": "00000000-0000-0000-0000-000000000003", 00:22:22.606 "is_configured": false, 00:22:22.606 "data_offset": 2048, 00:22:22.606 "data_size": 63488 00:22:22.606 }, 00:22:22.606 { 00:22:22.606 "name": null, 00:22:22.606 "uuid": "00000000-0000-0000-0000-000000000004", 00:22:22.606 "is_configured": false, 00:22:22.606 "data_offset": 2048, 00:22:22.606 "data_size": 63488 00:22:22.606 } 00:22:22.606 ] 00:22:22.606 }' 00:22:22.606 11:33:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:22:22.606 11:33:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:23.170 11:33:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@510 -- # (( i++ )) 00:22:23.170 11:33:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@510 -- # (( i < num_base_bdevs - 1 )) 00:22:23.170 11:33:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:22:23.427 [2024-07-15 11:33:06.895229] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:22:23.427 [2024-07-15 11:33:06.895282] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:23.427 [2024-07-15 11:33:06.895308] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x21c7a10 00:22:23.427 [2024-07-15 11:33:06.895321] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:23.427 [2024-07-15 11:33:06.895672] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:23.427 [2024-07-15 11:33:06.895690] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:22:23.427 [2024-07-15 11:33:06.895756] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:22:23.427 [2024-07-15 11:33:06.895776] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:22:23.427 pt3 00:22:23.427 11:33:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@514 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:22:23.427 11:33:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:22:23.427 11:33:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:22:23.427 11:33:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:22:23.427 11:33:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:22:23.427 11:33:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:22:23.427 11:33:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:22:23.427 11:33:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:22:23.427 11:33:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:22:23.427 11:33:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:22:23.427 11:33:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:23.427 11:33:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:23.684 11:33:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:22:23.684 "name": "raid_bdev1", 00:22:23.684 "uuid": "e56cdd97-02a3-4569-80cb-249866628e6f", 00:22:23.684 "strip_size_kb": 0, 00:22:23.684 "state": "configuring", 00:22:23.684 "raid_level": "raid1", 00:22:23.684 "superblock": true, 00:22:23.684 "num_base_bdevs": 4, 00:22:23.684 "num_base_bdevs_discovered": 2, 00:22:23.684 "num_base_bdevs_operational": 3, 00:22:23.684 "base_bdevs_list": [ 00:22:23.684 { 00:22:23.684 "name": null, 00:22:23.684 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:23.684 "is_configured": false, 00:22:23.684 "data_offset": 2048, 00:22:23.684 "data_size": 63488 00:22:23.684 }, 00:22:23.684 { 00:22:23.684 "name": "pt2", 00:22:23.684 "uuid": "00000000-0000-0000-0000-000000000002", 00:22:23.684 "is_configured": true, 00:22:23.684 "data_offset": 2048, 00:22:23.684 "data_size": 63488 00:22:23.684 }, 00:22:23.684 { 00:22:23.684 "name": "pt3", 00:22:23.684 "uuid": "00000000-0000-0000-0000-000000000003", 00:22:23.684 "is_configured": true, 00:22:23.684 "data_offset": 2048, 00:22:23.684 "data_size": 63488 00:22:23.684 }, 00:22:23.684 { 00:22:23.684 "name": null, 00:22:23.684 "uuid": "00000000-0000-0000-0000-000000000004", 00:22:23.684 "is_configured": false, 00:22:23.684 "data_offset": 2048, 00:22:23.684 "data_size": 63488 00:22:23.684 } 00:22:23.684 ] 00:22:23.684 }' 00:22:23.684 11:33:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:22:23.684 11:33:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:24.248 11:33:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@510 -- # (( i++ )) 00:22:24.249 11:33:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@510 -- # (( i < num_base_bdevs - 1 )) 00:22:24.249 11:33:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@518 -- # i=3 00:22:24.249 11:33:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:22:24.506 [2024-07-15 11:33:07.978098] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:22:24.506 [2024-07-15 11:33:07.978154] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:24.506 [2024-07-15 11:33:07.978175] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x236a520 00:22:24.506 [2024-07-15 11:33:07.978188] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:24.506 [2024-07-15 11:33:07.978544] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:24.506 [2024-07-15 11:33:07.978562] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:22:24.506 [2024-07-15 11:33:07.978627] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:22:24.506 [2024-07-15 11:33:07.978647] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:22:24.506 [2024-07-15 11:33:07.978760] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x21beea0 00:22:24.506 [2024-07-15 11:33:07.978771] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:22:24.506 [2024-07-15 11:33:07.978950] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x21c3600 00:22:24.506 [2024-07-15 11:33:07.979082] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x21beea0 00:22:24.506 [2024-07-15 11:33:07.979092] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x21beea0 00:22:24.506 [2024-07-15 11:33:07.979189] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:24.506 pt4 00:22:24.506 11:33:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@522 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:22:24.506 11:33:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:22:24.506 11:33:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:22:24.506 11:33:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:22:24.506 11:33:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:22:24.506 11:33:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:22:24.506 11:33:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:22:24.506 11:33:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:22:24.506 11:33:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:22:24.506 11:33:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:22:24.506 11:33:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:24.506 11:33:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:24.764 11:33:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:22:24.764 "name": "raid_bdev1", 00:22:24.764 "uuid": "e56cdd97-02a3-4569-80cb-249866628e6f", 00:22:24.764 "strip_size_kb": 0, 00:22:24.764 "state": "online", 00:22:24.764 "raid_level": "raid1", 00:22:24.764 "superblock": true, 00:22:24.764 "num_base_bdevs": 4, 00:22:24.764 "num_base_bdevs_discovered": 3, 00:22:24.764 "num_base_bdevs_operational": 3, 00:22:24.764 "base_bdevs_list": [ 00:22:24.764 { 00:22:24.764 "name": null, 00:22:24.764 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:24.764 "is_configured": false, 00:22:24.764 "data_offset": 2048, 00:22:24.764 "data_size": 63488 00:22:24.764 }, 00:22:24.764 { 00:22:24.764 "name": "pt2", 00:22:24.764 "uuid": "00000000-0000-0000-0000-000000000002", 00:22:24.764 "is_configured": true, 00:22:24.764 "data_offset": 2048, 00:22:24.764 "data_size": 63488 00:22:24.764 }, 00:22:24.764 { 00:22:24.764 "name": "pt3", 00:22:24.764 "uuid": "00000000-0000-0000-0000-000000000003", 00:22:24.764 "is_configured": true, 00:22:24.764 "data_offset": 2048, 00:22:24.764 "data_size": 63488 00:22:24.764 }, 00:22:24.764 { 00:22:24.764 "name": "pt4", 00:22:24.764 "uuid": "00000000-0000-0000-0000-000000000004", 00:22:24.764 "is_configured": true, 00:22:24.764 "data_offset": 2048, 00:22:24.764 "data_size": 63488 00:22:24.764 } 00:22:24.764 ] 00:22:24.764 }' 00:22:24.764 11:33:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:22:24.764 11:33:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:25.328 11:33:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@525 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:22:25.585 [2024-07-15 11:33:09.040882] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:22:25.585 [2024-07-15 11:33:09.040910] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:22:25.585 [2024-07-15 11:33:09.040971] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:22:25.585 [2024-07-15 11:33:09.041040] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:22:25.585 [2024-07-15 11:33:09.041052] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x21beea0 name raid_bdev1, state offline 00:22:25.585 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:25.585 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # jq -r '.[]' 00:22:25.842 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # raid_bdev= 00:22:25.842 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # '[' -n '' ']' 00:22:25.842 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@531 -- # '[' 4 -gt 2 ']' 00:22:25.842 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@533 -- # i=3 00:22:25.842 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt4 00:22:26.100 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@539 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:22:26.359 [2024-07-15 11:33:09.786831] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:22:26.359 [2024-07-15 11:33:09.786877] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:26.359 [2024-07-15 11:33:09.786895] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x236a520 00:22:26.359 [2024-07-15 11:33:09.786907] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:26.359 [2024-07-15 11:33:09.788530] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:26.359 [2024-07-15 11:33:09.788560] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:22:26.359 [2024-07-15 11:33:09.788629] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:22:26.359 [2024-07-15 11:33:09.788657] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:22:26.359 [2024-07-15 11:33:09.788761] bdev_raid.c:3547:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:22:26.359 [2024-07-15 11:33:09.788774] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:22:26.359 [2024-07-15 11:33:09.788789] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x21be060 name raid_bdev1, state configuring 00:22:26.359 [2024-07-15 11:33:09.788812] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:22:26.359 [2024-07-15 11:33:09.788887] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:22:26.359 pt1 00:22:26.359 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@541 -- # '[' 4 -gt 2 ']' 00:22:26.359 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@544 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:22:26.359 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:22:26.359 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:22:26.359 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:22:26.359 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:22:26.359 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:22:26.359 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:22:26.359 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:22:26.359 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:22:26.359 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:22:26.359 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:26.359 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:26.617 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:22:26.617 "name": "raid_bdev1", 00:22:26.617 "uuid": "e56cdd97-02a3-4569-80cb-249866628e6f", 00:22:26.617 "strip_size_kb": 0, 00:22:26.617 "state": "configuring", 00:22:26.617 "raid_level": "raid1", 00:22:26.617 "superblock": true, 00:22:26.617 "num_base_bdevs": 4, 00:22:26.617 "num_base_bdevs_discovered": 2, 00:22:26.617 "num_base_bdevs_operational": 3, 00:22:26.617 "base_bdevs_list": [ 00:22:26.617 { 00:22:26.617 "name": null, 00:22:26.617 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:26.617 "is_configured": false, 00:22:26.617 "data_offset": 2048, 00:22:26.617 "data_size": 63488 00:22:26.617 }, 00:22:26.617 { 00:22:26.617 "name": "pt2", 00:22:26.617 "uuid": "00000000-0000-0000-0000-000000000002", 00:22:26.617 "is_configured": true, 00:22:26.617 "data_offset": 2048, 00:22:26.617 "data_size": 63488 00:22:26.617 }, 00:22:26.617 { 00:22:26.617 "name": "pt3", 00:22:26.617 "uuid": "00000000-0000-0000-0000-000000000003", 00:22:26.617 "is_configured": true, 00:22:26.617 "data_offset": 2048, 00:22:26.617 "data_size": 63488 00:22:26.617 }, 00:22:26.617 { 00:22:26.617 "name": null, 00:22:26.617 "uuid": "00000000-0000-0000-0000-000000000004", 00:22:26.617 "is_configured": false, 00:22:26.617 "data_offset": 2048, 00:22:26.617 "data_size": 63488 00:22:26.617 } 00:22:26.617 ] 00:22:26.618 }' 00:22:26.618 11:33:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:22:26.618 11:33:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:27.183 11:33:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs configuring 00:22:27.184 11:33:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:22:27.184 11:33:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # [[ false == \f\a\l\s\e ]] 00:22:27.184 11:33:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@548 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:22:27.442 [2024-07-15 11:33:10.978003] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:22:27.442 [2024-07-15 11:33:10.978060] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:27.442 [2024-07-15 11:33:10.978081] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x21be310 00:22:27.442 [2024-07-15 11:33:10.978094] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:27.442 [2024-07-15 11:33:10.978462] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:27.442 [2024-07-15 11:33:10.978481] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:22:27.442 [2024-07-15 11:33:10.978551] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:22:27.442 [2024-07-15 11:33:10.978573] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:22:27.442 [2024-07-15 11:33:10.978691] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x21c1b40 00:22:27.442 [2024-07-15 11:33:10.978701] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:22:27.442 [2024-07-15 11:33:10.978878] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x2361990 00:22:27.442 [2024-07-15 11:33:10.979022] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x21c1b40 00:22:27.442 [2024-07-15 11:33:10.979033] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x21c1b40 00:22:27.442 [2024-07-15 11:33:10.979132] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:27.442 pt4 00:22:27.442 11:33:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@553 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:22:27.442 11:33:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:22:27.442 11:33:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:22:27.442 11:33:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:22:27.442 11:33:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:22:27.442 11:33:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:22:27.442 11:33:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:22:27.442 11:33:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:22:27.442 11:33:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:22:27.442 11:33:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:22:27.442 11:33:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:27.442 11:33:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:27.700 11:33:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:22:27.700 "name": "raid_bdev1", 00:22:27.700 "uuid": "e56cdd97-02a3-4569-80cb-249866628e6f", 00:22:27.700 "strip_size_kb": 0, 00:22:27.700 "state": "online", 00:22:27.700 "raid_level": "raid1", 00:22:27.700 "superblock": true, 00:22:27.700 "num_base_bdevs": 4, 00:22:27.700 "num_base_bdevs_discovered": 3, 00:22:27.700 "num_base_bdevs_operational": 3, 00:22:27.700 "base_bdevs_list": [ 00:22:27.700 { 00:22:27.700 "name": null, 00:22:27.700 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:27.700 "is_configured": false, 00:22:27.700 "data_offset": 2048, 00:22:27.700 "data_size": 63488 00:22:27.700 }, 00:22:27.700 { 00:22:27.700 "name": "pt2", 00:22:27.700 "uuid": "00000000-0000-0000-0000-000000000002", 00:22:27.700 "is_configured": true, 00:22:27.700 "data_offset": 2048, 00:22:27.700 "data_size": 63488 00:22:27.700 }, 00:22:27.700 { 00:22:27.700 "name": "pt3", 00:22:27.700 "uuid": "00000000-0000-0000-0000-000000000003", 00:22:27.700 "is_configured": true, 00:22:27.700 "data_offset": 2048, 00:22:27.700 "data_size": 63488 00:22:27.700 }, 00:22:27.700 { 00:22:27.700 "name": "pt4", 00:22:27.700 "uuid": "00000000-0000-0000-0000-000000000004", 00:22:27.700 "is_configured": true, 00:22:27.700 "data_offset": 2048, 00:22:27.700 "data_size": 63488 00:22:27.700 } 00:22:27.700 ] 00:22:27.700 }' 00:22:27.700 11:33:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:22:27.700 11:33:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:28.265 11:33:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs online 00:22:28.265 11:33:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:22:28.522 11:33:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # [[ false == \f\a\l\s\e ]] 00:22:28.522 11:33:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@557 -- # jq -r '.[] | .uuid' 00:22:28.522 11:33:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@557 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:22:28.780 [2024-07-15 11:33:12.317832] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:22:28.780 11:33:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@557 -- # '[' e56cdd97-02a3-4569-80cb-249866628e6f '!=' e56cdd97-02a3-4569-80cb-249866628e6f ']' 00:22:28.780 11:33:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@562 -- # killprocess 961900 00:22:28.780 11:33:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@948 -- # '[' -z 961900 ']' 00:22:28.780 11:33:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@952 -- # kill -0 961900 00:22:28.780 11:33:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@953 -- # uname 00:22:28.780 11:33:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:22:28.780 11:33:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 961900 00:22:29.038 11:33:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:22:29.038 11:33:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:22:29.038 11:33:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 961900' 00:22:29.038 killing process with pid 961900 00:22:29.038 11:33:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@967 -- # kill 961900 00:22:29.038 [2024-07-15 11:33:12.375564] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:22:29.038 [2024-07-15 11:33:12.375621] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:22:29.039 [2024-07-15 11:33:12.375689] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:22:29.039 [2024-07-15 11:33:12.375702] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x21c1b40 name raid_bdev1, state offline 00:22:29.039 11:33:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # wait 961900 00:22:29.039 [2024-07-15 11:33:12.418007] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:22:29.297 11:33:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@564 -- # return 0 00:22:29.297 00:22:29.297 real 0m24.733s 00:22:29.297 user 0m45.348s 00:22:29.297 sys 0m4.418s 00:22:29.297 11:33:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:22:29.297 11:33:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:29.297 ************************************ 00:22:29.297 END TEST raid_superblock_test 00:22:29.297 ************************************ 00:22:29.297 11:33:12 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:22:29.297 11:33:12 bdev_raid -- bdev/bdev_raid.sh@870 -- # run_test raid_read_error_test raid_io_error_test raid1 4 read 00:22:29.297 11:33:12 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:22:29.297 11:33:12 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:22:29.297 11:33:12 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:22:29.297 ************************************ 00:22:29.297 START TEST raid_read_error_test 00:22:29.297 ************************************ 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1123 -- # raid_io_error_test raid1 4 read 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@788 -- # local raid_level=raid1 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@789 -- # local num_base_bdevs=4 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local error_io_type=read 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i = 1 )) 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev1 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev2 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev3 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev4 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local base_bdevs 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local raid_bdev_name=raid_bdev1 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local strip_size 00:22:29.297 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local create_arg 00:22:29.298 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local bdevperf_log 00:22:29.298 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local fail_per_s 00:22:29.298 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # '[' raid1 '!=' raid1 ']' 00:22:29.298 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # strip_size=0 00:22:29.298 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@805 -- # mktemp -p /raidtest 00:22:29.298 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@805 -- # bdevperf_log=/raidtest/tmp.u89GAj1YwY 00:22:29.298 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@808 -- # raid_pid=966088 00:22:29.298 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # waitforlisten 966088 /var/tmp/spdk-raid.sock 00:22:29.298 11:33:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:22:29.298 11:33:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@829 -- # '[' -z 966088 ']' 00:22:29.298 11:33:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:22:29.298 11:33:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:22:29.298 11:33:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:22:29.298 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:22:29.298 11:33:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:22:29.298 11:33:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:29.298 [2024-07-15 11:33:12.806801] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:22:29.298 [2024-07-15 11:33:12.806870] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid966088 ] 00:22:29.556 [2024-07-15 11:33:12.937480] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:22:29.556 [2024-07-15 11:33:13.043430] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:22:29.556 [2024-07-15 11:33:13.110365] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:22:29.556 [2024-07-15 11:33:13.110408] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:22:30.492 11:33:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:22:30.492 11:33:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@862 -- # return 0 00:22:30.492 11:33:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:22:30.492 11:33:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:22:30.492 BaseBdev1_malloc 00:22:30.492 11:33:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev1_malloc 00:22:30.750 true 00:22:30.750 11:33:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:22:31.009 [2024-07-15 11:33:14.429359] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:22:31.009 [2024-07-15 11:33:14.429404] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:31.009 [2024-07-15 11:33:14.429425] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xe9b0d0 00:22:31.009 [2024-07-15 11:33:14.429437] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:31.009 [2024-07-15 11:33:14.431308] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:31.009 [2024-07-15 11:33:14.431338] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:22:31.009 BaseBdev1 00:22:31.009 11:33:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:22:31.009 11:33:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:22:31.267 BaseBdev2_malloc 00:22:31.267 11:33:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev2_malloc 00:22:31.525 true 00:22:31.525 11:33:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:22:31.784 [2024-07-15 11:33:15.161176] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:22:31.784 [2024-07-15 11:33:15.161220] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:31.784 [2024-07-15 11:33:15.161241] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xe9f910 00:22:31.784 [2024-07-15 11:33:15.161253] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:31.784 [2024-07-15 11:33:15.162819] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:31.784 [2024-07-15 11:33:15.162847] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:22:31.784 BaseBdev2 00:22:31.784 11:33:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:22:31.784 11:33:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:22:32.041 BaseBdev3_malloc 00:22:32.041 11:33:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev3_malloc 00:22:32.298 true 00:22:32.298 11:33:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:22:32.556 [2024-07-15 11:33:15.903987] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:22:32.556 [2024-07-15 11:33:15.904028] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:32.556 [2024-07-15 11:33:15.904050] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xea1bd0 00:22:32.556 [2024-07-15 11:33:15.904063] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:32.556 [2024-07-15 11:33:15.905646] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:32.556 [2024-07-15 11:33:15.905673] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:22:32.556 BaseBdev3 00:22:32.556 11:33:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:22:32.556 11:33:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:22:32.813 BaseBdev4_malloc 00:22:32.813 11:33:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev4_malloc 00:22:32.813 true 00:22:33.070 11:33:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:22:33.070 [2024-07-15 11:33:16.638474] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:22:33.070 [2024-07-15 11:33:16.638518] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:33.070 [2024-07-15 11:33:16.638540] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xea2aa0 00:22:33.070 [2024-07-15 11:33:16.638552] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:33.070 [2024-07-15 11:33:16.640162] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:33.070 [2024-07-15 11:33:16.640191] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:22:33.070 BaseBdev4 00:22:33.070 11:33:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@819 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n raid_bdev1 -s 00:22:33.327 [2024-07-15 11:33:16.883151] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:22:33.327 [2024-07-15 11:33:16.884522] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:22:33.327 [2024-07-15 11:33:16.884592] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:22:33.327 [2024-07-15 11:33:16.884653] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:22:33.327 [2024-07-15 11:33:16.884892] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xe9cc20 00:22:33.327 [2024-07-15 11:33:16.884903] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:22:33.327 [2024-07-15 11:33:16.885109] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xcf1260 00:22:33.327 [2024-07-15 11:33:16.885271] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xe9cc20 00:22:33.327 [2024-07-15 11:33:16.885281] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0xe9cc20 00:22:33.327 [2024-07-15 11:33:16.885389] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:33.327 11:33:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@820 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:22:33.327 11:33:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:22:33.327 11:33:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:22:33.327 11:33:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:22:33.327 11:33:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:22:33.327 11:33:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:22:33.327 11:33:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:22:33.327 11:33:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:22:33.327 11:33:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:22:33.327 11:33:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:22:33.327 11:33:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:33.327 11:33:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:33.584 11:33:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:22:33.584 "name": "raid_bdev1", 00:22:33.584 "uuid": "bebbf07d-7bf6-4a70-a914-0d369951f7c1", 00:22:33.584 "strip_size_kb": 0, 00:22:33.584 "state": "online", 00:22:33.584 "raid_level": "raid1", 00:22:33.584 "superblock": true, 00:22:33.584 "num_base_bdevs": 4, 00:22:33.584 "num_base_bdevs_discovered": 4, 00:22:33.584 "num_base_bdevs_operational": 4, 00:22:33.584 "base_bdevs_list": [ 00:22:33.584 { 00:22:33.584 "name": "BaseBdev1", 00:22:33.584 "uuid": "8be76aeb-68c5-5f5c-9d9a-80fd6aaec322", 00:22:33.584 "is_configured": true, 00:22:33.584 "data_offset": 2048, 00:22:33.584 "data_size": 63488 00:22:33.584 }, 00:22:33.584 { 00:22:33.585 "name": "BaseBdev2", 00:22:33.585 "uuid": "16b18648-e511-568f-84b1-982905886bc9", 00:22:33.585 "is_configured": true, 00:22:33.585 "data_offset": 2048, 00:22:33.585 "data_size": 63488 00:22:33.585 }, 00:22:33.585 { 00:22:33.585 "name": "BaseBdev3", 00:22:33.585 "uuid": "454b12c4-bb65-5e53-afd1-4a3bd1847f51", 00:22:33.585 "is_configured": true, 00:22:33.585 "data_offset": 2048, 00:22:33.585 "data_size": 63488 00:22:33.585 }, 00:22:33.585 { 00:22:33.585 "name": "BaseBdev4", 00:22:33.585 "uuid": "f3d3f32e-27eb-53f6-8679-8d0705124282", 00:22:33.585 "is_configured": true, 00:22:33.585 "data_offset": 2048, 00:22:33.585 "data_size": 63488 00:22:33.585 } 00:22:33.585 ] 00:22:33.585 }' 00:22:33.585 11:33:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:22:33.585 11:33:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:34.514 11:33:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@824 -- # sleep 1 00:22:34.514 11:33:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@823 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:22:34.514 [2024-07-15 11:33:17.853993] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xcf0c60 00:22:35.445 11:33:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@827 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:22:35.445 11:33:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # local expected_num_base_bdevs 00:22:35.445 11:33:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@830 -- # [[ raid1 = \r\a\i\d\1 ]] 00:22:35.445 11:33:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@830 -- # [[ read = \w\r\i\t\e ]] 00:22:35.445 11:33:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=4 00:22:35.445 11:33:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:22:35.445 11:33:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:22:35.445 11:33:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:22:35.445 11:33:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:22:35.445 11:33:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:22:35.445 11:33:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:22:35.445 11:33:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:22:35.445 11:33:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:22:35.445 11:33:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:22:35.445 11:33:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:22:35.445 11:33:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:35.445 11:33:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:35.703 11:33:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:22:35.703 "name": "raid_bdev1", 00:22:35.703 "uuid": "bebbf07d-7bf6-4a70-a914-0d369951f7c1", 00:22:35.703 "strip_size_kb": 0, 00:22:35.703 "state": "online", 00:22:35.703 "raid_level": "raid1", 00:22:35.703 "superblock": true, 00:22:35.703 "num_base_bdevs": 4, 00:22:35.703 "num_base_bdevs_discovered": 4, 00:22:35.703 "num_base_bdevs_operational": 4, 00:22:35.703 "base_bdevs_list": [ 00:22:35.703 { 00:22:35.703 "name": "BaseBdev1", 00:22:35.703 "uuid": "8be76aeb-68c5-5f5c-9d9a-80fd6aaec322", 00:22:35.703 "is_configured": true, 00:22:35.703 "data_offset": 2048, 00:22:35.703 "data_size": 63488 00:22:35.703 }, 00:22:35.703 { 00:22:35.703 "name": "BaseBdev2", 00:22:35.703 "uuid": "16b18648-e511-568f-84b1-982905886bc9", 00:22:35.703 "is_configured": true, 00:22:35.703 "data_offset": 2048, 00:22:35.703 "data_size": 63488 00:22:35.703 }, 00:22:35.703 { 00:22:35.704 "name": "BaseBdev3", 00:22:35.704 "uuid": "454b12c4-bb65-5e53-afd1-4a3bd1847f51", 00:22:35.704 "is_configured": true, 00:22:35.704 "data_offset": 2048, 00:22:35.704 "data_size": 63488 00:22:35.704 }, 00:22:35.704 { 00:22:35.704 "name": "BaseBdev4", 00:22:35.704 "uuid": "f3d3f32e-27eb-53f6-8679-8d0705124282", 00:22:35.704 "is_configured": true, 00:22:35.704 "data_offset": 2048, 00:22:35.704 "data_size": 63488 00:22:35.704 } 00:22:35.704 ] 00:22:35.704 }' 00:22:35.704 11:33:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:22:35.704 11:33:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:36.267 11:33:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:22:36.524 [2024-07-15 11:33:19.993211] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:22:36.524 [2024-07-15 11:33:19.993254] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:22:36.524 [2024-07-15 11:33:19.996524] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:22:36.524 [2024-07-15 11:33:19.996564] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:36.524 [2024-07-15 11:33:19.996683] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:22:36.524 [2024-07-15 11:33:19.996702] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xe9cc20 name raid_bdev1, state offline 00:22:36.524 0 00:22:36.524 11:33:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # killprocess 966088 00:22:36.524 11:33:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@948 -- # '[' -z 966088 ']' 00:22:36.524 11:33:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@952 -- # kill -0 966088 00:22:36.524 11:33:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@953 -- # uname 00:22:36.524 11:33:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:22:36.524 11:33:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 966088 00:22:36.524 11:33:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:22:36.524 11:33:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:22:36.524 11:33:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 966088' 00:22:36.524 killing process with pid 966088 00:22:36.524 11:33:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@967 -- # kill 966088 00:22:36.524 [2024-07-15 11:33:20.065876] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:22:36.524 11:33:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # wait 966088 00:22:36.524 [2024-07-15 11:33:20.096318] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:22:36.781 11:33:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # grep -v Job /raidtest/tmp.u89GAj1YwY 00:22:36.781 11:33:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # grep raid_bdev1 00:22:36.781 11:33:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # awk '{print $6}' 00:22:36.781 11:33:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@843 -- # fail_per_s=0.00 00:22:36.781 11:33:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@844 -- # has_redundancy raid1 00:22:36.781 11:33:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:22:36.781 11:33:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@214 -- # return 0 00:22:36.781 11:33:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # [[ 0.00 = \0\.\0\0 ]] 00:22:36.781 00:22:36.781 real 0m7.590s 00:22:36.781 user 0m12.056s 00:22:36.781 sys 0m1.415s 00:22:36.781 11:33:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:22:36.781 11:33:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:36.781 ************************************ 00:22:36.781 END TEST raid_read_error_test 00:22:36.781 ************************************ 00:22:36.781 11:33:20 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:22:36.782 11:33:20 bdev_raid -- bdev/bdev_raid.sh@871 -- # run_test raid_write_error_test raid_io_error_test raid1 4 write 00:22:36.782 11:33:20 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:22:36.782 11:33:20 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:22:36.782 11:33:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:22:37.040 ************************************ 00:22:37.040 START TEST raid_write_error_test 00:22:37.040 ************************************ 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1123 -- # raid_io_error_test raid1 4 write 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@788 -- # local raid_level=raid1 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@789 -- # local num_base_bdevs=4 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local error_io_type=write 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i = 1 )) 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev1 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev2 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev3 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # echo BaseBdev4 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i++ )) 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # (( i <= num_base_bdevs )) 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local base_bdevs 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local raid_bdev_name=raid_bdev1 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local strip_size 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local create_arg 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local bdevperf_log 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local fail_per_s 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # '[' raid1 '!=' raid1 ']' 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # strip_size=0 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@805 -- # mktemp -p /raidtest 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@805 -- # bdevperf_log=/raidtest/tmp.1CtYjjHcIy 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@808 -- # raid_pid=967229 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # waitforlisten 967229 /var/tmp/spdk-raid.sock 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@829 -- # '[' -z 967229 ']' 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:22:37.040 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:22:37.040 11:33:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:37.040 [2024-07-15 11:33:20.460225] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:22:37.040 [2024-07-15 11:33:20.460293] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid967229 ] 00:22:37.040 [2024-07-15 11:33:20.601078] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:22:37.352 [2024-07-15 11:33:20.739846] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:22:37.352 [2024-07-15 11:33:20.801187] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:22:37.352 [2024-07-15 11:33:20.801219] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:22:37.942 11:33:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:22:37.942 11:33:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@862 -- # return 0 00:22:37.942 11:33:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:22:37.942 11:33:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:22:38.200 BaseBdev1_malloc 00:22:38.200 11:33:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev1_malloc 00:22:38.458 true 00:22:38.458 11:33:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:22:38.716 [2024-07-15 11:33:22.178473] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:22:38.716 [2024-07-15 11:33:22.178516] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:38.716 [2024-07-15 11:33:22.178536] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1bcf0d0 00:22:38.716 [2024-07-15 11:33:22.178549] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:38.716 [2024-07-15 11:33:22.180366] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:38.716 [2024-07-15 11:33:22.180396] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:22:38.716 BaseBdev1 00:22:38.716 11:33:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:22:38.716 11:33:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:22:38.973 BaseBdev2_malloc 00:22:38.973 11:33:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev2_malloc 00:22:39.230 true 00:22:39.230 11:33:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:22:39.488 [2024-07-15 11:33:22.921534] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:22:39.488 [2024-07-15 11:33:22.921578] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:39.488 [2024-07-15 11:33:22.921599] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1bd3910 00:22:39.488 [2024-07-15 11:33:22.921611] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:39.488 [2024-07-15 11:33:22.923206] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:39.488 [2024-07-15 11:33:22.923232] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:22:39.488 BaseBdev2 00:22:39.488 11:33:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:22:39.488 11:33:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:22:39.758 BaseBdev3_malloc 00:22:39.758 11:33:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev3_malloc 00:22:40.016 true 00:22:40.016 11:33:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:22:40.273 [2024-07-15 11:33:23.661304] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:22:40.273 [2024-07-15 11:33:23.661351] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:40.273 [2024-07-15 11:33:23.661371] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1bd5bd0 00:22:40.273 [2024-07-15 11:33:23.661384] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:40.273 [2024-07-15 11:33:23.662977] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:40.273 [2024-07-15 11:33:23.663005] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:22:40.273 BaseBdev3 00:22:40.273 11:33:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@812 -- # for bdev in "${base_bdevs[@]}" 00:22:40.273 11:33:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@813 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:22:40.530 BaseBdev4_malloc 00:22:40.530 11:33:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_create BaseBdev4_malloc 00:22:40.787 true 00:22:40.787 11:33:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:22:40.787 [2024-07-15 11:33:24.379764] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:22:40.787 [2024-07-15 11:33:24.379808] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:40.787 [2024-07-15 11:33:24.379829] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1bd6aa0 00:22:40.787 [2024-07-15 11:33:24.379841] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:40.787 [2024-07-15 11:33:24.381396] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:40.787 [2024-07-15 11:33:24.381423] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:22:41.045 BaseBdev4 00:22:41.045 11:33:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@819 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n raid_bdev1 -s 00:22:41.045 [2024-07-15 11:33:24.624453] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:22:41.045 [2024-07-15 11:33:24.625829] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:22:41.045 [2024-07-15 11:33:24.625898] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:22:41.045 [2024-07-15 11:33:24.625966] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:22:41.045 [2024-07-15 11:33:24.626204] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1bd0c20 00:22:41.045 [2024-07-15 11:33:24.626215] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:22:41.045 [2024-07-15 11:33:24.626415] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1a25260 00:22:41.045 [2024-07-15 11:33:24.626574] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1bd0c20 00:22:41.045 [2024-07-15 11:33:24.626584] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x1bd0c20 00:22:41.045 [2024-07-15 11:33:24.626696] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:41.304 11:33:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@820 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:22:41.304 11:33:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:22:41.304 11:33:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:22:41.304 11:33:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:22:41.304 11:33:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:22:41.304 11:33:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:22:41.304 11:33:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:22:41.304 11:33:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:22:41.304 11:33:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:22:41.304 11:33:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:22:41.304 11:33:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:41.304 11:33:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:41.304 11:33:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:22:41.304 "name": "raid_bdev1", 00:22:41.304 "uuid": "223893fa-0bee-4c01-8ef3-335dbd8f54c9", 00:22:41.304 "strip_size_kb": 0, 00:22:41.304 "state": "online", 00:22:41.304 "raid_level": "raid1", 00:22:41.304 "superblock": true, 00:22:41.304 "num_base_bdevs": 4, 00:22:41.304 "num_base_bdevs_discovered": 4, 00:22:41.304 "num_base_bdevs_operational": 4, 00:22:41.304 "base_bdevs_list": [ 00:22:41.304 { 00:22:41.304 "name": "BaseBdev1", 00:22:41.304 "uuid": "5e113216-13da-5a81-9231-ce7cdc79d8c8", 00:22:41.304 "is_configured": true, 00:22:41.304 "data_offset": 2048, 00:22:41.304 "data_size": 63488 00:22:41.304 }, 00:22:41.304 { 00:22:41.304 "name": "BaseBdev2", 00:22:41.304 "uuid": "5931895c-b697-55c7-8f1b-7bf4ed5bf4f8", 00:22:41.304 "is_configured": true, 00:22:41.304 "data_offset": 2048, 00:22:41.304 "data_size": 63488 00:22:41.304 }, 00:22:41.304 { 00:22:41.304 "name": "BaseBdev3", 00:22:41.304 "uuid": "59fa5da0-c548-5cbf-bd36-4e5c5c4bdb19", 00:22:41.304 "is_configured": true, 00:22:41.304 "data_offset": 2048, 00:22:41.304 "data_size": 63488 00:22:41.304 }, 00:22:41.304 { 00:22:41.304 "name": "BaseBdev4", 00:22:41.304 "uuid": "c50dd6e5-15d7-513c-bd45-697c3477b393", 00:22:41.304 "is_configured": true, 00:22:41.304 "data_offset": 2048, 00:22:41.304 "data_size": 63488 00:22:41.304 } 00:22:41.304 ] 00:22:41.304 }' 00:22:41.304 11:33:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:22:41.304 11:33:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:41.871 11:33:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@824 -- # sleep 1 00:22:41.871 11:33:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@823 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:22:42.129 [2024-07-15 11:33:25.519105] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1a24c60 00:22:43.065 11:33:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@827 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:22:43.065 [2024-07-15 11:33:26.640495] bdev_raid.c:2221:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:22:43.065 [2024-07-15 11:33:26.640555] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:22:43.065 [2024-07-15 11:33:26.640770] bdev_raid.c:1919:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x1a24c60 00:22:43.323 11:33:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # local expected_num_base_bdevs 00:22:43.323 11:33:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@830 -- # [[ raid1 = \r\a\i\d\1 ]] 00:22:43.323 11:33:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@830 -- # [[ write = \w\r\i\t\e ]] 00:22:43.323 11:33:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # expected_num_base_bdevs=3 00:22:43.323 11:33:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:22:43.323 11:33:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:22:43.323 11:33:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:22:43.323 11:33:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:22:43.323 11:33:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:22:43.323 11:33:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:22:43.323 11:33:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:22:43.323 11:33:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:22:43.323 11:33:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:22:43.323 11:33:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:22:43.323 11:33:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:43.323 11:33:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:43.323 11:33:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:22:43.323 "name": "raid_bdev1", 00:22:43.323 "uuid": "223893fa-0bee-4c01-8ef3-335dbd8f54c9", 00:22:43.323 "strip_size_kb": 0, 00:22:43.323 "state": "online", 00:22:43.323 "raid_level": "raid1", 00:22:43.323 "superblock": true, 00:22:43.323 "num_base_bdevs": 4, 00:22:43.323 "num_base_bdevs_discovered": 3, 00:22:43.323 "num_base_bdevs_operational": 3, 00:22:43.323 "base_bdevs_list": [ 00:22:43.323 { 00:22:43.323 "name": null, 00:22:43.323 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:43.323 "is_configured": false, 00:22:43.323 "data_offset": 2048, 00:22:43.323 "data_size": 63488 00:22:43.323 }, 00:22:43.323 { 00:22:43.323 "name": "BaseBdev2", 00:22:43.323 "uuid": "5931895c-b697-55c7-8f1b-7bf4ed5bf4f8", 00:22:43.323 "is_configured": true, 00:22:43.323 "data_offset": 2048, 00:22:43.323 "data_size": 63488 00:22:43.323 }, 00:22:43.323 { 00:22:43.323 "name": "BaseBdev3", 00:22:43.323 "uuid": "59fa5da0-c548-5cbf-bd36-4e5c5c4bdb19", 00:22:43.323 "is_configured": true, 00:22:43.323 "data_offset": 2048, 00:22:43.323 "data_size": 63488 00:22:43.323 }, 00:22:43.323 { 00:22:43.323 "name": "BaseBdev4", 00:22:43.323 "uuid": "c50dd6e5-15d7-513c-bd45-697c3477b393", 00:22:43.323 "is_configured": true, 00:22:43.323 "data_offset": 2048, 00:22:43.323 "data_size": 63488 00:22:43.323 } 00:22:43.323 ] 00:22:43.323 }' 00:22:43.323 11:33:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:22:43.324 11:33:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:44.259 11:33:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:22:44.259 [2024-07-15 11:33:27.751893] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:22:44.259 [2024-07-15 11:33:27.751939] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:22:44.259 [2024-07-15 11:33:27.755122] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:22:44.259 [2024-07-15 11:33:27.755155] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:44.259 [2024-07-15 11:33:27.755251] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:22:44.259 [2024-07-15 11:33:27.755263] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1bd0c20 name raid_bdev1, state offline 00:22:44.259 0 00:22:44.259 11:33:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # killprocess 967229 00:22:44.259 11:33:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@948 -- # '[' -z 967229 ']' 00:22:44.259 11:33:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@952 -- # kill -0 967229 00:22:44.259 11:33:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@953 -- # uname 00:22:44.259 11:33:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:22:44.259 11:33:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 967229 00:22:44.259 11:33:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:22:44.259 11:33:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:22:44.259 11:33:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 967229' 00:22:44.259 killing process with pid 967229 00:22:44.259 11:33:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@967 -- # kill 967229 00:22:44.259 [2024-07-15 11:33:27.820621] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:22:44.259 11:33:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # wait 967229 00:22:44.259 [2024-07-15 11:33:27.852557] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:22:44.517 11:33:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # grep -v Job /raidtest/tmp.1CtYjjHcIy 00:22:44.517 11:33:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # grep raid_bdev1 00:22:44.518 11:33:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # awk '{print $6}' 00:22:44.518 11:33:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@843 -- # fail_per_s=0.00 00:22:44.518 11:33:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@844 -- # has_redundancy raid1 00:22:44.518 11:33:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@213 -- # case $1 in 00:22:44.518 11:33:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@214 -- # return 0 00:22:44.518 11:33:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # [[ 0.00 = \0\.\0\0 ]] 00:22:44.518 00:22:44.518 real 0m7.712s 00:22:44.518 user 0m12.334s 00:22:44.518 sys 0m1.374s 00:22:44.518 11:33:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:22:44.518 11:33:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:44.518 ************************************ 00:22:44.518 END TEST raid_write_error_test 00:22:44.518 ************************************ 00:22:44.775 11:33:28 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:22:44.775 11:33:28 bdev_raid -- bdev/bdev_raid.sh@875 -- # '[' true = true ']' 00:22:44.775 11:33:28 bdev_raid -- bdev/bdev_raid.sh@876 -- # for n in 2 4 00:22:44.775 11:33:28 bdev_raid -- bdev/bdev_raid.sh@877 -- # run_test raid_rebuild_test raid_rebuild_test raid1 2 false false true 00:22:44.775 11:33:28 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:22:44.775 11:33:28 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:22:44.775 11:33:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:22:44.775 ************************************ 00:22:44.775 START TEST raid_rebuild_test 00:22:44.776 ************************************ 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1123 -- # raid_rebuild_test raid1 2 false false true 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@568 -- # local raid_level=raid1 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local num_base_bdevs=2 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local superblock=false 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local background_io=false 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local verify=true 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # (( i = 1 )) 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # echo BaseBdev1 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # echo BaseBdev2 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local base_bdevs 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local raid_bdev_name=raid_bdev1 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local strip_size 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local create_arg 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local raid_bdev_size 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local data_offset 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@580 -- # '[' raid1 '!=' raid1 ']' 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@588 -- # strip_size=0 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@591 -- # '[' false = true ']' 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # raid_pid=968376 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # waitforlisten 968376 /var/tmp/spdk-raid.sock 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@595 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@829 -- # '[' -z 968376 ']' 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:22:44.776 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:22:44.776 11:33:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:22:44.776 [2024-07-15 11:33:28.255573] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:22:44.776 [2024-07-15 11:33:28.255646] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid968376 ] 00:22:44.776 I/O size of 3145728 is greater than zero copy threshold (65536). 00:22:44.776 Zero copy mechanism will not be used. 00:22:45.035 [2024-07-15 11:33:28.386133] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:22:45.035 [2024-07-15 11:33:28.486711] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:22:45.035 [2024-07-15 11:33:28.550158] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:22:45.035 [2024-07-15 11:33:28.550197] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:22:45.601 11:33:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:22:45.601 11:33:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@862 -- # return 0 00:22:45.601 11:33:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:22:45.601 11:33:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:22:45.861 BaseBdev1_malloc 00:22:45.861 11:33:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:22:46.119 [2024-07-15 11:33:29.659658] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:22:46.119 [2024-07-15 11:33:29.659707] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:46.119 [2024-07-15 11:33:29.659730] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2734d40 00:22:46.119 [2024-07-15 11:33:29.659743] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:46.119 [2024-07-15 11:33:29.661405] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:46.119 [2024-07-15 11:33:29.661433] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:22:46.119 BaseBdev1 00:22:46.119 11:33:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:22:46.119 11:33:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:22:46.378 BaseBdev2_malloc 00:22:46.378 11:33:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:22:46.637 [2024-07-15 11:33:30.161954] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:22:46.637 [2024-07-15 11:33:30.162011] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:46.637 [2024-07-15 11:33:30.162038] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2735860 00:22:46.637 [2024-07-15 11:33:30.162050] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:46.637 [2024-07-15 11:33:30.163637] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:46.637 [2024-07-15 11:33:30.163664] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:22:46.637 BaseBdev2 00:22:46.637 11:33:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@606 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b spare_malloc 00:22:46.895 spare_malloc 00:22:46.895 11:33:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:22:47.154 spare_delay 00:22:47.154 11:33:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:22:47.412 [2024-07-15 11:33:30.840384] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:22:47.412 [2024-07-15 11:33:30.840430] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:47.412 [2024-07-15 11:33:30.840449] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x28e3ec0 00:22:47.412 [2024-07-15 11:33:30.840462] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:47.412 [2024-07-15 11:33:30.841885] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:47.412 [2024-07-15 11:33:30.841911] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:22:47.412 spare 00:22:47.412 11:33:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@611 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2' -n raid_bdev1 00:22:47.671 [2024-07-15 11:33:31.089057] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:22:47.671 [2024-07-15 11:33:31.090303] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:22:47.671 [2024-07-15 11:33:31.090380] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x28e5070 00:22:47.671 [2024-07-15 11:33:31.090391] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:22:47.671 [2024-07-15 11:33:31.090596] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x28de490 00:22:47.671 [2024-07-15 11:33:31.090734] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x28e5070 00:22:47.671 [2024-07-15 11:33:31.090744] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x28e5070 00:22:47.671 [2024-07-15 11:33:31.090852] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:47.671 11:33:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:22:47.671 11:33:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:22:47.671 11:33:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:22:47.671 11:33:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:22:47.671 11:33:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:22:47.671 11:33:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:22:47.671 11:33:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:22:47.671 11:33:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:22:47.671 11:33:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:22:47.671 11:33:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:22:47.671 11:33:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:47.671 11:33:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:48.239 11:33:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:22:48.239 "name": "raid_bdev1", 00:22:48.239 "uuid": "bb9eb74a-c382-4d42-83de-60be80c9d41b", 00:22:48.239 "strip_size_kb": 0, 00:22:48.239 "state": "online", 00:22:48.239 "raid_level": "raid1", 00:22:48.239 "superblock": false, 00:22:48.239 "num_base_bdevs": 2, 00:22:48.239 "num_base_bdevs_discovered": 2, 00:22:48.239 "num_base_bdevs_operational": 2, 00:22:48.239 "base_bdevs_list": [ 00:22:48.239 { 00:22:48.239 "name": "BaseBdev1", 00:22:48.239 "uuid": "c82df40a-eaeb-5004-9026-bafc9774b226", 00:22:48.239 "is_configured": true, 00:22:48.239 "data_offset": 0, 00:22:48.239 "data_size": 65536 00:22:48.239 }, 00:22:48.239 { 00:22:48.239 "name": "BaseBdev2", 00:22:48.239 "uuid": "f2dacbb4-abc6-5d22-b739-a98d29925084", 00:22:48.239 "is_configured": true, 00:22:48.239 "data_offset": 0, 00:22:48.239 "data_size": 65536 00:22:48.239 } 00:22:48.239 ] 00:22:48.239 }' 00:22:48.239 11:33:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:22:48.239 11:33:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:22:48.806 11:33:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@615 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:22:48.806 11:33:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@615 -- # jq -r '.[].num_blocks' 00:22:49.064 [2024-07-15 11:33:32.440880] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:22:49.064 11:33:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@615 -- # raid_bdev_size=65536 00:22:49.064 11:33:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@618 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:49.064 11:33:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@618 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:22:49.322 11:33:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@618 -- # data_offset=0 00:22:49.322 11:33:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@620 -- # '[' false = true ']' 00:22:49.322 11:33:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@623 -- # '[' true = true ']' 00:22:49.322 11:33:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # local write_unit_size 00:22:49.322 11:33:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@627 -- # nbd_start_disks /var/tmp/spdk-raid.sock raid_bdev1 /dev/nbd0 00:22:49.322 11:33:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:22:49.322 11:33:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:22:49.322 11:33:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:22:49.322 11:33:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:22:49.322 11:33:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:22:49.322 11:33:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:22:49.322 11:33:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:22:49.322 11:33:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:22:49.322 11:33:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:22:49.885 [2024-07-15 11:33:33.174613] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x28de490 00:22:49.885 /dev/nbd0 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@867 -- # local i 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # break 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:22:49.885 1+0 records in 00:22:49.885 1+0 records out 00:22:49.885 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000235306 s, 17.4 MB/s 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # size=4096 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # return 0 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # '[' raid1 = raid5f ']' 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@632 -- # write_unit_size=1 00:22:49.885 11:33:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@634 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:22:55.140 65536+0 records in 00:22:55.140 65536+0 records out 00:22:55.140 33554432 bytes (34 MB, 32 MiB) copied, 5.22472 s, 6.4 MB/s 00:22:55.140 11:33:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # nbd_stop_disks /var/tmp/spdk-raid.sock /dev/nbd0 00:22:55.140 11:33:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:22:55.140 11:33:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:22:55.140 11:33:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:22:55.140 11:33:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:22:55.140 11:33:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:22:55.140 11:33:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd0 00:22:55.396 [2024-07-15 11:33:38.745847] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:55.396 11:33:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:22:55.396 11:33:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:22:55.396 11:33:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:22:55.396 11:33:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:22:55.396 11:33:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:22:55.396 11:33:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:22:55.396 11:33:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:22:55.396 11:33:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:22:55.396 11:33:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@639 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev1 00:22:55.396 [2024-07-15 11:33:38.974516] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:22:55.652 11:33:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@642 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:22:55.652 11:33:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:22:55.652 11:33:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:22:55.652 11:33:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:22:55.652 11:33:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:22:55.652 11:33:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:22:55.652 11:33:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:22:55.652 11:33:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:22:55.652 11:33:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:22:55.652 11:33:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:22:55.652 11:33:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:55.652 11:33:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:55.652 11:33:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:22:55.652 "name": "raid_bdev1", 00:22:55.652 "uuid": "bb9eb74a-c382-4d42-83de-60be80c9d41b", 00:22:55.652 "strip_size_kb": 0, 00:22:55.652 "state": "online", 00:22:55.652 "raid_level": "raid1", 00:22:55.652 "superblock": false, 00:22:55.652 "num_base_bdevs": 2, 00:22:55.652 "num_base_bdevs_discovered": 1, 00:22:55.652 "num_base_bdevs_operational": 1, 00:22:55.652 "base_bdevs_list": [ 00:22:55.652 { 00:22:55.652 "name": null, 00:22:55.652 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:55.652 "is_configured": false, 00:22:55.652 "data_offset": 0, 00:22:55.652 "data_size": 65536 00:22:55.652 }, 00:22:55.652 { 00:22:55.652 "name": "BaseBdev2", 00:22:55.652 "uuid": "f2dacbb4-abc6-5d22-b739-a98d29925084", 00:22:55.652 "is_configured": true, 00:22:55.652 "data_offset": 0, 00:22:55.652 "data_size": 65536 00:22:55.652 } 00:22:55.652 ] 00:22:55.652 }' 00:22:55.652 11:33:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:22:55.652 11:33:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:22:56.216 11:33:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@645 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:22:56.472 [2024-07-15 11:33:39.997226] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:22:56.472 [2024-07-15 11:33:40.002253] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x28e5880 00:22:56.472 [2024-07-15 11:33:40.004478] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:22:56.472 11:33:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # sleep 1 00:22:57.874 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@649 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:22:57.874 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:22:57.874 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:22:57.875 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@184 -- # local target=spare 00:22:57.875 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:22:57.875 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:57.875 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:57.875 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:22:57.875 "name": "raid_bdev1", 00:22:57.875 "uuid": "bb9eb74a-c382-4d42-83de-60be80c9d41b", 00:22:57.875 "strip_size_kb": 0, 00:22:57.875 "state": "online", 00:22:57.875 "raid_level": "raid1", 00:22:57.875 "superblock": false, 00:22:57.875 "num_base_bdevs": 2, 00:22:57.875 "num_base_bdevs_discovered": 2, 00:22:57.875 "num_base_bdevs_operational": 2, 00:22:57.875 "process": { 00:22:57.875 "type": "rebuild", 00:22:57.875 "target": "spare", 00:22:57.875 "progress": { 00:22:57.875 "blocks": 24576, 00:22:57.875 "percent": 37 00:22:57.875 } 00:22:57.875 }, 00:22:57.875 "base_bdevs_list": [ 00:22:57.875 { 00:22:57.875 "name": "spare", 00:22:57.875 "uuid": "9a0802f2-ea12-552d-bf00-a08a99ed951f", 00:22:57.875 "is_configured": true, 00:22:57.875 "data_offset": 0, 00:22:57.875 "data_size": 65536 00:22:57.875 }, 00:22:57.875 { 00:22:57.875 "name": "BaseBdev2", 00:22:57.875 "uuid": "f2dacbb4-abc6-5d22-b739-a98d29925084", 00:22:57.875 "is_configured": true, 00:22:57.875 "data_offset": 0, 00:22:57.875 "data_size": 65536 00:22:57.875 } 00:22:57.875 ] 00:22:57.875 }' 00:22:57.875 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:22:57.875 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:22:57.875 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:22:57.875 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:22:57.875 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@652 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev spare 00:22:58.133 [2024-07-15 11:33:41.578350] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:22:58.133 [2024-07-15 11:33:41.617202] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:22:58.133 [2024-07-15 11:33:41.617249] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:58.133 [2024-07-15 11:33:41.617265] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:22:58.133 [2024-07-15 11:33:41.617273] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:22:58.133 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@655 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:22:58.133 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:22:58.133 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:22:58.133 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:22:58.133 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:22:58.133 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:22:58.133 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:22:58.133 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:22:58.133 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:22:58.133 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:22:58.133 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:58.133 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:58.391 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:22:58.391 "name": "raid_bdev1", 00:22:58.391 "uuid": "bb9eb74a-c382-4d42-83de-60be80c9d41b", 00:22:58.391 "strip_size_kb": 0, 00:22:58.391 "state": "online", 00:22:58.391 "raid_level": "raid1", 00:22:58.391 "superblock": false, 00:22:58.391 "num_base_bdevs": 2, 00:22:58.391 "num_base_bdevs_discovered": 1, 00:22:58.391 "num_base_bdevs_operational": 1, 00:22:58.391 "base_bdevs_list": [ 00:22:58.391 { 00:22:58.391 "name": null, 00:22:58.391 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:58.391 "is_configured": false, 00:22:58.391 "data_offset": 0, 00:22:58.391 "data_size": 65536 00:22:58.391 }, 00:22:58.391 { 00:22:58.391 "name": "BaseBdev2", 00:22:58.391 "uuid": "f2dacbb4-abc6-5d22-b739-a98d29925084", 00:22:58.391 "is_configured": true, 00:22:58.391 "data_offset": 0, 00:22:58.391 "data_size": 65536 00:22:58.391 } 00:22:58.391 ] 00:22:58.391 }' 00:22:58.391 11:33:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:22:58.391 11:33:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:22:58.958 11:33:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@658 -- # verify_raid_bdev_process raid_bdev1 none none 00:22:58.958 11:33:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:22:58.958 11:33:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:22:58.958 11:33:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@184 -- # local target=none 00:22:58.958 11:33:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:22:58.958 11:33:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:58.958 11:33:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:22:59.216 11:33:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:22:59.216 "name": "raid_bdev1", 00:22:59.216 "uuid": "bb9eb74a-c382-4d42-83de-60be80c9d41b", 00:22:59.216 "strip_size_kb": 0, 00:22:59.216 "state": "online", 00:22:59.216 "raid_level": "raid1", 00:22:59.216 "superblock": false, 00:22:59.216 "num_base_bdevs": 2, 00:22:59.216 "num_base_bdevs_discovered": 1, 00:22:59.216 "num_base_bdevs_operational": 1, 00:22:59.216 "base_bdevs_list": [ 00:22:59.216 { 00:22:59.216 "name": null, 00:22:59.216 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:59.216 "is_configured": false, 00:22:59.216 "data_offset": 0, 00:22:59.216 "data_size": 65536 00:22:59.216 }, 00:22:59.216 { 00:22:59.216 "name": "BaseBdev2", 00:22:59.216 "uuid": "f2dacbb4-abc6-5d22-b739-a98d29925084", 00:22:59.216 "is_configured": true, 00:22:59.216 "data_offset": 0, 00:22:59.216 "data_size": 65536 00:22:59.216 } 00:22:59.216 ] 00:22:59.216 }' 00:22:59.216 11:33:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:22:59.216 11:33:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:22:59.216 11:33:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:22:59.216 11:33:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:22:59.216 11:33:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@661 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:22:59.474 [2024-07-15 11:33:42.969341] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:22:59.474 [2024-07-15 11:33:42.974295] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x28de490 00:22:59.474 [2024-07-15 11:33:42.975758] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:22:59.474 11:33:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # sleep 1 00:23:00.407 11:33:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:23:00.407 11:33:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:00.407 11:33:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:23:00.407 11:33:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@184 -- # local target=spare 00:23:00.407 11:33:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:00.407 11:33:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:00.407 11:33:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:00.665 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:00.665 "name": "raid_bdev1", 00:23:00.665 "uuid": "bb9eb74a-c382-4d42-83de-60be80c9d41b", 00:23:00.665 "strip_size_kb": 0, 00:23:00.665 "state": "online", 00:23:00.665 "raid_level": "raid1", 00:23:00.665 "superblock": false, 00:23:00.665 "num_base_bdevs": 2, 00:23:00.665 "num_base_bdevs_discovered": 2, 00:23:00.665 "num_base_bdevs_operational": 2, 00:23:00.665 "process": { 00:23:00.665 "type": "rebuild", 00:23:00.665 "target": "spare", 00:23:00.665 "progress": { 00:23:00.665 "blocks": 22528, 00:23:00.665 "percent": 34 00:23:00.665 } 00:23:00.666 }, 00:23:00.666 "base_bdevs_list": [ 00:23:00.666 { 00:23:00.666 "name": "spare", 00:23:00.666 "uuid": "9a0802f2-ea12-552d-bf00-a08a99ed951f", 00:23:00.666 "is_configured": true, 00:23:00.666 "data_offset": 0, 00:23:00.666 "data_size": 65536 00:23:00.666 }, 00:23:00.666 { 00:23:00.666 "name": "BaseBdev2", 00:23:00.666 "uuid": "f2dacbb4-abc6-5d22-b739-a98d29925084", 00:23:00.666 "is_configured": true, 00:23:00.666 "data_offset": 0, 00:23:00.666 "data_size": 65536 00:23:00.666 } 00:23:00.666 ] 00:23:00.666 }' 00:23:00.666 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:00.666 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:23:00.666 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:00.924 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:23:00.924 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@665 -- # '[' false = true ']' 00:23:00.924 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@690 -- # local num_base_bdevs_operational=2 00:23:00.924 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@692 -- # '[' raid1 = raid1 ']' 00:23:00.924 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@692 -- # '[' 2 -gt 2 ']' 00:23:00.924 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@705 -- # local timeout=752 00:23:00.924 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:23:00.924 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:23:00.924 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:00.925 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:23:00.925 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@184 -- # local target=spare 00:23:00.925 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:00.925 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:00.925 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:01.183 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:01.183 "name": "raid_bdev1", 00:23:01.183 "uuid": "bb9eb74a-c382-4d42-83de-60be80c9d41b", 00:23:01.183 "strip_size_kb": 0, 00:23:01.183 "state": "online", 00:23:01.183 "raid_level": "raid1", 00:23:01.183 "superblock": false, 00:23:01.183 "num_base_bdevs": 2, 00:23:01.183 "num_base_bdevs_discovered": 2, 00:23:01.183 "num_base_bdevs_operational": 2, 00:23:01.183 "process": { 00:23:01.183 "type": "rebuild", 00:23:01.183 "target": "spare", 00:23:01.183 "progress": { 00:23:01.183 "blocks": 30720, 00:23:01.183 "percent": 46 00:23:01.183 } 00:23:01.183 }, 00:23:01.183 "base_bdevs_list": [ 00:23:01.183 { 00:23:01.183 "name": "spare", 00:23:01.183 "uuid": "9a0802f2-ea12-552d-bf00-a08a99ed951f", 00:23:01.183 "is_configured": true, 00:23:01.183 "data_offset": 0, 00:23:01.183 "data_size": 65536 00:23:01.183 }, 00:23:01.183 { 00:23:01.183 "name": "BaseBdev2", 00:23:01.183 "uuid": "f2dacbb4-abc6-5d22-b739-a98d29925084", 00:23:01.183 "is_configured": true, 00:23:01.183 "data_offset": 0, 00:23:01.183 "data_size": 65536 00:23:01.183 } 00:23:01.183 ] 00:23:01.183 }' 00:23:01.183 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:01.183 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:23:01.183 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:01.183 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:23:01.183 11:33:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@710 -- # sleep 1 00:23:02.116 11:33:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:23:02.116 11:33:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:23:02.116 11:33:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:02.116 11:33:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:23:02.116 11:33:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@184 -- # local target=spare 00:23:02.116 11:33:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:02.116 11:33:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:02.116 11:33:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:02.374 11:33:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:02.374 "name": "raid_bdev1", 00:23:02.374 "uuid": "bb9eb74a-c382-4d42-83de-60be80c9d41b", 00:23:02.374 "strip_size_kb": 0, 00:23:02.374 "state": "online", 00:23:02.374 "raid_level": "raid1", 00:23:02.374 "superblock": false, 00:23:02.374 "num_base_bdevs": 2, 00:23:02.374 "num_base_bdevs_discovered": 2, 00:23:02.374 "num_base_bdevs_operational": 2, 00:23:02.374 "process": { 00:23:02.374 "type": "rebuild", 00:23:02.374 "target": "spare", 00:23:02.374 "progress": { 00:23:02.374 "blocks": 57344, 00:23:02.374 "percent": 87 00:23:02.374 } 00:23:02.374 }, 00:23:02.374 "base_bdevs_list": [ 00:23:02.374 { 00:23:02.374 "name": "spare", 00:23:02.374 "uuid": "9a0802f2-ea12-552d-bf00-a08a99ed951f", 00:23:02.374 "is_configured": true, 00:23:02.374 "data_offset": 0, 00:23:02.374 "data_size": 65536 00:23:02.374 }, 00:23:02.374 { 00:23:02.374 "name": "BaseBdev2", 00:23:02.374 "uuid": "f2dacbb4-abc6-5d22-b739-a98d29925084", 00:23:02.374 "is_configured": true, 00:23:02.374 "data_offset": 0, 00:23:02.374 "data_size": 65536 00:23:02.374 } 00:23:02.374 ] 00:23:02.374 }' 00:23:02.374 11:33:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:02.374 11:33:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:23:02.374 11:33:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:02.374 11:33:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:23:02.374 11:33:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@710 -- # sleep 1 00:23:02.631 [2024-07-15 11:33:46.200380] bdev_raid.c:2789:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:23:02.631 [2024-07-15 11:33:46.200440] bdev_raid.c:2504:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:23:02.631 [2024-07-15 11:33:46.200476] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:03.564 11:33:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:23:03.564 11:33:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:23:03.564 11:33:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:03.564 11:33:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:23:03.564 11:33:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@184 -- # local target=spare 00:23:03.564 11:33:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:03.564 11:33:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:03.564 11:33:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:03.823 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:03.823 "name": "raid_bdev1", 00:23:03.823 "uuid": "bb9eb74a-c382-4d42-83de-60be80c9d41b", 00:23:03.823 "strip_size_kb": 0, 00:23:03.823 "state": "online", 00:23:03.823 "raid_level": "raid1", 00:23:03.823 "superblock": false, 00:23:03.823 "num_base_bdevs": 2, 00:23:03.823 "num_base_bdevs_discovered": 2, 00:23:03.823 "num_base_bdevs_operational": 2, 00:23:03.823 "base_bdevs_list": [ 00:23:03.823 { 00:23:03.823 "name": "spare", 00:23:03.823 "uuid": "9a0802f2-ea12-552d-bf00-a08a99ed951f", 00:23:03.823 "is_configured": true, 00:23:03.823 "data_offset": 0, 00:23:03.823 "data_size": 65536 00:23:03.823 }, 00:23:03.823 { 00:23:03.823 "name": "BaseBdev2", 00:23:03.823 "uuid": "f2dacbb4-abc6-5d22-b739-a98d29925084", 00:23:03.823 "is_configured": true, 00:23:03.823 "data_offset": 0, 00:23:03.823 "data_size": 65536 00:23:03.823 } 00:23:03.823 ] 00:23:03.823 }' 00:23:03.823 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:03.823 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # [[ none == \r\e\b\u\i\l\d ]] 00:23:03.823 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:03.823 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # [[ none == \s\p\a\r\e ]] 00:23:03.823 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # break 00:23:03.823 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@714 -- # verify_raid_bdev_process raid_bdev1 none none 00:23:03.823 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:03.823 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:23:03.823 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@184 -- # local target=none 00:23:03.823 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:03.823 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:03.823 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:04.081 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:04.081 "name": "raid_bdev1", 00:23:04.081 "uuid": "bb9eb74a-c382-4d42-83de-60be80c9d41b", 00:23:04.081 "strip_size_kb": 0, 00:23:04.081 "state": "online", 00:23:04.081 "raid_level": "raid1", 00:23:04.081 "superblock": false, 00:23:04.081 "num_base_bdevs": 2, 00:23:04.081 "num_base_bdevs_discovered": 2, 00:23:04.081 "num_base_bdevs_operational": 2, 00:23:04.081 "base_bdevs_list": [ 00:23:04.081 { 00:23:04.081 "name": "spare", 00:23:04.081 "uuid": "9a0802f2-ea12-552d-bf00-a08a99ed951f", 00:23:04.081 "is_configured": true, 00:23:04.081 "data_offset": 0, 00:23:04.081 "data_size": 65536 00:23:04.081 }, 00:23:04.081 { 00:23:04.081 "name": "BaseBdev2", 00:23:04.081 "uuid": "f2dacbb4-abc6-5d22-b739-a98d29925084", 00:23:04.081 "is_configured": true, 00:23:04.081 "data_offset": 0, 00:23:04.081 "data_size": 65536 00:23:04.081 } 00:23:04.081 ] 00:23:04.081 }' 00:23:04.081 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:04.081 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:23:04.081 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:04.081 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:23:04.081 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:23:04.081 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:23:04.081 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:23:04.081 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:23:04.081 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:23:04.081 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:23:04.081 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:23:04.082 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:23:04.082 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:23:04.082 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:23:04.082 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:04.082 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:04.340 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:23:04.340 "name": "raid_bdev1", 00:23:04.340 "uuid": "bb9eb74a-c382-4d42-83de-60be80c9d41b", 00:23:04.340 "strip_size_kb": 0, 00:23:04.340 "state": "online", 00:23:04.340 "raid_level": "raid1", 00:23:04.340 "superblock": false, 00:23:04.340 "num_base_bdevs": 2, 00:23:04.340 "num_base_bdevs_discovered": 2, 00:23:04.340 "num_base_bdevs_operational": 2, 00:23:04.340 "base_bdevs_list": [ 00:23:04.340 { 00:23:04.340 "name": "spare", 00:23:04.340 "uuid": "9a0802f2-ea12-552d-bf00-a08a99ed951f", 00:23:04.340 "is_configured": true, 00:23:04.340 "data_offset": 0, 00:23:04.340 "data_size": 65536 00:23:04.340 }, 00:23:04.340 { 00:23:04.340 "name": "BaseBdev2", 00:23:04.340 "uuid": "f2dacbb4-abc6-5d22-b739-a98d29925084", 00:23:04.340 "is_configured": true, 00:23:04.340 "data_offset": 0, 00:23:04.340 "data_size": 65536 00:23:04.340 } 00:23:04.340 ] 00:23:04.340 }' 00:23:04.340 11:33:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:23:04.340 11:33:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:23:04.905 11:33:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@718 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:23:05.163 [2024-07-15 11:33:48.627840] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:23:05.163 [2024-07-15 11:33:48.627870] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:23:05.163 [2024-07-15 11:33:48.627938] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:23:05.163 [2024-07-15 11:33:48.627994] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:23:05.163 [2024-07-15 11:33:48.628007] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x28e5070 name raid_bdev1, state offline 00:23:05.163 11:33:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:05.163 11:33:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # jq length 00:23:05.421 11:33:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # [[ 0 == 0 ]] 00:23:05.421 11:33:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@721 -- # '[' true = true ']' 00:23:05.421 11:33:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' false = true ']' 00:23:05.421 11:33:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@736 -- # nbd_start_disks /var/tmp/spdk-raid.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:23:05.421 11:33:48 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:23:05.421 11:33:48 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:23:05.421 11:33:48 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:23:05.421 11:33:48 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:23:05.421 11:33:48 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:23:05.421 11:33:48 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:23:05.421 11:33:48 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:23:05.421 11:33:48 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:23:05.421 11:33:48 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:23:05.679 /dev/nbd0 00:23:05.679 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:23:05.679 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:23:05.679 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:23:05.679 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@867 -- # local i 00:23:05.679 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:23:05.679 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:23:05.679 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:23:05.679 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # break 00:23:05.679 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:23:05.679 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:23:05.679 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:23:05.679 1+0 records in 00:23:05.679 1+0 records out 00:23:05.679 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000157951 s, 25.9 MB/s 00:23:05.679 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:23:05.679 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # size=4096 00:23:05.679 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:23:05.679 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:23:05.679 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # return 0 00:23:05.679 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:23:05.679 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:23:05.679 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk spare /dev/nbd1 00:23:05.938 /dev/nbd1 00:23:05.938 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:23:05.938 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:23:05.938 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:23:05.938 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@867 -- # local i 00:23:05.938 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:23:05.938 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:23:05.938 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:23:05.938 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # break 00:23:05.938 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:23:05.938 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:23:05.938 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:23:05.938 1+0 records in 00:23:05.938 1+0 records out 00:23:05.938 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00033325 s, 12.3 MB/s 00:23:05.938 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:23:05.938 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # size=4096 00:23:05.938 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:23:05.938 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:23:05.938 11:33:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # return 0 00:23:05.938 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:23:05.938 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:23:05.938 11:33:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:23:06.196 11:33:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # nbd_stop_disks /var/tmp/spdk-raid.sock '/dev/nbd0 /dev/nbd1' 00:23:06.196 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:23:06.196 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:23:06.196 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:23:06.196 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:23:06.196 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:23:06.196 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd0 00:23:06.455 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:23:06.455 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:23:06.455 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:23:06.455 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:23:06.455 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:23:06.455 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:23:06.455 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@39 -- # sleep 0.1 00:23:06.455 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i++ )) 00:23:06.455 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:23:06.455 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:23:06.455 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:23:06.455 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:23:06.455 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:23:06.455 11:33:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd1 00:23:06.714 11:33:50 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:23:06.714 11:33:50 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:23:06.714 11:33:50 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:23:06.714 11:33:50 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:23:06.714 11:33:50 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:23:06.714 11:33:50 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:23:06.714 11:33:50 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:23:06.714 11:33:50 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:23:06.714 11:33:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@742 -- # '[' false = true ']' 00:23:06.714 11:33:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@782 -- # killprocess 968376 00:23:06.714 11:33:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@948 -- # '[' -z 968376 ']' 00:23:06.714 11:33:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@952 -- # kill -0 968376 00:23:06.714 11:33:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@953 -- # uname 00:23:06.714 11:33:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:23:06.714 11:33:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 968376 00:23:06.714 11:33:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:23:06.714 11:33:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:23:06.714 11:33:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 968376' 00:23:06.714 killing process with pid 968376 00:23:06.714 11:33:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@967 -- # kill 968376 00:23:06.714 Received shutdown signal, test time was about 60.000000 seconds 00:23:06.714 00:23:06.714 Latency(us) 00:23:06.714 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:23:06.714 =================================================================================================================== 00:23:06.714 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:23:06.714 [2024-07-15 11:33:50.272437] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:23:06.714 11:33:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@972 -- # wait 968376 00:23:06.714 [2024-07-15 11:33:50.300595] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:23:06.973 11:33:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # return 0 00:23:06.973 00:23:06.973 real 0m22.343s 00:23:06.973 user 0m30.054s 00:23:06.973 sys 0m5.142s 00:23:06.973 11:33:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:23:06.973 11:33:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:23:06.973 ************************************ 00:23:06.973 END TEST raid_rebuild_test 00:23:06.973 ************************************ 00:23:07.231 11:33:50 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:23:07.231 11:33:50 bdev_raid -- bdev/bdev_raid.sh@878 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 2 true false true 00:23:07.231 11:33:50 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:23:07.231 11:33:50 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:23:07.231 11:33:50 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:23:07.231 ************************************ 00:23:07.231 START TEST raid_rebuild_test_sb 00:23:07.231 ************************************ 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1123 -- # raid_rebuild_test raid1 2 true false true 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@568 -- # local raid_level=raid1 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local num_base_bdevs=2 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local superblock=true 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local background_io=false 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local verify=true 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # (( i = 1 )) 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # echo BaseBdev1 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # echo BaseBdev2 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local base_bdevs 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local raid_bdev_name=raid_bdev1 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local strip_size 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local create_arg 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local raid_bdev_size 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local data_offset 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@580 -- # '[' raid1 '!=' raid1 ']' 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@588 -- # strip_size=0 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@591 -- # '[' true = true ']' 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # create_arg+=' -s' 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # raid_pid=971434 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # waitforlisten 971434 /var/tmp/spdk-raid.sock 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@595 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@829 -- # '[' -z 971434 ']' 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@834 -- # local max_retries=100 00:23:07.231 11:33:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:23:07.232 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:23:07.232 11:33:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@838 -- # xtrace_disable 00:23:07.232 11:33:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:07.232 [2024-07-15 11:33:50.685607] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:23:07.232 [2024-07-15 11:33:50.685679] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid971434 ] 00:23:07.232 I/O size of 3145728 is greater than zero copy threshold (65536). 00:23:07.232 Zero copy mechanism will not be used. 00:23:07.232 [2024-07-15 11:33:50.818498] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:07.490 [2024-07-15 11:33:50.920998] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:23:07.490 [2024-07-15 11:33:50.981622] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:23:07.490 [2024-07-15 11:33:50.981658] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:23:08.056 11:33:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:23:08.056 11:33:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@862 -- # return 0 00:23:08.056 11:33:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:23:08.056 11:33:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:23:08.314 BaseBdev1_malloc 00:23:08.314 11:33:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:23:08.571 [2024-07-15 11:33:51.951748] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:23:08.571 [2024-07-15 11:33:51.951803] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:08.571 [2024-07-15 11:33:51.951827] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x10e5d40 00:23:08.571 [2024-07-15 11:33:51.951840] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:08.571 [2024-07-15 11:33:51.953556] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:08.571 [2024-07-15 11:33:51.953586] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:23:08.571 BaseBdev1 00:23:08.571 11:33:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:23:08.571 11:33:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:23:08.828 BaseBdev2_malloc 00:23:08.828 11:33:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:23:09.086 [2024-07-15 11:33:52.429914] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:23:09.086 [2024-07-15 11:33:52.429967] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:09.086 [2024-07-15 11:33:52.429990] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x10e6860 00:23:09.086 [2024-07-15 11:33:52.430003] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:09.086 [2024-07-15 11:33:52.431367] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:09.086 [2024-07-15 11:33:52.431394] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:23:09.086 BaseBdev2 00:23:09.086 11:33:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@606 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b spare_malloc 00:23:09.086 spare_malloc 00:23:09.343 11:33:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:23:09.602 spare_delay 00:23:09.859 11:33:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:23:09.859 [2024-07-15 11:33:53.418416] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:23:09.859 [2024-07-15 11:33:53.418463] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:09.859 [2024-07-15 11:33:53.418485] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1294ec0 00:23:09.860 [2024-07-15 11:33:53.418498] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:09.860 [2024-07-15 11:33:53.420083] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:09.860 [2024-07-15 11:33:53.420111] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:23:09.860 spare 00:23:09.860 11:33:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@611 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2' -n raid_bdev1 00:23:10.426 [2024-07-15 11:33:53.915755] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:23:10.426 [2024-07-15 11:33:53.917090] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:23:10.426 [2024-07-15 11:33:53.917260] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1296070 00:23:10.426 [2024-07-15 11:33:53.917273] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:23:10.426 [2024-07-15 11:33:53.917473] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x128f490 00:23:10.426 [2024-07-15 11:33:53.917614] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1296070 00:23:10.426 [2024-07-15 11:33:53.917624] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x1296070 00:23:10.426 [2024-07-15 11:33:53.917728] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:10.426 11:33:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:23:10.426 11:33:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:23:10.426 11:33:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:23:10.426 11:33:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:23:10.426 11:33:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:23:10.426 11:33:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:23:10.426 11:33:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:23:10.427 11:33:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:23:10.427 11:33:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:23:10.427 11:33:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:23:10.427 11:33:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:10.427 11:33:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:10.685 11:33:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:23:10.685 "name": "raid_bdev1", 00:23:10.685 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:10.685 "strip_size_kb": 0, 00:23:10.685 "state": "online", 00:23:10.685 "raid_level": "raid1", 00:23:10.685 "superblock": true, 00:23:10.685 "num_base_bdevs": 2, 00:23:10.685 "num_base_bdevs_discovered": 2, 00:23:10.685 "num_base_bdevs_operational": 2, 00:23:10.685 "base_bdevs_list": [ 00:23:10.685 { 00:23:10.685 "name": "BaseBdev1", 00:23:10.685 "uuid": "f53fd0cb-5397-5df8-b286-066ef36dffbf", 00:23:10.685 "is_configured": true, 00:23:10.685 "data_offset": 2048, 00:23:10.685 "data_size": 63488 00:23:10.685 }, 00:23:10.685 { 00:23:10.685 "name": "BaseBdev2", 00:23:10.685 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:10.685 "is_configured": true, 00:23:10.685 "data_offset": 2048, 00:23:10.685 "data_size": 63488 00:23:10.685 } 00:23:10.685 ] 00:23:10.685 }' 00:23:10.685 11:33:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:23:10.685 11:33:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:11.250 11:33:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@615 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:23:11.250 11:33:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@615 -- # jq -r '.[].num_blocks' 00:23:11.507 [2024-07-15 11:33:54.942676] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:23:11.507 11:33:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@615 -- # raid_bdev_size=63488 00:23:11.508 11:33:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@618 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:11.508 11:33:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@618 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:23:11.766 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@618 -- # data_offset=2048 00:23:11.766 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@620 -- # '[' false = true ']' 00:23:11.766 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@623 -- # '[' true = true ']' 00:23:11.766 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # local write_unit_size 00:23:11.766 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@627 -- # nbd_start_disks /var/tmp/spdk-raid.sock raid_bdev1 /dev/nbd0 00:23:11.766 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:23:11.766 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:23:11.766 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:23:11.766 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:23:11.766 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:23:11.766 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:23:11.766 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:23:11.766 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:23:11.766 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:23:12.024 [2024-07-15 11:33:55.435786] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x128f490 00:23:12.024 /dev/nbd0 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@867 -- # local i 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # break 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:23:12.024 1+0 records in 00:23:12.024 1+0 records out 00:23:12.024 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000253029 s, 16.2 MB/s 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # size=4096 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # return 0 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # '[' raid1 = raid5f ']' 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@632 -- # write_unit_size=1 00:23:12.024 11:33:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@634 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:23:17.287 63488+0 records in 00:23:17.287 63488+0 records out 00:23:17.287 32505856 bytes (33 MB, 31 MiB) copied, 5.32385 s, 6.1 MB/s 00:23:17.287 11:34:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # nbd_stop_disks /var/tmp/spdk-raid.sock /dev/nbd0 00:23:17.287 11:34:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:23:17.287 11:34:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:23:17.287 11:34:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:23:17.287 11:34:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:23:17.287 11:34:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:23:17.287 11:34:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd0 00:23:17.575 [2024-07-15 11:34:01.100368] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:17.575 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:23:17.575 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:23:17.575 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:23:17.575 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:23:17.575 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:23:17.575 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:23:17.575 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:23:17.575 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:23:17.575 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@639 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev1 00:23:17.861 [2024-07-15 11:34:01.341065] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:23:17.861 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@642 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:23:17.861 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:23:17.861 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:23:17.861 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:23:17.861 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:23:17.861 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:23:17.861 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:23:17.861 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:23:17.861 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:23:17.861 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:23:17.861 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:17.861 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:18.121 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:23:18.121 "name": "raid_bdev1", 00:23:18.121 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:18.121 "strip_size_kb": 0, 00:23:18.121 "state": "online", 00:23:18.121 "raid_level": "raid1", 00:23:18.121 "superblock": true, 00:23:18.121 "num_base_bdevs": 2, 00:23:18.121 "num_base_bdevs_discovered": 1, 00:23:18.121 "num_base_bdevs_operational": 1, 00:23:18.121 "base_bdevs_list": [ 00:23:18.121 { 00:23:18.121 "name": null, 00:23:18.121 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:18.121 "is_configured": false, 00:23:18.121 "data_offset": 2048, 00:23:18.121 "data_size": 63488 00:23:18.121 }, 00:23:18.121 { 00:23:18.121 "name": "BaseBdev2", 00:23:18.121 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:18.121 "is_configured": true, 00:23:18.121 "data_offset": 2048, 00:23:18.121 "data_size": 63488 00:23:18.121 } 00:23:18.121 ] 00:23:18.121 }' 00:23:18.121 11:34:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:23:18.121 11:34:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:18.686 11:34:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@645 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:23:18.944 [2024-07-15 11:34:02.435970] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:23:18.944 [2024-07-15 11:34:02.440917] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1296b30 00:23:18.944 [2024-07-15 11:34:02.443129] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:23:18.944 11:34:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # sleep 1 00:23:19.878 11:34:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@649 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:23:19.878 11:34:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:19.878 11:34:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:23:19.878 11:34:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=spare 00:23:19.878 11:34:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:19.878 11:34:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:19.878 11:34:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:20.136 11:34:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:20.136 "name": "raid_bdev1", 00:23:20.136 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:20.136 "strip_size_kb": 0, 00:23:20.136 "state": "online", 00:23:20.136 "raid_level": "raid1", 00:23:20.136 "superblock": true, 00:23:20.136 "num_base_bdevs": 2, 00:23:20.136 "num_base_bdevs_discovered": 2, 00:23:20.136 "num_base_bdevs_operational": 2, 00:23:20.136 "process": { 00:23:20.136 "type": "rebuild", 00:23:20.136 "target": "spare", 00:23:20.136 "progress": { 00:23:20.136 "blocks": 24576, 00:23:20.136 "percent": 38 00:23:20.136 } 00:23:20.136 }, 00:23:20.136 "base_bdevs_list": [ 00:23:20.136 { 00:23:20.136 "name": "spare", 00:23:20.136 "uuid": "e565947d-5b48-51dd-8764-c5371a138fd5", 00:23:20.136 "is_configured": true, 00:23:20.136 "data_offset": 2048, 00:23:20.136 "data_size": 63488 00:23:20.136 }, 00:23:20.136 { 00:23:20.136 "name": "BaseBdev2", 00:23:20.136 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:20.136 "is_configured": true, 00:23:20.136 "data_offset": 2048, 00:23:20.136 "data_size": 63488 00:23:20.136 } 00:23:20.136 ] 00:23:20.136 }' 00:23:20.136 11:34:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:20.394 11:34:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:23:20.394 11:34:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:20.394 11:34:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:23:20.394 11:34:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@652 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev spare 00:23:20.652 [2024-07-15 11:34:04.025951] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:23:20.652 [2024-07-15 11:34:04.055681] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:23:20.652 [2024-07-15 11:34:04.055725] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:20.652 [2024-07-15 11:34:04.055740] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:23:20.652 [2024-07-15 11:34:04.055749] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:23:20.652 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@655 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:23:20.652 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:23:20.652 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:23:20.652 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:23:20.652 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:23:20.652 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:23:20.653 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:23:20.653 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:23:20.653 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:23:20.653 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:23:20.653 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:20.653 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:20.911 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:23:20.911 "name": "raid_bdev1", 00:23:20.911 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:20.911 "strip_size_kb": 0, 00:23:20.911 "state": "online", 00:23:20.911 "raid_level": "raid1", 00:23:20.911 "superblock": true, 00:23:20.911 "num_base_bdevs": 2, 00:23:20.911 "num_base_bdevs_discovered": 1, 00:23:20.911 "num_base_bdevs_operational": 1, 00:23:20.911 "base_bdevs_list": [ 00:23:20.911 { 00:23:20.911 "name": null, 00:23:20.911 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:20.911 "is_configured": false, 00:23:20.911 "data_offset": 2048, 00:23:20.911 "data_size": 63488 00:23:20.911 }, 00:23:20.911 { 00:23:20.911 "name": "BaseBdev2", 00:23:20.911 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:20.911 "is_configured": true, 00:23:20.911 "data_offset": 2048, 00:23:20.911 "data_size": 63488 00:23:20.911 } 00:23:20.911 ] 00:23:20.911 }' 00:23:20.911 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:23:20.911 11:34:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:21.477 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@658 -- # verify_raid_bdev_process raid_bdev1 none none 00:23:21.477 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:21.477 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:23:21.477 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=none 00:23:21.477 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:21.477 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:21.477 11:34:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:21.736 11:34:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:21.736 "name": "raid_bdev1", 00:23:21.736 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:21.736 "strip_size_kb": 0, 00:23:21.736 "state": "online", 00:23:21.736 "raid_level": "raid1", 00:23:21.736 "superblock": true, 00:23:21.736 "num_base_bdevs": 2, 00:23:21.736 "num_base_bdevs_discovered": 1, 00:23:21.736 "num_base_bdevs_operational": 1, 00:23:21.736 "base_bdevs_list": [ 00:23:21.736 { 00:23:21.736 "name": null, 00:23:21.736 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:21.736 "is_configured": false, 00:23:21.736 "data_offset": 2048, 00:23:21.736 "data_size": 63488 00:23:21.736 }, 00:23:21.736 { 00:23:21.736 "name": "BaseBdev2", 00:23:21.736 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:21.736 "is_configured": true, 00:23:21.736 "data_offset": 2048, 00:23:21.736 "data_size": 63488 00:23:21.736 } 00:23:21.736 ] 00:23:21.736 }' 00:23:21.736 11:34:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:21.736 11:34:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:23:21.736 11:34:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:21.736 11:34:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:23:21.736 11:34:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@661 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:23:21.994 [2024-07-15 11:34:05.419755] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:23:21.994 [2024-07-15 11:34:05.424712] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1295ce0 00:23:21.994 [2024-07-15 11:34:05.426183] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:23:21.994 11:34:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # sleep 1 00:23:22.928 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:23:22.928 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:22.928 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:23:22.928 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=spare 00:23:22.928 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:22.928 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:22.928 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:23.186 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:23.186 "name": "raid_bdev1", 00:23:23.186 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:23.186 "strip_size_kb": 0, 00:23:23.186 "state": "online", 00:23:23.186 "raid_level": "raid1", 00:23:23.186 "superblock": true, 00:23:23.186 "num_base_bdevs": 2, 00:23:23.186 "num_base_bdevs_discovered": 2, 00:23:23.186 "num_base_bdevs_operational": 2, 00:23:23.186 "process": { 00:23:23.186 "type": "rebuild", 00:23:23.186 "target": "spare", 00:23:23.186 "progress": { 00:23:23.186 "blocks": 24576, 00:23:23.186 "percent": 38 00:23:23.186 } 00:23:23.186 }, 00:23:23.186 "base_bdevs_list": [ 00:23:23.186 { 00:23:23.186 "name": "spare", 00:23:23.186 "uuid": "e565947d-5b48-51dd-8764-c5371a138fd5", 00:23:23.186 "is_configured": true, 00:23:23.186 "data_offset": 2048, 00:23:23.186 "data_size": 63488 00:23:23.186 }, 00:23:23.186 { 00:23:23.186 "name": "BaseBdev2", 00:23:23.186 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:23.186 "is_configured": true, 00:23:23.186 "data_offset": 2048, 00:23:23.186 "data_size": 63488 00:23:23.186 } 00:23:23.186 ] 00:23:23.186 }' 00:23:23.186 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:23.186 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:23:23.186 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:23.445 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:23:23.445 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@665 -- # '[' true = true ']' 00:23:23.445 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@665 -- # '[' = false ']' 00:23:23.445 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev_raid.sh: line 665: [: =: unary operator expected 00:23:23.445 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@690 -- # local num_base_bdevs_operational=2 00:23:23.445 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@692 -- # '[' raid1 = raid1 ']' 00:23:23.445 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@692 -- # '[' 2 -gt 2 ']' 00:23:23.445 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@705 -- # local timeout=774 00:23:23.445 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:23:23.445 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:23:23.445 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:23.445 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:23:23.445 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=spare 00:23:23.445 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:23.445 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:23.445 11:34:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:23.703 11:34:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:23.703 "name": "raid_bdev1", 00:23:23.703 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:23.703 "strip_size_kb": 0, 00:23:23.703 "state": "online", 00:23:23.703 "raid_level": "raid1", 00:23:23.703 "superblock": true, 00:23:23.703 "num_base_bdevs": 2, 00:23:23.703 "num_base_bdevs_discovered": 2, 00:23:23.703 "num_base_bdevs_operational": 2, 00:23:23.703 "process": { 00:23:23.703 "type": "rebuild", 00:23:23.703 "target": "spare", 00:23:23.703 "progress": { 00:23:23.703 "blocks": 30720, 00:23:23.703 "percent": 48 00:23:23.703 } 00:23:23.703 }, 00:23:23.703 "base_bdevs_list": [ 00:23:23.703 { 00:23:23.703 "name": "spare", 00:23:23.703 "uuid": "e565947d-5b48-51dd-8764-c5371a138fd5", 00:23:23.703 "is_configured": true, 00:23:23.703 "data_offset": 2048, 00:23:23.703 "data_size": 63488 00:23:23.703 }, 00:23:23.703 { 00:23:23.703 "name": "BaseBdev2", 00:23:23.703 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:23.703 "is_configured": true, 00:23:23.703 "data_offset": 2048, 00:23:23.703 "data_size": 63488 00:23:23.703 } 00:23:23.703 ] 00:23:23.703 }' 00:23:23.703 11:34:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:23.703 11:34:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:23:23.703 11:34:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:23.703 11:34:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:23:23.703 11:34:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@710 -- # sleep 1 00:23:24.640 11:34:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:23:24.640 11:34:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:23:24.640 11:34:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:24.640 11:34:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:23:24.640 11:34:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=spare 00:23:24.640 11:34:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:24.640 11:34:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:24.640 11:34:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:24.898 11:34:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:24.898 "name": "raid_bdev1", 00:23:24.898 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:24.898 "strip_size_kb": 0, 00:23:24.898 "state": "online", 00:23:24.898 "raid_level": "raid1", 00:23:24.898 "superblock": true, 00:23:24.898 "num_base_bdevs": 2, 00:23:24.898 "num_base_bdevs_discovered": 2, 00:23:24.898 "num_base_bdevs_operational": 2, 00:23:24.898 "process": { 00:23:24.898 "type": "rebuild", 00:23:24.898 "target": "spare", 00:23:24.898 "progress": { 00:23:24.898 "blocks": 59392, 00:23:24.898 "percent": 93 00:23:24.898 } 00:23:24.898 }, 00:23:24.898 "base_bdevs_list": [ 00:23:24.898 { 00:23:24.898 "name": "spare", 00:23:24.898 "uuid": "e565947d-5b48-51dd-8764-c5371a138fd5", 00:23:24.898 "is_configured": true, 00:23:24.898 "data_offset": 2048, 00:23:24.898 "data_size": 63488 00:23:24.898 }, 00:23:24.898 { 00:23:24.898 "name": "BaseBdev2", 00:23:24.899 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:24.899 "is_configured": true, 00:23:24.899 "data_offset": 2048, 00:23:24.899 "data_size": 63488 00:23:24.899 } 00:23:24.899 ] 00:23:24.899 }' 00:23:24.899 11:34:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:24.899 11:34:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:23:24.899 11:34:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:24.899 11:34:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:23:24.899 11:34:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@710 -- # sleep 1 00:23:25.156 [2024-07-15 11:34:08.549797] bdev_raid.c:2789:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:23:25.156 [2024-07-15 11:34:08.549866] bdev_raid.c:2504:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:23:25.156 [2024-07-15 11:34:08.549954] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:26.096 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:23:26.096 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:23:26.096 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:26.096 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:23:26.096 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=spare 00:23:26.096 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:26.096 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:26.096 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:26.096 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:26.096 "name": "raid_bdev1", 00:23:26.096 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:26.096 "strip_size_kb": 0, 00:23:26.096 "state": "online", 00:23:26.096 "raid_level": "raid1", 00:23:26.096 "superblock": true, 00:23:26.096 "num_base_bdevs": 2, 00:23:26.096 "num_base_bdevs_discovered": 2, 00:23:26.096 "num_base_bdevs_operational": 2, 00:23:26.096 "base_bdevs_list": [ 00:23:26.096 { 00:23:26.096 "name": "spare", 00:23:26.096 "uuid": "e565947d-5b48-51dd-8764-c5371a138fd5", 00:23:26.096 "is_configured": true, 00:23:26.096 "data_offset": 2048, 00:23:26.096 "data_size": 63488 00:23:26.096 }, 00:23:26.096 { 00:23:26.096 "name": "BaseBdev2", 00:23:26.096 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:26.096 "is_configured": true, 00:23:26.096 "data_offset": 2048, 00:23:26.096 "data_size": 63488 00:23:26.096 } 00:23:26.096 ] 00:23:26.096 }' 00:23:26.096 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:26.355 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ none == \r\e\b\u\i\l\d ]] 00:23:26.355 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:26.355 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ none == \s\p\a\r\e ]] 00:23:26.355 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # break 00:23:26.355 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@714 -- # verify_raid_bdev_process raid_bdev1 none none 00:23:26.355 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:26.355 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:23:26.355 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=none 00:23:26.355 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:26.355 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:26.355 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:26.613 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:26.613 "name": "raid_bdev1", 00:23:26.613 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:26.613 "strip_size_kb": 0, 00:23:26.613 "state": "online", 00:23:26.613 "raid_level": "raid1", 00:23:26.613 "superblock": true, 00:23:26.613 "num_base_bdevs": 2, 00:23:26.613 "num_base_bdevs_discovered": 2, 00:23:26.613 "num_base_bdevs_operational": 2, 00:23:26.613 "base_bdevs_list": [ 00:23:26.613 { 00:23:26.613 "name": "spare", 00:23:26.613 "uuid": "e565947d-5b48-51dd-8764-c5371a138fd5", 00:23:26.613 "is_configured": true, 00:23:26.613 "data_offset": 2048, 00:23:26.613 "data_size": 63488 00:23:26.613 }, 00:23:26.613 { 00:23:26.613 "name": "BaseBdev2", 00:23:26.613 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:26.613 "is_configured": true, 00:23:26.613 "data_offset": 2048, 00:23:26.613 "data_size": 63488 00:23:26.613 } 00:23:26.613 ] 00:23:26.613 }' 00:23:26.613 11:34:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:26.613 11:34:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:23:26.613 11:34:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:26.613 11:34:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:23:26.613 11:34:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:23:26.613 11:34:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:23:26.613 11:34:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:23:26.613 11:34:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:23:26.613 11:34:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:23:26.613 11:34:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:23:26.613 11:34:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:23:26.613 11:34:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:23:26.613 11:34:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:23:26.613 11:34:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:23:26.613 11:34:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:26.613 11:34:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:26.872 11:34:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:23:26.872 "name": "raid_bdev1", 00:23:26.872 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:26.872 "strip_size_kb": 0, 00:23:26.872 "state": "online", 00:23:26.872 "raid_level": "raid1", 00:23:26.872 "superblock": true, 00:23:26.872 "num_base_bdevs": 2, 00:23:26.872 "num_base_bdevs_discovered": 2, 00:23:26.872 "num_base_bdevs_operational": 2, 00:23:26.872 "base_bdevs_list": [ 00:23:26.872 { 00:23:26.872 "name": "spare", 00:23:26.872 "uuid": "e565947d-5b48-51dd-8764-c5371a138fd5", 00:23:26.872 "is_configured": true, 00:23:26.872 "data_offset": 2048, 00:23:26.872 "data_size": 63488 00:23:26.872 }, 00:23:26.872 { 00:23:26.872 "name": "BaseBdev2", 00:23:26.872 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:26.872 "is_configured": true, 00:23:26.872 "data_offset": 2048, 00:23:26.872 "data_size": 63488 00:23:26.872 } 00:23:26.872 ] 00:23:26.872 }' 00:23:26.872 11:34:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:23:26.872 11:34:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:27.437 11:34:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@718 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:23:27.694 [2024-07-15 11:34:11.069479] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:23:27.694 [2024-07-15 11:34:11.069506] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:23:27.694 [2024-07-15 11:34:11.069571] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:23:27.694 [2024-07-15 11:34:11.069629] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:23:27.694 [2024-07-15 11:34:11.069641] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1296070 name raid_bdev1, state offline 00:23:27.694 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:27.694 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # jq length 00:23:27.951 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # [[ 0 == 0 ]] 00:23:27.951 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@721 -- # '[' true = true ']' 00:23:27.951 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' false = true ']' 00:23:27.951 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@736 -- # nbd_start_disks /var/tmp/spdk-raid.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:23:27.951 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:23:27.951 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:23:27.951 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:23:27.951 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:23:27.951 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:23:27.951 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:23:27.951 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:23:27.951 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:23:27.951 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:23:28.208 /dev/nbd0 00:23:28.208 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:23:28.208 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:23:28.208 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:23:28.208 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@867 -- # local i 00:23:28.208 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:23:28.208 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:23:28.208 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:23:28.208 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # break 00:23:28.208 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:23:28.208 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:23:28.208 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:23:28.208 1+0 records in 00:23:28.208 1+0 records out 00:23:28.208 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00026074 s, 15.7 MB/s 00:23:28.208 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:23:28.208 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # size=4096 00:23:28.208 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:23:28.208 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:23:28.208 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # return 0 00:23:28.208 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:23:28.208 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:23:28.208 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk spare /dev/nbd1 00:23:28.465 /dev/nbd1 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@867 -- # local i 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # break 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:23:28.465 1+0 records in 00:23:28.465 1+0 records out 00:23:28.465 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000276386 s, 14.8 MB/s 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # size=4096 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # return 0 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # nbd_stop_disks /var/tmp/spdk-raid.sock '/dev/nbd0 /dev/nbd1' 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:23:28.465 11:34:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd0 00:23:28.722 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:23:28.722 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:23:28.722 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:23:28.722 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:23:28.722 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:23:28.722 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:23:28.722 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:23:28.722 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:23:28.722 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:23:28.722 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd1 00:23:28.979 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:23:28.979 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:23:28.979 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:23:28.979 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:23:28.979 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:23:28.979 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:23:28.979 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:23:28.979 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:23:28.979 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@742 -- # '[' true = true ']' 00:23:28.979 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@744 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:23:29.236 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:23:29.493 [2024-07-15 11:34:12.877044] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:23:29.493 [2024-07-15 11:34:12.877090] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:29.493 [2024-07-15 11:34:12.877116] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1290670 00:23:29.493 [2024-07-15 11:34:12.877129] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:29.493 [2024-07-15 11:34:12.878779] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:29.493 [2024-07-15 11:34:12.878809] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:23:29.493 [2024-07-15 11:34:12.878898] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:23:29.493 [2024-07-15 11:34:12.878935] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:23:29.493 [2024-07-15 11:34:12.879044] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:23:29.493 spare 00:23:29.493 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:23:29.493 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:23:29.493 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:23:29.493 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:23:29.493 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:23:29.493 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:23:29.493 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:23:29.493 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:23:29.493 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:23:29.493 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:23:29.493 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:29.493 11:34:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:29.493 [2024-07-15 11:34:12.979359] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x10e4900 00:23:29.493 [2024-07-15 11:34:12.979379] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:23:29.493 [2024-07-15 11:34:12.979584] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x12983d0 00:23:29.493 [2024-07-15 11:34:12.979734] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x10e4900 00:23:29.493 [2024-07-15 11:34:12.979744] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x10e4900 00:23:29.493 [2024-07-15 11:34:12.979846] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:29.751 11:34:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:23:29.751 "name": "raid_bdev1", 00:23:29.751 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:29.751 "strip_size_kb": 0, 00:23:29.751 "state": "online", 00:23:29.751 "raid_level": "raid1", 00:23:29.751 "superblock": true, 00:23:29.751 "num_base_bdevs": 2, 00:23:29.751 "num_base_bdevs_discovered": 2, 00:23:29.751 "num_base_bdevs_operational": 2, 00:23:29.751 "base_bdevs_list": [ 00:23:29.751 { 00:23:29.751 "name": "spare", 00:23:29.751 "uuid": "e565947d-5b48-51dd-8764-c5371a138fd5", 00:23:29.751 "is_configured": true, 00:23:29.751 "data_offset": 2048, 00:23:29.751 "data_size": 63488 00:23:29.751 }, 00:23:29.751 { 00:23:29.751 "name": "BaseBdev2", 00:23:29.751 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:29.751 "is_configured": true, 00:23:29.751 "data_offset": 2048, 00:23:29.751 "data_size": 63488 00:23:29.751 } 00:23:29.751 ] 00:23:29.751 }' 00:23:29.751 11:34:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:23:29.751 11:34:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:30.314 11:34:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@748 -- # verify_raid_bdev_process raid_bdev1 none none 00:23:30.314 11:34:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:30.314 11:34:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:23:30.314 11:34:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=none 00:23:30.314 11:34:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:30.314 11:34:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:30.314 11:34:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:30.572 11:34:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:30.572 "name": "raid_bdev1", 00:23:30.572 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:30.572 "strip_size_kb": 0, 00:23:30.572 "state": "online", 00:23:30.572 "raid_level": "raid1", 00:23:30.572 "superblock": true, 00:23:30.572 "num_base_bdevs": 2, 00:23:30.572 "num_base_bdevs_discovered": 2, 00:23:30.572 "num_base_bdevs_operational": 2, 00:23:30.572 "base_bdevs_list": [ 00:23:30.572 { 00:23:30.572 "name": "spare", 00:23:30.572 "uuid": "e565947d-5b48-51dd-8764-c5371a138fd5", 00:23:30.572 "is_configured": true, 00:23:30.572 "data_offset": 2048, 00:23:30.572 "data_size": 63488 00:23:30.572 }, 00:23:30.572 { 00:23:30.572 "name": "BaseBdev2", 00:23:30.572 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:30.572 "is_configured": true, 00:23:30.572 "data_offset": 2048, 00:23:30.572 "data_size": 63488 00:23:30.572 } 00:23:30.572 ] 00:23:30.572 }' 00:23:30.572 11:34:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:30.572 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:23:30.572 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:30.572 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:23:30.572 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # jq -r '.[].base_bdevs_list[0].name' 00:23:30.572 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:30.829 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # [[ spare == \s\p\a\r\e ]] 00:23:30.829 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@752 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev spare 00:23:31.087 [2024-07-15 11:34:14.517504] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:23:31.087 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@753 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:23:31.087 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:23:31.087 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:23:31.087 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:23:31.087 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:23:31.087 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:23:31.087 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:23:31.087 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:23:31.087 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:23:31.087 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:23:31.087 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:31.087 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:31.345 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:23:31.345 "name": "raid_bdev1", 00:23:31.345 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:31.345 "strip_size_kb": 0, 00:23:31.345 "state": "online", 00:23:31.345 "raid_level": "raid1", 00:23:31.345 "superblock": true, 00:23:31.345 "num_base_bdevs": 2, 00:23:31.345 "num_base_bdevs_discovered": 1, 00:23:31.345 "num_base_bdevs_operational": 1, 00:23:31.345 "base_bdevs_list": [ 00:23:31.345 { 00:23:31.345 "name": null, 00:23:31.345 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:31.345 "is_configured": false, 00:23:31.345 "data_offset": 2048, 00:23:31.345 "data_size": 63488 00:23:31.345 }, 00:23:31.345 { 00:23:31.345 "name": "BaseBdev2", 00:23:31.345 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:31.345 "is_configured": true, 00:23:31.345 "data_offset": 2048, 00:23:31.345 "data_size": 63488 00:23:31.345 } 00:23:31.345 ] 00:23:31.345 }' 00:23:31.345 11:34:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:23:31.345 11:34:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:31.910 11:34:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:23:32.166 [2024-07-15 11:34:15.616585] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:23:32.166 [2024-07-15 11:34:15.616738] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:23:32.166 [2024-07-15 11:34:15.616755] bdev_raid.c:3620:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:23:32.166 [2024-07-15 11:34:15.616782] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:23:32.166 [2024-07-15 11:34:15.621662] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1295d00 00:23:32.166 [2024-07-15 11:34:15.623993] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:23:32.166 11:34:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # sleep 1 00:23:33.096 11:34:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:23:33.096 11:34:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:33.096 11:34:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:23:33.096 11:34:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=spare 00:23:33.096 11:34:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:33.096 11:34:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:33.096 11:34:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:33.353 11:34:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:33.353 "name": "raid_bdev1", 00:23:33.353 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:33.353 "strip_size_kb": 0, 00:23:33.353 "state": "online", 00:23:33.353 "raid_level": "raid1", 00:23:33.353 "superblock": true, 00:23:33.353 "num_base_bdevs": 2, 00:23:33.353 "num_base_bdevs_discovered": 2, 00:23:33.353 "num_base_bdevs_operational": 2, 00:23:33.353 "process": { 00:23:33.353 "type": "rebuild", 00:23:33.353 "target": "spare", 00:23:33.353 "progress": { 00:23:33.353 "blocks": 22528, 00:23:33.353 "percent": 35 00:23:33.353 } 00:23:33.353 }, 00:23:33.353 "base_bdevs_list": [ 00:23:33.353 { 00:23:33.353 "name": "spare", 00:23:33.353 "uuid": "e565947d-5b48-51dd-8764-c5371a138fd5", 00:23:33.353 "is_configured": true, 00:23:33.353 "data_offset": 2048, 00:23:33.353 "data_size": 63488 00:23:33.353 }, 00:23:33.353 { 00:23:33.353 "name": "BaseBdev2", 00:23:33.353 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:33.353 "is_configured": true, 00:23:33.353 "data_offset": 2048, 00:23:33.353 "data_size": 63488 00:23:33.353 } 00:23:33.353 ] 00:23:33.353 }' 00:23:33.353 11:34:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:33.353 11:34:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:23:33.353 11:34:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:33.353 11:34:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:23:33.353 11:34:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@759 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:23:33.610 [2024-07-15 11:34:17.126124] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:23:33.610 [2024-07-15 11:34:17.136010] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:23:33.610 [2024-07-15 11:34:17.136053] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:33.610 [2024-07-15 11:34:17.136068] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:23:33.610 [2024-07-15 11:34:17.136077] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:23:33.610 11:34:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@760 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:23:33.610 11:34:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:23:33.610 11:34:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:23:33.610 11:34:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:23:33.610 11:34:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:23:33.610 11:34:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:23:33.610 11:34:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:23:33.610 11:34:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:23:33.610 11:34:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:23:33.610 11:34:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:23:33.610 11:34:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:33.610 11:34:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:33.868 11:34:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:23:33.868 "name": "raid_bdev1", 00:23:33.868 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:33.868 "strip_size_kb": 0, 00:23:33.868 "state": "online", 00:23:33.868 "raid_level": "raid1", 00:23:33.868 "superblock": true, 00:23:33.868 "num_base_bdevs": 2, 00:23:33.868 "num_base_bdevs_discovered": 1, 00:23:33.868 "num_base_bdevs_operational": 1, 00:23:33.868 "base_bdevs_list": [ 00:23:33.868 { 00:23:33.868 "name": null, 00:23:33.868 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:33.868 "is_configured": false, 00:23:33.868 "data_offset": 2048, 00:23:33.868 "data_size": 63488 00:23:33.868 }, 00:23:33.868 { 00:23:33.868 "name": "BaseBdev2", 00:23:33.868 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:33.868 "is_configured": true, 00:23:33.868 "data_offset": 2048, 00:23:33.868 "data_size": 63488 00:23:33.868 } 00:23:33.868 ] 00:23:33.868 }' 00:23:33.868 11:34:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:23:33.868 11:34:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:34.435 11:34:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:23:34.703 [2024-07-15 11:34:18.236089] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:23:34.703 [2024-07-15 11:34:18.236143] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:34.703 [2024-07-15 11:34:18.236166] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x10e4c80 00:23:34.703 [2024-07-15 11:34:18.236178] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:34.703 [2024-07-15 11:34:18.236552] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:34.703 [2024-07-15 11:34:18.236570] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:23:34.703 [2024-07-15 11:34:18.236652] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:23:34.703 [2024-07-15 11:34:18.236665] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:23:34.703 [2024-07-15 11:34:18.236676] bdev_raid.c:3620:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:23:34.703 [2024-07-15 11:34:18.236694] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:23:34.703 [2024-07-15 11:34:18.241555] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1295d00 00:23:34.703 spare 00:23:34.703 [2024-07-15 11:34:18.243018] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:23:34.703 11:34:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # sleep 1 00:23:36.093 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:23:36.093 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:36.093 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:23:36.093 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=spare 00:23:36.093 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:36.093 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:36.093 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:36.093 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:36.093 "name": "raid_bdev1", 00:23:36.093 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:36.093 "strip_size_kb": 0, 00:23:36.093 "state": "online", 00:23:36.093 "raid_level": "raid1", 00:23:36.093 "superblock": true, 00:23:36.093 "num_base_bdevs": 2, 00:23:36.093 "num_base_bdevs_discovered": 2, 00:23:36.093 "num_base_bdevs_operational": 2, 00:23:36.093 "process": { 00:23:36.093 "type": "rebuild", 00:23:36.093 "target": "spare", 00:23:36.093 "progress": { 00:23:36.093 "blocks": 24576, 00:23:36.093 "percent": 38 00:23:36.093 } 00:23:36.093 }, 00:23:36.093 "base_bdevs_list": [ 00:23:36.093 { 00:23:36.093 "name": "spare", 00:23:36.093 "uuid": "e565947d-5b48-51dd-8764-c5371a138fd5", 00:23:36.093 "is_configured": true, 00:23:36.093 "data_offset": 2048, 00:23:36.093 "data_size": 63488 00:23:36.093 }, 00:23:36.093 { 00:23:36.093 "name": "BaseBdev2", 00:23:36.093 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:36.093 "is_configured": true, 00:23:36.093 "data_offset": 2048, 00:23:36.093 "data_size": 63488 00:23:36.093 } 00:23:36.093 ] 00:23:36.093 }' 00:23:36.093 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:36.093 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:23:36.093 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:36.093 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:23:36.093 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@766 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:23:36.350 [2024-07-15 11:34:19.822440] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:23:36.350 [2024-07-15 11:34:19.855994] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:23:36.350 [2024-07-15 11:34:19.856041] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:36.350 [2024-07-15 11:34:19.856056] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:23:36.350 [2024-07-15 11:34:19.856065] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:23:36.350 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@767 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:23:36.350 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:23:36.350 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:23:36.350 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:23:36.350 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:23:36.350 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:23:36.350 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:23:36.350 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:23:36.350 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:23:36.350 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:23:36.350 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:36.350 11:34:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:36.607 11:34:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:23:36.607 "name": "raid_bdev1", 00:23:36.607 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:36.607 "strip_size_kb": 0, 00:23:36.607 "state": "online", 00:23:36.607 "raid_level": "raid1", 00:23:36.607 "superblock": true, 00:23:36.607 "num_base_bdevs": 2, 00:23:36.607 "num_base_bdevs_discovered": 1, 00:23:36.607 "num_base_bdevs_operational": 1, 00:23:36.607 "base_bdevs_list": [ 00:23:36.607 { 00:23:36.607 "name": null, 00:23:36.607 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:36.607 "is_configured": false, 00:23:36.607 "data_offset": 2048, 00:23:36.607 "data_size": 63488 00:23:36.607 }, 00:23:36.607 { 00:23:36.607 "name": "BaseBdev2", 00:23:36.607 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:36.607 "is_configured": true, 00:23:36.607 "data_offset": 2048, 00:23:36.607 "data_size": 63488 00:23:36.607 } 00:23:36.607 ] 00:23:36.607 }' 00:23:36.607 11:34:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:23:36.607 11:34:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:37.169 11:34:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # verify_raid_bdev_process raid_bdev1 none none 00:23:37.169 11:34:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:37.169 11:34:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:23:37.169 11:34:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=none 00:23:37.169 11:34:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:37.169 11:34:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:37.169 11:34:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:37.425 11:34:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:37.425 "name": "raid_bdev1", 00:23:37.425 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:37.425 "strip_size_kb": 0, 00:23:37.425 "state": "online", 00:23:37.425 "raid_level": "raid1", 00:23:37.425 "superblock": true, 00:23:37.425 "num_base_bdevs": 2, 00:23:37.425 "num_base_bdevs_discovered": 1, 00:23:37.425 "num_base_bdevs_operational": 1, 00:23:37.425 "base_bdevs_list": [ 00:23:37.425 { 00:23:37.425 "name": null, 00:23:37.425 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:37.425 "is_configured": false, 00:23:37.425 "data_offset": 2048, 00:23:37.425 "data_size": 63488 00:23:37.425 }, 00:23:37.425 { 00:23:37.425 "name": "BaseBdev2", 00:23:37.425 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:37.425 "is_configured": true, 00:23:37.425 "data_offset": 2048, 00:23:37.425 "data_size": 63488 00:23:37.425 } 00:23:37.425 ] 00:23:37.425 }' 00:23:37.425 11:34:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:37.681 11:34:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:23:37.681 11:34:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:37.681 11:34:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:23:37.681 11:34:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@771 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete BaseBdev1 00:23:37.937 11:34:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@772 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:23:38.194 [2024-07-15 11:34:21.537592] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:23:38.194 [2024-07-15 11:34:21.537644] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:38.194 [2024-07-15 11:34:21.537666] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1290190 00:23:38.194 [2024-07-15 11:34:21.537679] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:38.194 [2024-07-15 11:34:21.538048] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:38.194 [2024-07-15 11:34:21.538068] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:23:38.194 [2024-07-15 11:34:21.538139] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:23:38.194 [2024-07-15 11:34:21.538152] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:23:38.194 [2024-07-15 11:34:21.538162] bdev_raid.c:3581:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:23:38.194 BaseBdev1 00:23:38.194 11:34:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # sleep 1 00:23:39.123 11:34:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:23:39.123 11:34:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:23:39.123 11:34:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:23:39.123 11:34:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:23:39.123 11:34:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:23:39.123 11:34:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:23:39.123 11:34:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:23:39.123 11:34:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:23:39.123 11:34:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:23:39.123 11:34:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:23:39.123 11:34:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:39.123 11:34:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:39.380 11:34:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:23:39.380 "name": "raid_bdev1", 00:23:39.380 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:39.380 "strip_size_kb": 0, 00:23:39.380 "state": "online", 00:23:39.380 "raid_level": "raid1", 00:23:39.380 "superblock": true, 00:23:39.380 "num_base_bdevs": 2, 00:23:39.380 "num_base_bdevs_discovered": 1, 00:23:39.380 "num_base_bdevs_operational": 1, 00:23:39.380 "base_bdevs_list": [ 00:23:39.380 { 00:23:39.380 "name": null, 00:23:39.380 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:39.380 "is_configured": false, 00:23:39.380 "data_offset": 2048, 00:23:39.380 "data_size": 63488 00:23:39.380 }, 00:23:39.380 { 00:23:39.380 "name": "BaseBdev2", 00:23:39.380 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:39.380 "is_configured": true, 00:23:39.380 "data_offset": 2048, 00:23:39.380 "data_size": 63488 00:23:39.380 } 00:23:39.380 ] 00:23:39.380 }' 00:23:39.380 11:34:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:23:39.380 11:34:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:39.942 11:34:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # verify_raid_bdev_process raid_bdev1 none none 00:23:39.942 11:34:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:39.942 11:34:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:23:39.942 11:34:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=none 00:23:39.942 11:34:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:39.942 11:34:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:39.942 11:34:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:40.199 11:34:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:40.199 "name": "raid_bdev1", 00:23:40.199 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:40.199 "strip_size_kb": 0, 00:23:40.199 "state": "online", 00:23:40.199 "raid_level": "raid1", 00:23:40.199 "superblock": true, 00:23:40.199 "num_base_bdevs": 2, 00:23:40.199 "num_base_bdevs_discovered": 1, 00:23:40.199 "num_base_bdevs_operational": 1, 00:23:40.199 "base_bdevs_list": [ 00:23:40.199 { 00:23:40.199 "name": null, 00:23:40.199 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:40.199 "is_configured": false, 00:23:40.199 "data_offset": 2048, 00:23:40.199 "data_size": 63488 00:23:40.199 }, 00:23:40.199 { 00:23:40.199 "name": "BaseBdev2", 00:23:40.199 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:40.199 "is_configured": true, 00:23:40.199 "data_offset": 2048, 00:23:40.199 "data_size": 63488 00:23:40.199 } 00:23:40.199 ] 00:23:40.199 }' 00:23:40.199 11:34:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:40.199 11:34:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:23:40.199 11:34:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:40.199 11:34:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:23:40.199 11:34:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:23:40.199 11:34:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@648 -- # local es=0 00:23:40.199 11:34:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:23:40.199 11:34:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:23:40.456 11:34:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:23:40.456 11:34:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:23:40.456 11:34:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:23:40.456 11:34:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:23:40.456 11:34:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:23:40.456 11:34:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:23:40.456 11:34:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py ]] 00:23:40.456 11:34:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:23:40.456 [2024-07-15 11:34:24.024255] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:23:40.456 [2024-07-15 11:34:24.024374] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:23:40.456 [2024-07-15 11:34:24.024390] bdev_raid.c:3581:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:23:40.456 request: 00:23:40.456 { 00:23:40.456 "base_bdev": "BaseBdev1", 00:23:40.456 "raid_bdev": "raid_bdev1", 00:23:40.456 "method": "bdev_raid_add_base_bdev", 00:23:40.456 "req_id": 1 00:23:40.456 } 00:23:40.456 Got JSON-RPC error response 00:23:40.456 response: 00:23:40.456 { 00:23:40.456 "code": -22, 00:23:40.456 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:23:40.456 } 00:23:40.456 11:34:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@651 -- # es=1 00:23:40.456 11:34:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:23:40.456 11:34:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:23:40.456 11:34:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:23:40.456 11:34:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # sleep 1 00:23:41.829 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:23:41.829 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:23:41.829 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:23:41.829 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:23:41.829 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:23:41.829 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:23:41.829 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:23:41.829 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:23:41.829 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:23:41.829 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:23:41.829 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:41.829 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:41.829 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:23:41.829 "name": "raid_bdev1", 00:23:41.829 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:41.829 "strip_size_kb": 0, 00:23:41.829 "state": "online", 00:23:41.829 "raid_level": "raid1", 00:23:41.829 "superblock": true, 00:23:41.829 "num_base_bdevs": 2, 00:23:41.829 "num_base_bdevs_discovered": 1, 00:23:41.829 "num_base_bdevs_operational": 1, 00:23:41.829 "base_bdevs_list": [ 00:23:41.829 { 00:23:41.829 "name": null, 00:23:41.829 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:41.829 "is_configured": false, 00:23:41.829 "data_offset": 2048, 00:23:41.829 "data_size": 63488 00:23:41.829 }, 00:23:41.829 { 00:23:41.829 "name": "BaseBdev2", 00:23:41.829 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:41.829 "is_configured": true, 00:23:41.829 "data_offset": 2048, 00:23:41.829 "data_size": 63488 00:23:41.829 } 00:23:41.829 ] 00:23:41.829 }' 00:23:41.829 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:23:41.829 11:34:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:42.396 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # verify_raid_bdev_process raid_bdev1 none none 00:23:42.396 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:42.396 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:23:42.396 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=none 00:23:42.396 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:42.396 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:42.396 11:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:42.652 11:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:42.652 "name": "raid_bdev1", 00:23:42.652 "uuid": "9a690ca4-7bc0-42bf-917c-9acb5a292c74", 00:23:42.652 "strip_size_kb": 0, 00:23:42.652 "state": "online", 00:23:42.652 "raid_level": "raid1", 00:23:42.652 "superblock": true, 00:23:42.652 "num_base_bdevs": 2, 00:23:42.652 "num_base_bdevs_discovered": 1, 00:23:42.652 "num_base_bdevs_operational": 1, 00:23:42.652 "base_bdevs_list": [ 00:23:42.652 { 00:23:42.652 "name": null, 00:23:42.652 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:42.652 "is_configured": false, 00:23:42.652 "data_offset": 2048, 00:23:42.652 "data_size": 63488 00:23:42.652 }, 00:23:42.652 { 00:23:42.653 "name": "BaseBdev2", 00:23:42.653 "uuid": "d3e5fbe5-7621-5cba-8e9d-79daae41c937", 00:23:42.653 "is_configured": true, 00:23:42.653 "data_offset": 2048, 00:23:42.653 "data_size": 63488 00:23:42.653 } 00:23:42.653 ] 00:23:42.653 }' 00:23:42.653 11:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:42.653 11:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:23:42.653 11:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:42.653 11:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:23:42.653 11:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@782 -- # killprocess 971434 00:23:42.653 11:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@948 -- # '[' -z 971434 ']' 00:23:42.653 11:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@952 -- # kill -0 971434 00:23:42.653 11:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@953 -- # uname 00:23:42.653 11:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:23:42.653 11:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 971434 00:23:42.911 11:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:23:42.911 11:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:23:42.911 11:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@966 -- # echo 'killing process with pid 971434' 00:23:42.911 killing process with pid 971434 00:23:42.911 11:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@967 -- # kill 971434 00:23:42.911 Received shutdown signal, test time was about 60.000000 seconds 00:23:42.911 00:23:42.911 Latency(us) 00:23:42.911 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:23:42.911 =================================================================================================================== 00:23:42.911 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:23:42.911 [2024-07-15 11:34:26.287464] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:23:42.911 [2024-07-15 11:34:26.287566] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:23:42.911 11:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@972 -- # wait 971434 00:23:42.911 [2024-07-15 11:34:26.287612] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:23:42.911 [2024-07-15 11:34:26.287627] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x10e4900 name raid_bdev1, state offline 00:23:42.911 [2024-07-15 11:34:26.318029] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # return 0 00:23:43.170 00:23:43.170 real 0m35.930s 00:23:43.170 user 0m52.025s 00:23:43.170 sys 0m6.749s 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1124 -- # xtrace_disable 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:43.170 ************************************ 00:23:43.170 END TEST raid_rebuild_test_sb 00:23:43.170 ************************************ 00:23:43.170 11:34:26 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:23:43.170 11:34:26 bdev_raid -- bdev/bdev_raid.sh@879 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 2 false true true 00:23:43.170 11:34:26 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:23:43.170 11:34:26 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:23:43.170 11:34:26 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:23:43.170 ************************************ 00:23:43.170 START TEST raid_rebuild_test_io 00:23:43.170 ************************************ 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1123 -- # raid_rebuild_test raid1 2 false true true 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@568 -- # local raid_level=raid1 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local num_base_bdevs=2 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local superblock=false 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local background_io=true 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local verify=true 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # (( i = 1 )) 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # echo BaseBdev1 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # echo BaseBdev2 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local base_bdevs 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local raid_bdev_name=raid_bdev1 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local strip_size 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local create_arg 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local raid_bdev_size 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local data_offset 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@580 -- # '[' raid1 '!=' raid1 ']' 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@588 -- # strip_size=0 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@591 -- # '[' false = true ']' 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # raid_pid=976615 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # waitforlisten 976615 /var/tmp/spdk-raid.sock 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@595 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@829 -- # '[' -z 976615 ']' 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@834 -- # local max_retries=100 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:23:43.170 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@838 -- # xtrace_disable 00:23:43.170 11:34:26 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:23:43.170 [2024-07-15 11:34:26.709253] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:23:43.170 [2024-07-15 11:34:26.709324] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid976615 ] 00:23:43.170 I/O size of 3145728 is greater than zero copy threshold (65536). 00:23:43.170 Zero copy mechanism will not be used. 00:23:43.485 [2024-07-15 11:34:26.837178] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:43.485 [2024-07-15 11:34:26.939767] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:23:43.485 [2024-07-15 11:34:27.001294] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:23:43.485 [2024-07-15 11:34:27.001351] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:23:44.058 11:34:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:23:44.058 11:34:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@862 -- # return 0 00:23:44.058 11:34:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:23:44.058 11:34:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:23:44.315 BaseBdev1_malloc 00:23:44.315 11:34:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:23:44.573 [2024-07-15 11:34:28.116759] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:23:44.573 [2024-07-15 11:34:28.116809] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:44.573 [2024-07-15 11:34:28.116831] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1a7bd40 00:23:44.573 [2024-07-15 11:34:28.116843] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:44.573 [2024-07-15 11:34:28.118418] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:44.573 [2024-07-15 11:34:28.118446] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:23:44.573 BaseBdev1 00:23:44.573 11:34:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:23:44.573 11:34:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:23:44.832 BaseBdev2_malloc 00:23:44.832 11:34:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:23:45.089 [2024-07-15 11:34:28.598779] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:23:45.089 [2024-07-15 11:34:28.598822] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:45.089 [2024-07-15 11:34:28.598842] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1a7c860 00:23:45.089 [2024-07-15 11:34:28.598855] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:45.089 [2024-07-15 11:34:28.600230] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:45.089 [2024-07-15 11:34:28.600255] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:23:45.089 BaseBdev2 00:23:45.089 11:34:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@606 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b spare_malloc 00:23:45.347 spare_malloc 00:23:45.347 11:34:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:23:45.605 spare_delay 00:23:45.605 11:34:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:23:45.863 [2024-07-15 11:34:29.337189] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:23:45.863 [2024-07-15 11:34:29.337233] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:45.863 [2024-07-15 11:34:29.337251] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1c2aec0 00:23:45.863 [2024-07-15 11:34:29.337264] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:45.863 [2024-07-15 11:34:29.338676] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:45.863 [2024-07-15 11:34:29.338702] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:23:45.863 spare 00:23:45.863 11:34:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@611 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2' -n raid_bdev1 00:23:46.122 [2024-07-15 11:34:29.581849] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:23:46.122 [2024-07-15 11:34:29.583044] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:23:46.122 [2024-07-15 11:34:29.583117] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1c2c070 00:23:46.122 [2024-07-15 11:34:29.583128] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:23:46.122 [2024-07-15 11:34:29.583322] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1c25490 00:23:46.122 [2024-07-15 11:34:29.583458] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1c2c070 00:23:46.122 [2024-07-15 11:34:29.583468] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x1c2c070 00:23:46.122 [2024-07-15 11:34:29.583572] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:46.122 11:34:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:23:46.122 11:34:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:23:46.122 11:34:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:23:46.122 11:34:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:23:46.122 11:34:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:23:46.122 11:34:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:23:46.122 11:34:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:23:46.122 11:34:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:23:46.122 11:34:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:23:46.122 11:34:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:23:46.122 11:34:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:46.122 11:34:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:46.380 11:34:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:23:46.380 "name": "raid_bdev1", 00:23:46.380 "uuid": "1746909c-81cc-4286-a264-8fe6432320b4", 00:23:46.380 "strip_size_kb": 0, 00:23:46.380 "state": "online", 00:23:46.380 "raid_level": "raid1", 00:23:46.380 "superblock": false, 00:23:46.380 "num_base_bdevs": 2, 00:23:46.380 "num_base_bdevs_discovered": 2, 00:23:46.380 "num_base_bdevs_operational": 2, 00:23:46.380 "base_bdevs_list": [ 00:23:46.380 { 00:23:46.380 "name": "BaseBdev1", 00:23:46.380 "uuid": "93037170-f96a-5a45-984a-5f0986d567ff", 00:23:46.380 "is_configured": true, 00:23:46.380 "data_offset": 0, 00:23:46.380 "data_size": 65536 00:23:46.380 }, 00:23:46.380 { 00:23:46.380 "name": "BaseBdev2", 00:23:46.380 "uuid": "b87ddd44-5920-53c0-8456-f91c0a57ceb6", 00:23:46.380 "is_configured": true, 00:23:46.380 "data_offset": 0, 00:23:46.380 "data_size": 65536 00:23:46.380 } 00:23:46.380 ] 00:23:46.381 }' 00:23:46.381 11:34:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:23:46.381 11:34:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:23:46.946 11:34:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@615 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:23:46.946 11:34:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@615 -- # jq -r '.[].num_blocks' 00:23:47.204 [2024-07-15 11:34:30.705060] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:23:47.204 11:34:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@615 -- # raid_bdev_size=65536 00:23:47.204 11:34:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@618 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:47.204 11:34:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@618 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:23:47.462 11:34:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@618 -- # data_offset=0 00:23:47.462 11:34:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@620 -- # '[' true = true ']' 00:23:47.462 11:34:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@639 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev1 00:23:47.462 11:34:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@622 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:23:47.721 [2024-07-15 11:34:31.079871] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1c26bd0 00:23:47.721 I/O size of 3145728 is greater than zero copy threshold (65536). 00:23:47.721 Zero copy mechanism will not be used. 00:23:47.721 Running I/O for 60 seconds... 00:23:47.721 [2024-07-15 11:34:31.203229] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:23:47.721 [2024-07-15 11:34:31.219362] bdev_raid.c:1919:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x1c26bd0 00:23:47.721 11:34:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@642 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:23:47.721 11:34:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:23:47.721 11:34:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:23:47.721 11:34:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:23:47.721 11:34:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:23:47.721 11:34:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:23:47.721 11:34:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:23:47.721 11:34:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:23:47.721 11:34:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:23:47.721 11:34:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:23:47.721 11:34:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:47.721 11:34:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:47.979 11:34:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:23:47.980 "name": "raid_bdev1", 00:23:47.980 "uuid": "1746909c-81cc-4286-a264-8fe6432320b4", 00:23:47.980 "strip_size_kb": 0, 00:23:47.980 "state": "online", 00:23:47.980 "raid_level": "raid1", 00:23:47.980 "superblock": false, 00:23:47.980 "num_base_bdevs": 2, 00:23:47.980 "num_base_bdevs_discovered": 1, 00:23:47.980 "num_base_bdevs_operational": 1, 00:23:47.980 "base_bdevs_list": [ 00:23:47.980 { 00:23:47.980 "name": null, 00:23:47.980 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:47.980 "is_configured": false, 00:23:47.980 "data_offset": 0, 00:23:47.980 "data_size": 65536 00:23:47.980 }, 00:23:47.980 { 00:23:47.980 "name": "BaseBdev2", 00:23:47.980 "uuid": "b87ddd44-5920-53c0-8456-f91c0a57ceb6", 00:23:47.980 "is_configured": true, 00:23:47.980 "data_offset": 0, 00:23:47.980 "data_size": 65536 00:23:47.980 } 00:23:47.980 ] 00:23:47.980 }' 00:23:47.980 11:34:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:23:47.980 11:34:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:23:48.914 11:34:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@645 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:23:48.914 [2024-07-15 11:34:32.377532] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:23:48.914 11:34:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # sleep 1 00:23:48.914 [2024-07-15 11:34:32.460385] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1bae8b0 00:23:48.914 [2024-07-15 11:34:32.462780] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:23:49.172 [2024-07-15 11:34:32.589614] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:23:49.172 [2024-07-15 11:34:32.590017] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:23:49.430 [2024-07-15 11:34:32.801373] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:23:49.430 [2024-07-15 11:34:32.801536] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:23:49.689 [2024-07-15 11:34:33.116648] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:23:49.689 [2024-07-15 11:34:33.264740] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:23:49.945 11:34:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@649 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:23:49.945 11:34:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:49.945 11:34:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:23:49.945 11:34:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:23:49.945 11:34:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:49.945 11:34:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:49.945 11:34:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:50.203 [2024-07-15 11:34:33.620732] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:23:50.203 11:34:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:50.203 "name": "raid_bdev1", 00:23:50.203 "uuid": "1746909c-81cc-4286-a264-8fe6432320b4", 00:23:50.203 "strip_size_kb": 0, 00:23:50.203 "state": "online", 00:23:50.203 "raid_level": "raid1", 00:23:50.203 "superblock": false, 00:23:50.203 "num_base_bdevs": 2, 00:23:50.203 "num_base_bdevs_discovered": 2, 00:23:50.203 "num_base_bdevs_operational": 2, 00:23:50.203 "process": { 00:23:50.203 "type": "rebuild", 00:23:50.203 "target": "spare", 00:23:50.203 "progress": { 00:23:50.203 "blocks": 14336, 00:23:50.203 "percent": 21 00:23:50.203 } 00:23:50.203 }, 00:23:50.203 "base_bdevs_list": [ 00:23:50.203 { 00:23:50.203 "name": "spare", 00:23:50.203 "uuid": "cbfa7db1-5755-5490-ae84-8c0d9d2f7ef7", 00:23:50.203 "is_configured": true, 00:23:50.203 "data_offset": 0, 00:23:50.203 "data_size": 65536 00:23:50.203 }, 00:23:50.203 { 00:23:50.203 "name": "BaseBdev2", 00:23:50.203 "uuid": "b87ddd44-5920-53c0-8456-f91c0a57ceb6", 00:23:50.203 "is_configured": true, 00:23:50.203 "data_offset": 0, 00:23:50.203 "data_size": 65536 00:23:50.203 } 00:23:50.203 ] 00:23:50.203 }' 00:23:50.203 11:34:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:50.203 11:34:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:23:50.203 11:34:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:50.203 11:34:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:23:50.203 11:34:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@652 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev spare 00:23:50.461 [2024-07-15 11:34:33.840553] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:23:50.718 [2024-07-15 11:34:34.171722] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:23:50.719 [2024-07-15 11:34:34.272846] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:23:50.976 [2024-07-15 11:34:34.328978] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:23:50.976 [2024-07-15 11:34:34.338989] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:50.976 [2024-07-15 11:34:34.339017] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:23:50.976 [2024-07-15 11:34:34.339029] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:23:50.976 [2024-07-15 11:34:34.360735] bdev_raid.c:1919:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x1c26bd0 00:23:50.976 11:34:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@655 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:23:50.976 11:34:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:23:50.976 11:34:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:23:50.976 11:34:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:23:50.976 11:34:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:23:50.976 11:34:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:23:50.976 11:34:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:23:50.976 11:34:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:23:50.976 11:34:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:23:50.976 11:34:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:23:50.976 11:34:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:50.976 11:34:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:51.234 11:34:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:23:51.234 "name": "raid_bdev1", 00:23:51.234 "uuid": "1746909c-81cc-4286-a264-8fe6432320b4", 00:23:51.234 "strip_size_kb": 0, 00:23:51.234 "state": "online", 00:23:51.234 "raid_level": "raid1", 00:23:51.234 "superblock": false, 00:23:51.234 "num_base_bdevs": 2, 00:23:51.234 "num_base_bdevs_discovered": 1, 00:23:51.234 "num_base_bdevs_operational": 1, 00:23:51.234 "base_bdevs_list": [ 00:23:51.234 { 00:23:51.234 "name": null, 00:23:51.234 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:51.234 "is_configured": false, 00:23:51.234 "data_offset": 0, 00:23:51.234 "data_size": 65536 00:23:51.234 }, 00:23:51.234 { 00:23:51.234 "name": "BaseBdev2", 00:23:51.234 "uuid": "b87ddd44-5920-53c0-8456-f91c0a57ceb6", 00:23:51.234 "is_configured": true, 00:23:51.234 "data_offset": 0, 00:23:51.234 "data_size": 65536 00:23:51.234 } 00:23:51.234 ] 00:23:51.234 }' 00:23:51.234 11:34:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:23:51.234 11:34:34 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:23:51.802 11:34:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@658 -- # verify_raid_bdev_process raid_bdev1 none none 00:23:51.802 11:34:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:51.802 11:34:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:23:51.802 11:34:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@184 -- # local target=none 00:23:51.802 11:34:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:51.802 11:34:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:51.802 11:34:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:52.109 11:34:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:52.109 "name": "raid_bdev1", 00:23:52.109 "uuid": "1746909c-81cc-4286-a264-8fe6432320b4", 00:23:52.109 "strip_size_kb": 0, 00:23:52.109 "state": "online", 00:23:52.109 "raid_level": "raid1", 00:23:52.109 "superblock": false, 00:23:52.109 "num_base_bdevs": 2, 00:23:52.109 "num_base_bdevs_discovered": 1, 00:23:52.109 "num_base_bdevs_operational": 1, 00:23:52.109 "base_bdevs_list": [ 00:23:52.109 { 00:23:52.109 "name": null, 00:23:52.109 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:52.109 "is_configured": false, 00:23:52.109 "data_offset": 0, 00:23:52.109 "data_size": 65536 00:23:52.109 }, 00:23:52.109 { 00:23:52.109 "name": "BaseBdev2", 00:23:52.109 "uuid": "b87ddd44-5920-53c0-8456-f91c0a57ceb6", 00:23:52.109 "is_configured": true, 00:23:52.109 "data_offset": 0, 00:23:52.109 "data_size": 65536 00:23:52.109 } 00:23:52.109 ] 00:23:52.109 }' 00:23:52.109 11:34:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:52.109 11:34:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:23:52.109 11:34:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:52.109 11:34:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:23:52.109 11:34:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@661 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:23:52.370 [2024-07-15 11:34:35.868795] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:23:52.370 11:34:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # sleep 1 00:23:52.370 [2024-07-15 11:34:35.929606] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1c2c450 00:23:52.370 [2024-07-15 11:34:35.931180] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:23:52.628 [2024-07-15 11:34:36.049585] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:23:52.628 [2024-07-15 11:34:36.050020] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:23:52.886 [2024-07-15 11:34:36.253238] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:23:52.886 [2024-07-15 11:34:36.253505] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:23:53.144 [2024-07-15 11:34:36.584236] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:23:53.144 [2024-07-15 11:34:36.584512] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:23:53.402 [2024-07-15 11:34:36.797023] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:23:53.402 11:34:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:23:53.402 11:34:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:53.402 11:34:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:23:53.402 11:34:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:23:53.402 11:34:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:53.402 11:34:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:53.402 11:34:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:53.660 [2024-07-15 11:34:37.034190] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:23:53.660 [2024-07-15 11:34:37.034683] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:23:53.660 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:53.660 "name": "raid_bdev1", 00:23:53.660 "uuid": "1746909c-81cc-4286-a264-8fe6432320b4", 00:23:53.660 "strip_size_kb": 0, 00:23:53.660 "state": "online", 00:23:53.660 "raid_level": "raid1", 00:23:53.660 "superblock": false, 00:23:53.660 "num_base_bdevs": 2, 00:23:53.660 "num_base_bdevs_discovered": 2, 00:23:53.660 "num_base_bdevs_operational": 2, 00:23:53.660 "process": { 00:23:53.660 "type": "rebuild", 00:23:53.660 "target": "spare", 00:23:53.660 "progress": { 00:23:53.660 "blocks": 14336, 00:23:53.660 "percent": 21 00:23:53.660 } 00:23:53.660 }, 00:23:53.660 "base_bdevs_list": [ 00:23:53.660 { 00:23:53.660 "name": "spare", 00:23:53.660 "uuid": "cbfa7db1-5755-5490-ae84-8c0d9d2f7ef7", 00:23:53.660 "is_configured": true, 00:23:53.660 "data_offset": 0, 00:23:53.660 "data_size": 65536 00:23:53.660 }, 00:23:53.660 { 00:23:53.660 "name": "BaseBdev2", 00:23:53.660 "uuid": "b87ddd44-5920-53c0-8456-f91c0a57ceb6", 00:23:53.660 "is_configured": true, 00:23:53.660 "data_offset": 0, 00:23:53.660 "data_size": 65536 00:23:53.660 } 00:23:53.660 ] 00:23:53.660 }' 00:23:53.660 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:53.660 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:23:53.660 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:53.918 [2024-07-15 11:34:37.255255] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:23:53.918 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:23:53.918 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@665 -- # '[' false = true ']' 00:23:53.918 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@690 -- # local num_base_bdevs_operational=2 00:23:53.918 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@692 -- # '[' raid1 = raid1 ']' 00:23:53.918 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@692 -- # '[' 2 -gt 2 ']' 00:23:53.918 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@705 -- # local timeout=805 00:23:53.918 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:23:53.918 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:23:53.918 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:53.918 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:23:53.918 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:23:53.918 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:53.918 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:53.918 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:53.918 [2024-07-15 11:34:37.488852] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:23:54.177 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:54.177 "name": "raid_bdev1", 00:23:54.177 "uuid": "1746909c-81cc-4286-a264-8fe6432320b4", 00:23:54.177 "strip_size_kb": 0, 00:23:54.177 "state": "online", 00:23:54.177 "raid_level": "raid1", 00:23:54.177 "superblock": false, 00:23:54.177 "num_base_bdevs": 2, 00:23:54.177 "num_base_bdevs_discovered": 2, 00:23:54.177 "num_base_bdevs_operational": 2, 00:23:54.177 "process": { 00:23:54.177 "type": "rebuild", 00:23:54.177 "target": "spare", 00:23:54.177 "progress": { 00:23:54.177 "blocks": 20480, 00:23:54.177 "percent": 31 00:23:54.177 } 00:23:54.177 }, 00:23:54.177 "base_bdevs_list": [ 00:23:54.177 { 00:23:54.177 "name": "spare", 00:23:54.177 "uuid": "cbfa7db1-5755-5490-ae84-8c0d9d2f7ef7", 00:23:54.177 "is_configured": true, 00:23:54.177 "data_offset": 0, 00:23:54.177 "data_size": 65536 00:23:54.177 }, 00:23:54.177 { 00:23:54.177 "name": "BaseBdev2", 00:23:54.177 "uuid": "b87ddd44-5920-53c0-8456-f91c0a57ceb6", 00:23:54.177 "is_configured": true, 00:23:54.177 "data_offset": 0, 00:23:54.177 "data_size": 65536 00:23:54.177 } 00:23:54.177 ] 00:23:54.177 }' 00:23:54.177 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:54.177 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:23:54.177 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:54.177 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:23:54.177 11:34:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@710 -- # sleep 1 00:23:54.177 [2024-07-15 11:34:37.616103] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:23:54.177 [2024-07-15 11:34:37.616348] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:23:54.435 [2024-07-15 11:34:37.957891] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:23:55.000 [2024-07-15 11:34:38.322033] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:23:55.000 [2024-07-15 11:34:38.322403] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:23:55.000 [2024-07-15 11:34:38.540735] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:23:55.000 [2024-07-15 11:34:38.540867] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:23:55.257 11:34:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:23:55.257 11:34:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:23:55.257 11:34:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:55.257 11:34:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:23:55.257 11:34:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:23:55.257 11:34:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:55.257 11:34:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:55.257 11:34:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:55.514 11:34:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:55.514 "name": "raid_bdev1", 00:23:55.514 "uuid": "1746909c-81cc-4286-a264-8fe6432320b4", 00:23:55.514 "strip_size_kb": 0, 00:23:55.514 "state": "online", 00:23:55.514 "raid_level": "raid1", 00:23:55.514 "superblock": false, 00:23:55.514 "num_base_bdevs": 2, 00:23:55.514 "num_base_bdevs_discovered": 2, 00:23:55.514 "num_base_bdevs_operational": 2, 00:23:55.514 "process": { 00:23:55.514 "type": "rebuild", 00:23:55.514 "target": "spare", 00:23:55.514 "progress": { 00:23:55.514 "blocks": 36864, 00:23:55.514 "percent": 56 00:23:55.514 } 00:23:55.514 }, 00:23:55.514 "base_bdevs_list": [ 00:23:55.514 { 00:23:55.514 "name": "spare", 00:23:55.514 "uuid": "cbfa7db1-5755-5490-ae84-8c0d9d2f7ef7", 00:23:55.514 "is_configured": true, 00:23:55.514 "data_offset": 0, 00:23:55.514 "data_size": 65536 00:23:55.514 }, 00:23:55.514 { 00:23:55.514 "name": "BaseBdev2", 00:23:55.514 "uuid": "b87ddd44-5920-53c0-8456-f91c0a57ceb6", 00:23:55.514 "is_configured": true, 00:23:55.514 "data_offset": 0, 00:23:55.514 "data_size": 65536 00:23:55.514 } 00:23:55.514 ] 00:23:55.514 }' 00:23:55.514 11:34:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:55.514 11:34:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:23:55.514 11:34:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:55.514 11:34:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:23:55.515 11:34:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@710 -- # sleep 1 00:23:55.515 [2024-07-15 11:34:39.025494] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:23:55.771 [2024-07-15 11:34:39.280673] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 45056 offset_begin: 43008 offset_end: 49152 00:23:56.029 [2024-07-15 11:34:39.506634] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:23:56.029 [2024-07-15 11:34:39.506780] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:23:56.286 [2024-07-15 11:34:39.837075] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 51200 offset_begin: 49152 offset_end: 55296 00:23:56.543 11:34:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:23:56.543 11:34:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:23:56.543 11:34:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:56.543 11:34:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:23:56.543 11:34:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:23:56.543 11:34:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:56.543 11:34:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:56.543 11:34:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:56.799 11:34:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:56.799 "name": "raid_bdev1", 00:23:56.799 "uuid": "1746909c-81cc-4286-a264-8fe6432320b4", 00:23:56.799 "strip_size_kb": 0, 00:23:56.799 "state": "online", 00:23:56.799 "raid_level": "raid1", 00:23:56.799 "superblock": false, 00:23:56.799 "num_base_bdevs": 2, 00:23:56.799 "num_base_bdevs_discovered": 2, 00:23:56.799 "num_base_bdevs_operational": 2, 00:23:56.799 "process": { 00:23:56.799 "type": "rebuild", 00:23:56.799 "target": "spare", 00:23:56.799 "progress": { 00:23:56.799 "blocks": 57344, 00:23:56.799 "percent": 87 00:23:56.799 } 00:23:56.799 }, 00:23:56.799 "base_bdevs_list": [ 00:23:56.799 { 00:23:56.799 "name": "spare", 00:23:56.799 "uuid": "cbfa7db1-5755-5490-ae84-8c0d9d2f7ef7", 00:23:56.799 "is_configured": true, 00:23:56.799 "data_offset": 0, 00:23:56.799 "data_size": 65536 00:23:56.799 }, 00:23:56.799 { 00:23:56.799 "name": "BaseBdev2", 00:23:56.799 "uuid": "b87ddd44-5920-53c0-8456-f91c0a57ceb6", 00:23:56.799 "is_configured": true, 00:23:56.799 "data_offset": 0, 00:23:56.799 "data_size": 65536 00:23:56.799 } 00:23:56.799 ] 00:23:56.799 }' 00:23:56.799 11:34:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:56.799 11:34:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:23:56.799 11:34:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:56.799 11:34:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:23:56.799 11:34:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@710 -- # sleep 1 00:23:57.056 [2024-07-15 11:34:40.554289] bdev_raid.c:2789:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:23:57.056 [2024-07-15 11:34:40.615518] bdev_raid.c:2504:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:23:57.056 [2024-07-15 11:34:40.625234] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:57.987 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:23:57.987 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:23:57.987 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:57.987 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:23:57.987 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:23:57.987 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:57.987 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:57.987 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:57.987 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:57.987 "name": "raid_bdev1", 00:23:57.987 "uuid": "1746909c-81cc-4286-a264-8fe6432320b4", 00:23:57.987 "strip_size_kb": 0, 00:23:57.987 "state": "online", 00:23:57.987 "raid_level": "raid1", 00:23:57.987 "superblock": false, 00:23:57.987 "num_base_bdevs": 2, 00:23:57.987 "num_base_bdevs_discovered": 2, 00:23:57.987 "num_base_bdevs_operational": 2, 00:23:57.987 "base_bdevs_list": [ 00:23:57.987 { 00:23:57.987 "name": "spare", 00:23:57.987 "uuid": "cbfa7db1-5755-5490-ae84-8c0d9d2f7ef7", 00:23:57.987 "is_configured": true, 00:23:57.987 "data_offset": 0, 00:23:57.987 "data_size": 65536 00:23:57.987 }, 00:23:57.987 { 00:23:57.987 "name": "BaseBdev2", 00:23:57.987 "uuid": "b87ddd44-5920-53c0-8456-f91c0a57ceb6", 00:23:57.987 "is_configured": true, 00:23:57.987 "data_offset": 0, 00:23:57.987 "data_size": 65536 00:23:57.987 } 00:23:57.987 ] 00:23:57.987 }' 00:23:57.987 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:57.987 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # [[ none == \r\e\b\u\i\l\d ]] 00:23:58.244 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:58.244 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # [[ none == \s\p\a\r\e ]] 00:23:58.244 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # break 00:23:58.244 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@714 -- # verify_raid_bdev_process raid_bdev1 none none 00:23:58.244 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:23:58.244 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:23:58.244 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@184 -- # local target=none 00:23:58.244 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:23:58.244 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:58.244 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:58.501 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:58.501 "name": "raid_bdev1", 00:23:58.501 "uuid": "1746909c-81cc-4286-a264-8fe6432320b4", 00:23:58.501 "strip_size_kb": 0, 00:23:58.501 "state": "online", 00:23:58.501 "raid_level": "raid1", 00:23:58.501 "superblock": false, 00:23:58.501 "num_base_bdevs": 2, 00:23:58.501 "num_base_bdevs_discovered": 2, 00:23:58.501 "num_base_bdevs_operational": 2, 00:23:58.501 "base_bdevs_list": [ 00:23:58.501 { 00:23:58.501 "name": "spare", 00:23:58.501 "uuid": "cbfa7db1-5755-5490-ae84-8c0d9d2f7ef7", 00:23:58.501 "is_configured": true, 00:23:58.501 "data_offset": 0, 00:23:58.501 "data_size": 65536 00:23:58.501 }, 00:23:58.501 { 00:23:58.501 "name": "BaseBdev2", 00:23:58.501 "uuid": "b87ddd44-5920-53c0-8456-f91c0a57ceb6", 00:23:58.501 "is_configured": true, 00:23:58.501 "data_offset": 0, 00:23:58.501 "data_size": 65536 00:23:58.501 } 00:23:58.501 ] 00:23:58.501 }' 00:23:58.501 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:23:58.501 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:23:58.501 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:23:58.501 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:23:58.501 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:23:58.501 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:23:58.501 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:23:58.501 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:23:58.501 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:23:58.501 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:23:58.501 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:23:58.501 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:23:58.501 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:23:58.501 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:23:58.501 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:58.501 11:34:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:58.757 11:34:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:23:58.757 "name": "raid_bdev1", 00:23:58.757 "uuid": "1746909c-81cc-4286-a264-8fe6432320b4", 00:23:58.757 "strip_size_kb": 0, 00:23:58.757 "state": "online", 00:23:58.757 "raid_level": "raid1", 00:23:58.757 "superblock": false, 00:23:58.757 "num_base_bdevs": 2, 00:23:58.757 "num_base_bdevs_discovered": 2, 00:23:58.757 "num_base_bdevs_operational": 2, 00:23:58.757 "base_bdevs_list": [ 00:23:58.757 { 00:23:58.757 "name": "spare", 00:23:58.757 "uuid": "cbfa7db1-5755-5490-ae84-8c0d9d2f7ef7", 00:23:58.757 "is_configured": true, 00:23:58.757 "data_offset": 0, 00:23:58.757 "data_size": 65536 00:23:58.757 }, 00:23:58.757 { 00:23:58.757 "name": "BaseBdev2", 00:23:58.757 "uuid": "b87ddd44-5920-53c0-8456-f91c0a57ceb6", 00:23:58.757 "is_configured": true, 00:23:58.757 "data_offset": 0, 00:23:58.757 "data_size": 65536 00:23:58.757 } 00:23:58.757 ] 00:23:58.757 }' 00:23:58.757 11:34:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:23:58.757 11:34:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:23:59.322 11:34:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@718 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:23:59.579 [2024-07-15 11:34:43.068592] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:23:59.579 [2024-07-15 11:34:43.068629] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:23:59.579 00:23:59.579 Latency(us) 00:23:59.579 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:23:59.579 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:23:59.579 raid_bdev1 : 12.04 94.01 282.03 0.00 0.00 13587.23 284.94 111240.24 00:23:59.579 =================================================================================================================== 00:23:59.579 Total : 94.01 282.03 0.00 0.00 13587.23 284.94 111240.24 00:23:59.579 [2024-07-15 11:34:43.156743] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:59.579 [2024-07-15 11:34:43.156772] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:23:59.579 [2024-07-15 11:34:43.156845] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:23:59.579 [2024-07-15 11:34:43.156858] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1c2c070 name raid_bdev1, state offline 00:23:59.579 0 00:23:59.835 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:23:59.835 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # jq length 00:23:59.835 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # [[ 0 == 0 ]] 00:23:59.835 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@721 -- # '[' true = true ']' 00:23:59.835 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:23:59.835 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@724 -- # nbd_start_disks /var/tmp/spdk-raid.sock spare /dev/nbd0 00:23:59.835 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:23:59.835 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:23:59.835 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:23:59.835 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:23:59.835 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:23:59.835 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:23:59.835 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:23:59.835 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:23:59.835 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk spare /dev/nbd0 00:24:00.092 /dev/nbd0 00:24:00.092 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:24:00.092 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:24:00.092 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:24:00.092 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@867 -- # local i 00:24:00.092 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:24:00.092 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:24:00.092 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # break 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:24:00.351 1+0 records in 00:24:00.351 1+0 records out 00:24:00.351 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000265081 s, 15.5 MB/s 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # size=4096 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # return 0 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # for bdev in "${base_bdevs[@]:1}" 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # '[' -z BaseBdev2 ']' 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@729 -- # nbd_start_disks /var/tmp/spdk-raid.sock BaseBdev2 /dev/nbd1 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:24:00.351 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:24:00.609 /dev/nbd1 00:24:00.609 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:24:00.609 11:34:43 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:24:00.609 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:24:00.609 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@867 -- # local i 00:24:00.609 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:24:00.609 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:24:00.609 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:24:00.609 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # break 00:24:00.609 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:24:00.609 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:24:00.609 11:34:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:24:00.609 1+0 records in 00:24:00.609 1+0 records out 00:24:00.609 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000262089 s, 15.6 MB/s 00:24:00.609 11:34:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:24:00.609 11:34:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # size=4096 00:24:00.609 11:34:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:24:00.609 11:34:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:24:00.609 11:34:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # return 0 00:24:00.609 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:24:00.609 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:24:00.609 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:24:00.609 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # nbd_stop_disks /var/tmp/spdk-raid.sock /dev/nbd1 00:24:00.609 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:24:00.609 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:24:00.609 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:24:00.609 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:24:00.609 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:24:00.609 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd1 00:24:00.867 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:24:00.867 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:24:00.867 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:24:00.867 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:24:00.867 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:24:00.867 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:24:00.867 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:24:00.867 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:24:00.867 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@733 -- # nbd_stop_disks /var/tmp/spdk-raid.sock /dev/nbd0 00:24:00.867 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:24:00.867 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:24:00.867 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:24:00.867 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:24:00.867 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:24:00.867 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd0 00:24:01.124 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:24:01.124 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:24:01.124 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:24:01.124 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:24:01.124 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:24:01.124 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:24:01.124 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:24:01.124 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:24:01.124 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@742 -- # '[' false = true ']' 00:24:01.124 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@782 -- # killprocess 976615 00:24:01.124 11:34:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@948 -- # '[' -z 976615 ']' 00:24:01.124 11:34:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@952 -- # kill -0 976615 00:24:01.124 11:34:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@953 -- # uname 00:24:01.124 11:34:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:24:01.124 11:34:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 976615 00:24:01.124 11:34:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:24:01.124 11:34:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:24:01.124 11:34:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@966 -- # echo 'killing process with pid 976615' 00:24:01.124 killing process with pid 976615 00:24:01.124 11:34:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@967 -- # kill 976615 00:24:01.124 Received shutdown signal, test time was about 13.597882 seconds 00:24:01.124 00:24:01.124 Latency(us) 00:24:01.124 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:01.124 =================================================================================================================== 00:24:01.124 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:24:01.124 [2024-07-15 11:34:44.713216] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:24:01.124 11:34:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@972 -- # wait 976615 00:24:01.383 [2024-07-15 11:34:44.736048] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:24:01.383 11:34:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # return 0 00:24:01.383 00:24:01.383 real 0m18.332s 00:24:01.383 user 0m28.103s 00:24:01.383 sys 0m2.844s 00:24:01.383 11:34:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1124 -- # xtrace_disable 00:24:01.383 11:34:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:24:01.383 ************************************ 00:24:01.383 END TEST raid_rebuild_test_io 00:24:01.383 ************************************ 00:24:01.640 11:34:45 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:24:01.640 11:34:45 bdev_raid -- bdev/bdev_raid.sh@880 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 2 true true true 00:24:01.640 11:34:45 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:24:01.640 11:34:45 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:24:01.640 11:34:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:24:01.640 ************************************ 00:24:01.640 START TEST raid_rebuild_test_sb_io 00:24:01.640 ************************************ 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1123 -- # raid_rebuild_test raid1 2 true true true 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@568 -- # local raid_level=raid1 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local num_base_bdevs=2 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local superblock=true 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local background_io=true 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local verify=true 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # (( i = 1 )) 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # echo BaseBdev1 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # echo BaseBdev2 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local base_bdevs 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local raid_bdev_name=raid_bdev1 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local strip_size 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local create_arg 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local raid_bdev_size 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local data_offset 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@580 -- # '[' raid1 '!=' raid1 ']' 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@588 -- # strip_size=0 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@591 -- # '[' true = true ']' 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # create_arg+=' -s' 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # raid_pid=979138 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # waitforlisten 979138 /var/tmp/spdk-raid.sock 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@595 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@829 -- # '[' -z 979138 ']' 00:24:01.640 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:24:01.641 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@834 -- # local max_retries=100 00:24:01.641 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:24:01.641 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:24:01.641 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@838 -- # xtrace_disable 00:24:01.641 11:34:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:24:01.641 [2024-07-15 11:34:45.124228] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:24:01.641 [2024-07-15 11:34:45.124287] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid979138 ] 00:24:01.641 I/O size of 3145728 is greater than zero copy threshold (65536). 00:24:01.641 Zero copy mechanism will not be used. 00:24:01.897 [2024-07-15 11:34:45.237080] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:01.897 [2024-07-15 11:34:45.334588] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:24:01.897 [2024-07-15 11:34:45.400049] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:24:01.897 [2024-07-15 11:34:45.400088] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:24:02.519 11:34:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:24:02.519 11:34:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@862 -- # return 0 00:24:02.519 11:34:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:24:02.519 11:34:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:24:02.775 BaseBdev1_malloc 00:24:02.775 11:34:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:24:03.032 [2024-07-15 11:34:46.529395] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:24:03.032 [2024-07-15 11:34:46.529447] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:03.032 [2024-07-15 11:34:46.529470] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xacdd40 00:24:03.032 [2024-07-15 11:34:46.529487] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:03.032 [2024-07-15 11:34:46.531245] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:03.032 [2024-07-15 11:34:46.531275] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:24:03.032 BaseBdev1 00:24:03.032 11:34:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:24:03.032 11:34:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:24:03.290 BaseBdev2_malloc 00:24:03.290 11:34:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:24:03.547 [2024-07-15 11:34:46.963440] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:24:03.547 [2024-07-15 11:34:46.963487] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:03.547 [2024-07-15 11:34:46.963510] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xace860 00:24:03.547 [2024-07-15 11:34:46.963523] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:03.547 [2024-07-15 11:34:46.964985] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:03.547 [2024-07-15 11:34:46.965014] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:24:03.547 BaseBdev2 00:24:03.547 11:34:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@606 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b spare_malloc 00:24:03.805 spare_malloc 00:24:03.805 11:34:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:24:04.062 spare_delay 00:24:04.062 11:34:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:24:04.320 [2024-07-15 11:34:47.701976] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:24:04.320 [2024-07-15 11:34:47.702019] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:04.320 [2024-07-15 11:34:47.702038] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xc7cec0 00:24:04.320 [2024-07-15 11:34:47.702050] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:04.320 [2024-07-15 11:34:47.703426] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:04.320 [2024-07-15 11:34:47.703452] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:24:04.320 spare 00:24:04.320 11:34:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@611 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2' -n raid_bdev1 00:24:04.577 [2024-07-15 11:34:47.946647] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:24:04.577 [2024-07-15 11:34:47.947900] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:24:04.577 [2024-07-15 11:34:47.948071] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xc7e070 00:24:04.577 [2024-07-15 11:34:47.948084] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:24:04.577 [2024-07-15 11:34:47.948275] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xc77490 00:24:04.577 [2024-07-15 11:34:47.948414] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xc7e070 00:24:04.577 [2024-07-15 11:34:47.948424] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0xc7e070 00:24:04.577 [2024-07-15 11:34:47.948522] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:04.577 11:34:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:24:04.577 11:34:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:24:04.577 11:34:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:24:04.577 11:34:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:24:04.577 11:34:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:24:04.577 11:34:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:24:04.577 11:34:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:24:04.577 11:34:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:24:04.577 11:34:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:24:04.577 11:34:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:24:04.577 11:34:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:04.577 11:34:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:04.835 11:34:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:24:04.835 "name": "raid_bdev1", 00:24:04.835 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:04.835 "strip_size_kb": 0, 00:24:04.835 "state": "online", 00:24:04.835 "raid_level": "raid1", 00:24:04.835 "superblock": true, 00:24:04.835 "num_base_bdevs": 2, 00:24:04.835 "num_base_bdevs_discovered": 2, 00:24:04.835 "num_base_bdevs_operational": 2, 00:24:04.835 "base_bdevs_list": [ 00:24:04.835 { 00:24:04.835 "name": "BaseBdev1", 00:24:04.835 "uuid": "98f54a85-c193-5fbc-9039-eb8a7996f1b4", 00:24:04.835 "is_configured": true, 00:24:04.835 "data_offset": 2048, 00:24:04.835 "data_size": 63488 00:24:04.835 }, 00:24:04.835 { 00:24:04.835 "name": "BaseBdev2", 00:24:04.835 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:04.835 "is_configured": true, 00:24:04.835 "data_offset": 2048, 00:24:04.835 "data_size": 63488 00:24:04.835 } 00:24:04.835 ] 00:24:04.835 }' 00:24:04.835 11:34:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:24:04.835 11:34:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:24:05.401 11:34:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@615 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:24:05.401 11:34:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@615 -- # jq -r '.[].num_blocks' 00:24:05.659 [2024-07-15 11:34:49.017922] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:24:05.659 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@615 -- # raid_bdev_size=63488 00:24:05.659 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@618 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:05.659 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@618 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:24:05.917 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@618 -- # data_offset=2048 00:24:05.917 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@620 -- # '[' true = true ']' 00:24:05.917 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@639 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev1 00:24:05.917 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@622 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:24:05.917 [2024-07-15 11:34:49.397108] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xc7ec50 00:24:05.917 I/O size of 3145728 is greater than zero copy threshold (65536). 00:24:05.917 Zero copy mechanism will not be used. 00:24:05.917 Running I/O for 60 seconds... 00:24:06.175 [2024-07-15 11:34:49.514463] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:24:06.175 [2024-07-15 11:34:49.522647] bdev_raid.c:1919:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0xc7ec50 00:24:06.175 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@642 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:24:06.175 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:24:06.175 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:24:06.175 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:24:06.175 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:24:06.175 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:24:06.175 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:24:06.175 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:24:06.175 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:24:06.175 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:24:06.175 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:06.175 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:06.433 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:24:06.433 "name": "raid_bdev1", 00:24:06.433 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:06.433 "strip_size_kb": 0, 00:24:06.433 "state": "online", 00:24:06.433 "raid_level": "raid1", 00:24:06.433 "superblock": true, 00:24:06.433 "num_base_bdevs": 2, 00:24:06.433 "num_base_bdevs_discovered": 1, 00:24:06.433 "num_base_bdevs_operational": 1, 00:24:06.433 "base_bdevs_list": [ 00:24:06.433 { 00:24:06.433 "name": null, 00:24:06.433 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:06.433 "is_configured": false, 00:24:06.433 "data_offset": 2048, 00:24:06.433 "data_size": 63488 00:24:06.433 }, 00:24:06.433 { 00:24:06.433 "name": "BaseBdev2", 00:24:06.433 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:06.433 "is_configured": true, 00:24:06.433 "data_offset": 2048, 00:24:06.433 "data_size": 63488 00:24:06.433 } 00:24:06.433 ] 00:24:06.433 }' 00:24:06.433 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:24:06.433 11:34:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:24:07.000 11:34:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@645 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:24:07.259 [2024-07-15 11:34:50.670052] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:24:07.259 [2024-07-15 11:34:50.721019] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xbea230 00:24:07.259 [2024-07-15 11:34:50.723372] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:24:07.259 11:34:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # sleep 1 00:24:07.259 [2024-07-15 11:34:50.842299] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:24:07.259 [2024-07-15 11:34:50.842726] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:24:07.517 [2024-07-15 11:34:51.079015] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:24:07.517 [2024-07-15 11:34:51.079154] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:24:08.083 [2024-07-15 11:34:51.426475] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:24:08.083 [2024-07-15 11:34:51.426798] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:24:08.083 [2024-07-15 11:34:51.546326] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:24:08.340 11:34:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@649 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:24:08.340 11:34:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:08.340 11:34:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:24:08.340 11:34:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:24:08.340 11:34:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:08.340 11:34:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:08.340 11:34:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:08.340 [2024-07-15 11:34:51.868792] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:24:08.340 [2024-07-15 11:34:51.869258] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:24:08.633 11:34:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:08.633 "name": "raid_bdev1", 00:24:08.633 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:08.633 "strip_size_kb": 0, 00:24:08.633 "state": "online", 00:24:08.633 "raid_level": "raid1", 00:24:08.633 "superblock": true, 00:24:08.633 "num_base_bdevs": 2, 00:24:08.633 "num_base_bdevs_discovered": 2, 00:24:08.633 "num_base_bdevs_operational": 2, 00:24:08.633 "process": { 00:24:08.633 "type": "rebuild", 00:24:08.633 "target": "spare", 00:24:08.633 "progress": { 00:24:08.633 "blocks": 14336, 00:24:08.633 "percent": 22 00:24:08.633 } 00:24:08.633 }, 00:24:08.633 "base_bdevs_list": [ 00:24:08.633 { 00:24:08.633 "name": "spare", 00:24:08.633 "uuid": "9e0bf645-776a-5b4f-992d-f1615f3c1e06", 00:24:08.633 "is_configured": true, 00:24:08.633 "data_offset": 2048, 00:24:08.633 "data_size": 63488 00:24:08.633 }, 00:24:08.633 { 00:24:08.633 "name": "BaseBdev2", 00:24:08.633 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:08.633 "is_configured": true, 00:24:08.633 "data_offset": 2048, 00:24:08.633 "data_size": 63488 00:24:08.633 } 00:24:08.633 ] 00:24:08.633 }' 00:24:08.633 11:34:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:08.633 11:34:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:24:08.633 11:34:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:08.633 11:34:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:24:08.633 11:34:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@652 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev spare 00:24:08.633 [2024-07-15 11:34:52.090797] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:24:08.905 [2024-07-15 11:34:52.301418] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:24:08.905 [2024-07-15 11:34:52.420843] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:24:08.905 [2024-07-15 11:34:52.430868] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:08.905 [2024-07-15 11:34:52.430898] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:24:08.905 [2024-07-15 11:34:52.430908] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:24:08.905 [2024-07-15 11:34:52.460794] bdev_raid.c:1919:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0xc7ec50 00:24:09.163 11:34:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@655 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:24:09.163 11:34:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:24:09.163 11:34:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:24:09.163 11:34:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:24:09.163 11:34:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:24:09.163 11:34:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:24:09.163 11:34:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:24:09.163 11:34:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:24:09.163 11:34:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:24:09.163 11:34:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:24:09.163 11:34:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:09.163 11:34:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:09.422 11:34:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:24:09.422 "name": "raid_bdev1", 00:24:09.422 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:09.422 "strip_size_kb": 0, 00:24:09.422 "state": "online", 00:24:09.422 "raid_level": "raid1", 00:24:09.422 "superblock": true, 00:24:09.422 "num_base_bdevs": 2, 00:24:09.422 "num_base_bdevs_discovered": 1, 00:24:09.422 "num_base_bdevs_operational": 1, 00:24:09.422 "base_bdevs_list": [ 00:24:09.422 { 00:24:09.422 "name": null, 00:24:09.422 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:09.422 "is_configured": false, 00:24:09.422 "data_offset": 2048, 00:24:09.422 "data_size": 63488 00:24:09.422 }, 00:24:09.422 { 00:24:09.422 "name": "BaseBdev2", 00:24:09.422 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:09.422 "is_configured": true, 00:24:09.422 "data_offset": 2048, 00:24:09.422 "data_size": 63488 00:24:09.422 } 00:24:09.422 ] 00:24:09.422 }' 00:24:09.422 11:34:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:24:09.422 11:34:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:24:09.988 11:34:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@658 -- # verify_raid_bdev_process raid_bdev1 none none 00:24:09.988 11:34:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:09.988 11:34:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:24:09.988 11:34:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=none 00:24:09.988 11:34:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:09.988 11:34:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:09.988 11:34:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:10.246 11:34:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:10.246 "name": "raid_bdev1", 00:24:10.246 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:10.246 "strip_size_kb": 0, 00:24:10.246 "state": "online", 00:24:10.246 "raid_level": "raid1", 00:24:10.246 "superblock": true, 00:24:10.246 "num_base_bdevs": 2, 00:24:10.246 "num_base_bdevs_discovered": 1, 00:24:10.246 "num_base_bdevs_operational": 1, 00:24:10.246 "base_bdevs_list": [ 00:24:10.246 { 00:24:10.246 "name": null, 00:24:10.246 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:10.246 "is_configured": false, 00:24:10.246 "data_offset": 2048, 00:24:10.246 "data_size": 63488 00:24:10.246 }, 00:24:10.246 { 00:24:10.246 "name": "BaseBdev2", 00:24:10.246 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:10.246 "is_configured": true, 00:24:10.246 "data_offset": 2048, 00:24:10.246 "data_size": 63488 00:24:10.246 } 00:24:10.246 ] 00:24:10.246 }' 00:24:10.246 11:34:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:10.246 11:34:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:24:10.246 11:34:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:10.246 11:34:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:24:10.246 11:34:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@661 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:24:10.504 [2024-07-15 11:34:53.971841] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:24:10.504 [2024-07-15 11:34:54.006576] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xc7ee60 00:24:10.504 [2024-07-15 11:34:54.008058] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:24:10.504 11:34:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # sleep 1 00:24:10.763 [2024-07-15 11:34:54.118683] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:24:10.763 [2024-07-15 11:34:54.119087] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:24:10.763 [2024-07-15 11:34:54.321649] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:24:10.763 [2024-07-15 11:34:54.321772] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:24:11.021 [2024-07-15 11:34:54.588277] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:24:11.586 [2024-07-15 11:34:54.947334] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:24:11.586 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:24:11.586 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:11.586 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:24:11.586 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:24:11.587 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:11.587 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:11.587 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:11.587 [2024-07-15 11:34:55.059584] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:24:11.844 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:11.844 "name": "raid_bdev1", 00:24:11.844 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:11.844 "strip_size_kb": 0, 00:24:11.844 "state": "online", 00:24:11.844 "raid_level": "raid1", 00:24:11.844 "superblock": true, 00:24:11.844 "num_base_bdevs": 2, 00:24:11.844 "num_base_bdevs_discovered": 2, 00:24:11.844 "num_base_bdevs_operational": 2, 00:24:11.844 "process": { 00:24:11.844 "type": "rebuild", 00:24:11.844 "target": "spare", 00:24:11.844 "progress": { 00:24:11.845 "blocks": 16384, 00:24:11.845 "percent": 25 00:24:11.845 } 00:24:11.845 }, 00:24:11.845 "base_bdevs_list": [ 00:24:11.845 { 00:24:11.845 "name": "spare", 00:24:11.845 "uuid": "9e0bf645-776a-5b4f-992d-f1615f3c1e06", 00:24:11.845 "is_configured": true, 00:24:11.845 "data_offset": 2048, 00:24:11.845 "data_size": 63488 00:24:11.845 }, 00:24:11.845 { 00:24:11.845 "name": "BaseBdev2", 00:24:11.845 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:11.845 "is_configured": true, 00:24:11.845 "data_offset": 2048, 00:24:11.845 "data_size": 63488 00:24:11.845 } 00:24:11.845 ] 00:24:11.845 }' 00:24:11.845 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:11.845 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:24:11.845 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:11.845 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:24:11.845 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@665 -- # '[' true = true ']' 00:24:11.845 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@665 -- # '[' = false ']' 00:24:11.845 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev_raid.sh: line 665: [: =: unary operator expected 00:24:11.845 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@690 -- # local num_base_bdevs_operational=2 00:24:11.845 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@692 -- # '[' raid1 = raid1 ']' 00:24:11.845 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@692 -- # '[' 2 -gt 2 ']' 00:24:11.845 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@705 -- # local timeout=823 00:24:11.845 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:24:11.845 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:24:11.845 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:11.845 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:24:11.845 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:24:11.845 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:11.845 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:11.845 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:11.845 [2024-07-15 11:34:55.416536] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:24:11.845 [2024-07-15 11:34:55.416870] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:24:12.103 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:12.103 "name": "raid_bdev1", 00:24:12.103 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:12.103 "strip_size_kb": 0, 00:24:12.103 "state": "online", 00:24:12.103 "raid_level": "raid1", 00:24:12.103 "superblock": true, 00:24:12.103 "num_base_bdevs": 2, 00:24:12.103 "num_base_bdevs_discovered": 2, 00:24:12.103 "num_base_bdevs_operational": 2, 00:24:12.103 "process": { 00:24:12.103 "type": "rebuild", 00:24:12.103 "target": "spare", 00:24:12.103 "progress": { 00:24:12.103 "blocks": 20480, 00:24:12.103 "percent": 32 00:24:12.103 } 00:24:12.103 }, 00:24:12.103 "base_bdevs_list": [ 00:24:12.103 { 00:24:12.103 "name": "spare", 00:24:12.103 "uuid": "9e0bf645-776a-5b4f-992d-f1615f3c1e06", 00:24:12.103 "is_configured": true, 00:24:12.103 "data_offset": 2048, 00:24:12.103 "data_size": 63488 00:24:12.103 }, 00:24:12.103 { 00:24:12.103 "name": "BaseBdev2", 00:24:12.103 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:12.103 "is_configured": true, 00:24:12.103 "data_offset": 2048, 00:24:12.103 "data_size": 63488 00:24:12.103 } 00:24:12.103 ] 00:24:12.103 }' 00:24:12.103 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:12.104 [2024-07-15 11:34:55.636990] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:24:12.104 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:24:12.104 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:12.361 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:24:12.361 11:34:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@710 -- # sleep 1 00:24:12.618 [2024-07-15 11:34:55.977973] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:24:12.618 [2024-07-15 11:34:56.105139] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:24:12.876 [2024-07-15 11:34:56.452040] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:24:13.135 11:34:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:24:13.135 11:34:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:24:13.135 11:34:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:13.135 11:34:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:24:13.135 11:34:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:24:13.135 11:34:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:13.135 11:34:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:13.135 11:34:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:13.393 11:34:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:13.393 "name": "raid_bdev1", 00:24:13.393 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:13.393 "strip_size_kb": 0, 00:24:13.393 "state": "online", 00:24:13.393 "raid_level": "raid1", 00:24:13.393 "superblock": true, 00:24:13.393 "num_base_bdevs": 2, 00:24:13.393 "num_base_bdevs_discovered": 2, 00:24:13.393 "num_base_bdevs_operational": 2, 00:24:13.393 "process": { 00:24:13.393 "type": "rebuild", 00:24:13.393 "target": "spare", 00:24:13.393 "progress": { 00:24:13.393 "blocks": 40960, 00:24:13.393 "percent": 64 00:24:13.393 } 00:24:13.393 }, 00:24:13.393 "base_bdevs_list": [ 00:24:13.393 { 00:24:13.393 "name": "spare", 00:24:13.393 "uuid": "9e0bf645-776a-5b4f-992d-f1615f3c1e06", 00:24:13.393 "is_configured": true, 00:24:13.393 "data_offset": 2048, 00:24:13.393 "data_size": 63488 00:24:13.393 }, 00:24:13.393 { 00:24:13.393 "name": "BaseBdev2", 00:24:13.393 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:13.393 "is_configured": true, 00:24:13.393 "data_offset": 2048, 00:24:13.393 "data_size": 63488 00:24:13.393 } 00:24:13.393 ] 00:24:13.393 }' 00:24:13.393 11:34:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:13.651 11:34:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:24:13.651 11:34:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:13.651 11:34:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:24:13.651 11:34:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@710 -- # sleep 1 00:24:14.215 [2024-07-15 11:34:57.766176] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:24:14.215 [2024-07-15 11:34:57.766610] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:24:14.472 11:34:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:24:14.472 11:34:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:24:14.472 11:34:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:14.472 11:34:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:24:14.472 11:34:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:24:14.472 11:34:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:14.472 11:34:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:14.472 11:34:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:14.728 [2024-07-15 11:34:58.216738] bdev_raid.c:2789:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:24:14.728 11:34:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:14.728 "name": "raid_bdev1", 00:24:14.728 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:14.728 "strip_size_kb": 0, 00:24:14.728 "state": "online", 00:24:14.728 "raid_level": "raid1", 00:24:14.728 "superblock": true, 00:24:14.728 "num_base_bdevs": 2, 00:24:14.728 "num_base_bdevs_discovered": 2, 00:24:14.728 "num_base_bdevs_operational": 2, 00:24:14.728 "process": { 00:24:14.728 "type": "rebuild", 00:24:14.728 "target": "spare", 00:24:14.728 "progress": { 00:24:14.728 "blocks": 63488, 00:24:14.728 "percent": 100 00:24:14.728 } 00:24:14.728 }, 00:24:14.728 "base_bdevs_list": [ 00:24:14.728 { 00:24:14.728 "name": "spare", 00:24:14.728 "uuid": "9e0bf645-776a-5b4f-992d-f1615f3c1e06", 00:24:14.728 "is_configured": true, 00:24:14.728 "data_offset": 2048, 00:24:14.728 "data_size": 63488 00:24:14.728 }, 00:24:14.728 { 00:24:14.728 "name": "BaseBdev2", 00:24:14.728 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:14.728 "is_configured": true, 00:24:14.728 "data_offset": 2048, 00:24:14.728 "data_size": 63488 00:24:14.728 } 00:24:14.728 ] 00:24:14.728 }' 00:24:14.728 11:34:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:14.728 [2024-07-15 11:34:58.317002] bdev_raid.c:2504:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:24:14.728 [2024-07-15 11:34:58.318698] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:14.984 11:34:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:24:14.984 11:34:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:14.984 11:34:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:24:14.984 11:34:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@710 -- # sleep 1 00:24:15.914 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:24:15.914 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:24:15.914 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:15.914 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:24:15.914 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:24:15.914 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:15.914 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:15.914 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:16.171 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:16.171 "name": "raid_bdev1", 00:24:16.171 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:16.171 "strip_size_kb": 0, 00:24:16.171 "state": "online", 00:24:16.171 "raid_level": "raid1", 00:24:16.171 "superblock": true, 00:24:16.171 "num_base_bdevs": 2, 00:24:16.171 "num_base_bdevs_discovered": 2, 00:24:16.171 "num_base_bdevs_operational": 2, 00:24:16.171 "base_bdevs_list": [ 00:24:16.171 { 00:24:16.171 "name": "spare", 00:24:16.171 "uuid": "9e0bf645-776a-5b4f-992d-f1615f3c1e06", 00:24:16.171 "is_configured": true, 00:24:16.171 "data_offset": 2048, 00:24:16.171 "data_size": 63488 00:24:16.171 }, 00:24:16.171 { 00:24:16.171 "name": "BaseBdev2", 00:24:16.171 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:16.171 "is_configured": true, 00:24:16.171 "data_offset": 2048, 00:24:16.171 "data_size": 63488 00:24:16.171 } 00:24:16.171 ] 00:24:16.171 }' 00:24:16.171 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:16.171 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ none == \r\e\b\u\i\l\d ]] 00:24:16.171 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:16.171 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ none == \s\p\a\r\e ]] 00:24:16.171 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # break 00:24:16.171 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@714 -- # verify_raid_bdev_process raid_bdev1 none none 00:24:16.171 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:16.171 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:24:16.171 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=none 00:24:16.171 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:16.171 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:16.171 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:16.429 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:16.429 "name": "raid_bdev1", 00:24:16.429 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:16.429 "strip_size_kb": 0, 00:24:16.429 "state": "online", 00:24:16.429 "raid_level": "raid1", 00:24:16.429 "superblock": true, 00:24:16.429 "num_base_bdevs": 2, 00:24:16.429 "num_base_bdevs_discovered": 2, 00:24:16.429 "num_base_bdevs_operational": 2, 00:24:16.429 "base_bdevs_list": [ 00:24:16.429 { 00:24:16.429 "name": "spare", 00:24:16.429 "uuid": "9e0bf645-776a-5b4f-992d-f1615f3c1e06", 00:24:16.429 "is_configured": true, 00:24:16.429 "data_offset": 2048, 00:24:16.429 "data_size": 63488 00:24:16.429 }, 00:24:16.429 { 00:24:16.429 "name": "BaseBdev2", 00:24:16.429 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:16.429 "is_configured": true, 00:24:16.429 "data_offset": 2048, 00:24:16.429 "data_size": 63488 00:24:16.429 } 00:24:16.429 ] 00:24:16.429 }' 00:24:16.429 11:34:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:16.686 11:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:24:16.686 11:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:16.687 11:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:24:16.687 11:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:24:16.687 11:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:24:16.687 11:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:24:16.687 11:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:24:16.687 11:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:24:16.687 11:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:24:16.687 11:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:24:16.687 11:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:24:16.687 11:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:24:16.687 11:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:24:16.687 11:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:16.687 11:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:16.687 11:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:24:16.687 "name": "raid_bdev1", 00:24:16.687 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:16.687 "strip_size_kb": 0, 00:24:16.687 "state": "online", 00:24:16.687 "raid_level": "raid1", 00:24:16.687 "superblock": true, 00:24:16.687 "num_base_bdevs": 2, 00:24:16.687 "num_base_bdevs_discovered": 2, 00:24:16.687 "num_base_bdevs_operational": 2, 00:24:16.687 "base_bdevs_list": [ 00:24:16.687 { 00:24:16.687 "name": "spare", 00:24:16.687 "uuid": "9e0bf645-776a-5b4f-992d-f1615f3c1e06", 00:24:16.687 "is_configured": true, 00:24:16.687 "data_offset": 2048, 00:24:16.687 "data_size": 63488 00:24:16.687 }, 00:24:16.687 { 00:24:16.687 "name": "BaseBdev2", 00:24:16.687 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:16.687 "is_configured": true, 00:24:16.687 "data_offset": 2048, 00:24:16.687 "data_size": 63488 00:24:16.687 } 00:24:16.687 ] 00:24:16.687 }' 00:24:16.687 11:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:24:16.687 11:35:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:24:17.619 11:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@718 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:24:17.619 [2024-07-15 11:35:01.087488] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:24:17.619 [2024-07-15 11:35:01.087525] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:24:17.619 00:24:17.619 Latency(us) 00:24:17.619 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:17.619 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:24:17.619 raid_bdev1 : 11.75 95.22 285.66 0.00 0.00 14692.83 292.06 110328.43 00:24:17.619 =================================================================================================================== 00:24:17.619 Total : 95.22 285.66 0.00 0.00 14692.83 292.06 110328.43 00:24:17.619 [2024-07-15 11:35:01.183719] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:17.619 [2024-07-15 11:35:01.183749] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:24:17.619 [2024-07-15 11:35:01.183823] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:24:17.619 [2024-07-15 11:35:01.183835] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xc7e070 name raid_bdev1, state offline 00:24:17.619 0 00:24:17.619 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:17.619 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # jq length 00:24:17.877 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # [[ 0 == 0 ]] 00:24:17.877 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@721 -- # '[' true = true ']' 00:24:17.877 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:24:17.877 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@724 -- # nbd_start_disks /var/tmp/spdk-raid.sock spare /dev/nbd0 00:24:17.877 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:24:17.877 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:24:17.877 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:24:17.877 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:24:17.877 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:24:17.877 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:24:17.877 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:24:17.877 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:24:17.877 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk spare /dev/nbd0 00:24:18.133 /dev/nbd0 00:24:18.133 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:24:18.133 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:24:18.133 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:24:18.133 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@867 -- # local i 00:24:18.133 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:24:18.133 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:24:18.133 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:24:18.133 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # break 00:24:18.133 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:24:18.133 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:24:18.133 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:24:18.390 1+0 records in 00:24:18.390 1+0 records out 00:24:18.390 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000283058 s, 14.5 MB/s 00:24:18.390 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:24:18.390 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # size=4096 00:24:18.390 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:24:18.390 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:24:18.390 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # return 0 00:24:18.390 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:24:18.390 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:24:18.390 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # for bdev in "${base_bdevs[@]:1}" 00:24:18.390 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # '[' -z BaseBdev2 ']' 00:24:18.391 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@729 -- # nbd_start_disks /var/tmp/spdk-raid.sock BaseBdev2 /dev/nbd1 00:24:18.391 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:24:18.391 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:24:18.391 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:24:18.391 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:24:18.391 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:24:18.391 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:24:18.391 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:24:18.391 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:24:18.391 11:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:24:18.647 /dev/nbd1 00:24:18.647 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:24:18.647 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:24:18.647 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:24:18.647 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@867 -- # local i 00:24:18.647 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:24:18.647 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:24:18.647 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:24:18.647 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # break 00:24:18.647 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:24:18.647 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:24:18.647 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:24:18.647 1+0 records in 00:24:18.647 1+0 records out 00:24:18.647 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000276254 s, 14.8 MB/s 00:24:18.647 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:24:18.647 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # size=4096 00:24:18.647 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:24:18.648 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:24:18.648 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # return 0 00:24:18.648 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:24:18.648 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:24:18.648 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:24:18.648 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # nbd_stop_disks /var/tmp/spdk-raid.sock /dev/nbd1 00:24:18.648 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:24:18.648 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:24:18.648 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:24:18.648 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:24:18.648 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:24:18.648 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd1 00:24:18.905 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:24:18.905 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:24:18.905 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:24:18.905 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:24:18.905 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:24:18.905 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:24:18.905 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:24:18.905 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:24:18.905 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@733 -- # nbd_stop_disks /var/tmp/spdk-raid.sock /dev/nbd0 00:24:18.905 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:24:18.905 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:24:18.905 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:24:18.905 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:24:18.905 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:24:18.905 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd0 00:24:19.162 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:24:19.162 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:24:19.162 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:24:19.162 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:24:19.162 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:24:19.162 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:24:19.162 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:24:19.162 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:24:19.162 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@742 -- # '[' true = true ']' 00:24:19.162 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@744 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:24:19.419 11:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:24:19.419 [2024-07-15 11:35:03.012556] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:24:19.420 [2024-07-15 11:35:03.012606] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:19.420 [2024-07-15 11:35:03.012631] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xadee70 00:24:19.420 [2024-07-15 11:35:03.012644] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:19.678 [2024-07-15 11:35:03.014284] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:19.678 [2024-07-15 11:35:03.014314] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:24:19.678 [2024-07-15 11:35:03.014393] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:24:19.678 [2024-07-15 11:35:03.014424] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:24:19.678 [2024-07-15 11:35:03.014524] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:24:19.678 spare 00:24:19.678 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:24:19.678 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:24:19.678 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:24:19.678 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:24:19.678 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:24:19.678 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:24:19.678 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:24:19.678 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:24:19.678 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:24:19.678 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:24:19.678 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:19.678 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:19.678 [2024-07-15 11:35:03.114838] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xc7e2f0 00:24:19.678 [2024-07-15 11:35:03.114856] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:24:19.678 [2024-07-15 11:35:03.115050] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xc77490 00:24:19.678 [2024-07-15 11:35:03.115195] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xc7e2f0 00:24:19.678 [2024-07-15 11:35:03.115205] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0xc7e2f0 00:24:19.678 [2024-07-15 11:35:03.115310] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:19.678 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:24:19.678 "name": "raid_bdev1", 00:24:19.678 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:19.678 "strip_size_kb": 0, 00:24:19.678 "state": "online", 00:24:19.678 "raid_level": "raid1", 00:24:19.678 "superblock": true, 00:24:19.678 "num_base_bdevs": 2, 00:24:19.678 "num_base_bdevs_discovered": 2, 00:24:19.678 "num_base_bdevs_operational": 2, 00:24:19.678 "base_bdevs_list": [ 00:24:19.678 { 00:24:19.678 "name": "spare", 00:24:19.678 "uuid": "9e0bf645-776a-5b4f-992d-f1615f3c1e06", 00:24:19.678 "is_configured": true, 00:24:19.678 "data_offset": 2048, 00:24:19.678 "data_size": 63488 00:24:19.678 }, 00:24:19.678 { 00:24:19.678 "name": "BaseBdev2", 00:24:19.678 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:19.678 "is_configured": true, 00:24:19.678 "data_offset": 2048, 00:24:19.678 "data_size": 63488 00:24:19.678 } 00:24:19.678 ] 00:24:19.678 }' 00:24:19.678 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:24:19.678 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:24:20.242 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@748 -- # verify_raid_bdev_process raid_bdev1 none none 00:24:20.243 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:20.243 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:24:20.243 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=none 00:24:20.243 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:20.243 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:20.243 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:20.500 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:20.500 "name": "raid_bdev1", 00:24:20.500 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:20.500 "strip_size_kb": 0, 00:24:20.500 "state": "online", 00:24:20.500 "raid_level": "raid1", 00:24:20.500 "superblock": true, 00:24:20.500 "num_base_bdevs": 2, 00:24:20.500 "num_base_bdevs_discovered": 2, 00:24:20.500 "num_base_bdevs_operational": 2, 00:24:20.500 "base_bdevs_list": [ 00:24:20.500 { 00:24:20.500 "name": "spare", 00:24:20.500 "uuid": "9e0bf645-776a-5b4f-992d-f1615f3c1e06", 00:24:20.500 "is_configured": true, 00:24:20.500 "data_offset": 2048, 00:24:20.500 "data_size": 63488 00:24:20.500 }, 00:24:20.500 { 00:24:20.500 "name": "BaseBdev2", 00:24:20.500 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:20.500 "is_configured": true, 00:24:20.500 "data_offset": 2048, 00:24:20.500 "data_size": 63488 00:24:20.500 } 00:24:20.500 ] 00:24:20.500 }' 00:24:20.500 11:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:20.500 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:24:20.500 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:20.500 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:24:20.500 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:20.500 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # jq -r '.[].base_bdevs_list[0].name' 00:24:20.757 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # [[ spare == \s\p\a\r\e ]] 00:24:20.757 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@752 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev spare 00:24:21.014 [2024-07-15 11:35:04.557000] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:24:21.014 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@753 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:24:21.014 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:24:21.014 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:24:21.014 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:24:21.014 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:24:21.014 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:24:21.014 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:24:21.014 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:24:21.014 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:24:21.014 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:24:21.014 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:21.014 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:21.271 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:24:21.271 "name": "raid_bdev1", 00:24:21.271 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:21.271 "strip_size_kb": 0, 00:24:21.271 "state": "online", 00:24:21.271 "raid_level": "raid1", 00:24:21.271 "superblock": true, 00:24:21.271 "num_base_bdevs": 2, 00:24:21.271 "num_base_bdevs_discovered": 1, 00:24:21.271 "num_base_bdevs_operational": 1, 00:24:21.271 "base_bdevs_list": [ 00:24:21.271 { 00:24:21.271 "name": null, 00:24:21.271 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:21.271 "is_configured": false, 00:24:21.271 "data_offset": 2048, 00:24:21.271 "data_size": 63488 00:24:21.271 }, 00:24:21.271 { 00:24:21.271 "name": "BaseBdev2", 00:24:21.271 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:21.271 "is_configured": true, 00:24:21.271 "data_offset": 2048, 00:24:21.271 "data_size": 63488 00:24:21.271 } 00:24:21.271 ] 00:24:21.271 }' 00:24:21.271 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:24:21.271 11:35:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:24:21.834 11:35:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:24:22.091 [2024-07-15 11:35:05.636021] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:24:22.091 [2024-07-15 11:35:05.636178] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:24:22.091 [2024-07-15 11:35:05.636195] bdev_raid.c:3620:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:24:22.091 [2024-07-15 11:35:05.636224] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:24:22.091 [2024-07-15 11:35:05.641552] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xc77490 00:24:22.091 [2024-07-15 11:35:05.643889] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:24:22.091 11:35:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # sleep 1 00:24:23.462 11:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:24:23.462 11:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:23.462 11:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:24:23.462 11:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:24:23.462 11:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:23.462 11:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:23.462 11:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:23.462 11:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:23.462 "name": "raid_bdev1", 00:24:23.462 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:23.462 "strip_size_kb": 0, 00:24:23.462 "state": "online", 00:24:23.462 "raid_level": "raid1", 00:24:23.462 "superblock": true, 00:24:23.462 "num_base_bdevs": 2, 00:24:23.462 "num_base_bdevs_discovered": 2, 00:24:23.462 "num_base_bdevs_operational": 2, 00:24:23.462 "process": { 00:24:23.462 "type": "rebuild", 00:24:23.462 "target": "spare", 00:24:23.462 "progress": { 00:24:23.462 "blocks": 24576, 00:24:23.462 "percent": 38 00:24:23.462 } 00:24:23.462 }, 00:24:23.462 "base_bdevs_list": [ 00:24:23.462 { 00:24:23.462 "name": "spare", 00:24:23.462 "uuid": "9e0bf645-776a-5b4f-992d-f1615f3c1e06", 00:24:23.462 "is_configured": true, 00:24:23.462 "data_offset": 2048, 00:24:23.462 "data_size": 63488 00:24:23.462 }, 00:24:23.462 { 00:24:23.462 "name": "BaseBdev2", 00:24:23.462 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:23.462 "is_configured": true, 00:24:23.462 "data_offset": 2048, 00:24:23.462 "data_size": 63488 00:24:23.462 } 00:24:23.462 ] 00:24:23.462 }' 00:24:23.462 11:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:23.462 11:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:24:23.462 11:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:23.463 11:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:24:23.463 11:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@759 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:24:23.719 [2024-07-15 11:35:07.234413] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:24:23.719 [2024-07-15 11:35:07.256751] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:24:23.719 [2024-07-15 11:35:07.256799] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:23.719 [2024-07-15 11:35:07.256814] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:24:23.719 [2024-07-15 11:35:07.256822] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:24:23.719 11:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@760 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:24:23.719 11:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:24:23.719 11:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:24:23.719 11:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:24:23.719 11:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:24:23.719 11:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:24:23.719 11:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:24:23.719 11:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:24:23.719 11:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:24:23.719 11:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:24:23.719 11:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:23.719 11:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:23.976 11:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:24:23.976 "name": "raid_bdev1", 00:24:23.976 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:23.976 "strip_size_kb": 0, 00:24:23.976 "state": "online", 00:24:23.976 "raid_level": "raid1", 00:24:23.976 "superblock": true, 00:24:23.976 "num_base_bdevs": 2, 00:24:23.976 "num_base_bdevs_discovered": 1, 00:24:23.976 "num_base_bdevs_operational": 1, 00:24:23.976 "base_bdevs_list": [ 00:24:23.976 { 00:24:23.976 "name": null, 00:24:23.976 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:23.976 "is_configured": false, 00:24:23.976 "data_offset": 2048, 00:24:23.976 "data_size": 63488 00:24:23.976 }, 00:24:23.976 { 00:24:23.976 "name": "BaseBdev2", 00:24:23.976 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:23.976 "is_configured": true, 00:24:23.976 "data_offset": 2048, 00:24:23.976 "data_size": 63488 00:24:23.976 } 00:24:23.976 ] 00:24:23.976 }' 00:24:23.976 11:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:24:23.976 11:35:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:24:24.541 11:35:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:24:24.799 [2024-07-15 11:35:08.341144] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:24:24.799 [2024-07-15 11:35:08.341198] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:24.799 [2024-07-15 11:35:08.341223] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xc77da0 00:24:24.799 [2024-07-15 11:35:08.341235] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:24.799 [2024-07-15 11:35:08.341618] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:24.799 [2024-07-15 11:35:08.341637] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:24:24.799 [2024-07-15 11:35:08.341723] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:24:24.799 [2024-07-15 11:35:08.341737] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:24:24.799 [2024-07-15 11:35:08.341748] bdev_raid.c:3620:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:24:24.799 [2024-07-15 11:35:08.341766] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:24:24.799 [2024-07-15 11:35:08.347067] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xc77490 00:24:24.799 spare 00:24:24.799 [2024-07-15 11:35:08.348532] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:24:24.799 11:35:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # sleep 1 00:24:26.206 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:24:26.206 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:26.206 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:24:26.206 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:24:26.206 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:26.206 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:26.206 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:26.206 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:26.206 "name": "raid_bdev1", 00:24:26.206 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:26.206 "strip_size_kb": 0, 00:24:26.206 "state": "online", 00:24:26.206 "raid_level": "raid1", 00:24:26.206 "superblock": true, 00:24:26.206 "num_base_bdevs": 2, 00:24:26.206 "num_base_bdevs_discovered": 2, 00:24:26.206 "num_base_bdevs_operational": 2, 00:24:26.206 "process": { 00:24:26.206 "type": "rebuild", 00:24:26.206 "target": "spare", 00:24:26.206 "progress": { 00:24:26.206 "blocks": 24576, 00:24:26.206 "percent": 38 00:24:26.206 } 00:24:26.206 }, 00:24:26.206 "base_bdevs_list": [ 00:24:26.206 { 00:24:26.206 "name": "spare", 00:24:26.206 "uuid": "9e0bf645-776a-5b4f-992d-f1615f3c1e06", 00:24:26.206 "is_configured": true, 00:24:26.206 "data_offset": 2048, 00:24:26.206 "data_size": 63488 00:24:26.206 }, 00:24:26.206 { 00:24:26.206 "name": "BaseBdev2", 00:24:26.206 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:26.206 "is_configured": true, 00:24:26.206 "data_offset": 2048, 00:24:26.206 "data_size": 63488 00:24:26.206 } 00:24:26.206 ] 00:24:26.206 }' 00:24:26.206 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:26.206 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:24:26.206 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:26.206 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:24:26.206 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@766 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:24:26.463 [2024-07-15 11:35:09.928802] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:24:26.463 [2024-07-15 11:35:09.961303] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:24:26.463 [2024-07-15 11:35:09.961347] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:26.463 [2024-07-15 11:35:09.961363] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:24:26.463 [2024-07-15 11:35:09.961371] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:24:26.463 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@767 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:24:26.463 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:24:26.463 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:24:26.463 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:24:26.463 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:24:26.463 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:24:26.463 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:24:26.463 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:24:26.463 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:24:26.463 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:24:26.463 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:26.463 11:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:26.719 11:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:24:26.719 "name": "raid_bdev1", 00:24:26.719 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:26.719 "strip_size_kb": 0, 00:24:26.719 "state": "online", 00:24:26.719 "raid_level": "raid1", 00:24:26.719 "superblock": true, 00:24:26.719 "num_base_bdevs": 2, 00:24:26.719 "num_base_bdevs_discovered": 1, 00:24:26.719 "num_base_bdevs_operational": 1, 00:24:26.719 "base_bdevs_list": [ 00:24:26.719 { 00:24:26.719 "name": null, 00:24:26.719 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:26.719 "is_configured": false, 00:24:26.719 "data_offset": 2048, 00:24:26.719 "data_size": 63488 00:24:26.719 }, 00:24:26.719 { 00:24:26.719 "name": "BaseBdev2", 00:24:26.719 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:26.719 "is_configured": true, 00:24:26.719 "data_offset": 2048, 00:24:26.719 "data_size": 63488 00:24:26.719 } 00:24:26.719 ] 00:24:26.719 }' 00:24:26.719 11:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:24:26.719 11:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:24:27.283 11:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # verify_raid_bdev_process raid_bdev1 none none 00:24:27.283 11:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:27.283 11:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:24:27.283 11:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=none 00:24:27.283 11:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:27.283 11:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:27.283 11:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:27.539 11:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:27.539 "name": "raid_bdev1", 00:24:27.539 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:27.539 "strip_size_kb": 0, 00:24:27.539 "state": "online", 00:24:27.539 "raid_level": "raid1", 00:24:27.539 "superblock": true, 00:24:27.539 "num_base_bdevs": 2, 00:24:27.539 "num_base_bdevs_discovered": 1, 00:24:27.539 "num_base_bdevs_operational": 1, 00:24:27.539 "base_bdevs_list": [ 00:24:27.539 { 00:24:27.539 "name": null, 00:24:27.539 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:27.539 "is_configured": false, 00:24:27.539 "data_offset": 2048, 00:24:27.539 "data_size": 63488 00:24:27.539 }, 00:24:27.539 { 00:24:27.539 "name": "BaseBdev2", 00:24:27.539 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:27.539 "is_configured": true, 00:24:27.539 "data_offset": 2048, 00:24:27.539 "data_size": 63488 00:24:27.539 } 00:24:27.539 ] 00:24:27.539 }' 00:24:27.539 11:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:27.539 11:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:24:27.539 11:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:27.796 11:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:24:27.796 11:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@771 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete BaseBdev1 00:24:28.053 11:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@772 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:24:28.053 [2024-07-15 11:35:11.642592] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:24:28.053 [2024-07-15 11:35:11.642645] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:28.053 [2024-07-15 11:35:11.642667] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xadfab0 00:24:28.053 [2024-07-15 11:35:11.642679] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:28.053 [2024-07-15 11:35:11.643031] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:28.053 [2024-07-15 11:35:11.643050] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:24:28.053 [2024-07-15 11:35:11.643116] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:24:28.053 [2024-07-15 11:35:11.643129] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:24:28.053 [2024-07-15 11:35:11.643140] bdev_raid.c:3581:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:24:28.053 BaseBdev1 00:24:28.309 11:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # sleep 1 00:24:29.239 11:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:24:29.239 11:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:24:29.239 11:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:24:29.239 11:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:24:29.239 11:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:24:29.239 11:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:24:29.239 11:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:24:29.239 11:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:24:29.239 11:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:24:29.239 11:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:24:29.239 11:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:29.239 11:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:29.496 11:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:24:29.496 "name": "raid_bdev1", 00:24:29.496 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:29.496 "strip_size_kb": 0, 00:24:29.496 "state": "online", 00:24:29.496 "raid_level": "raid1", 00:24:29.496 "superblock": true, 00:24:29.496 "num_base_bdevs": 2, 00:24:29.496 "num_base_bdevs_discovered": 1, 00:24:29.496 "num_base_bdevs_operational": 1, 00:24:29.496 "base_bdevs_list": [ 00:24:29.496 { 00:24:29.496 "name": null, 00:24:29.496 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:29.496 "is_configured": false, 00:24:29.496 "data_offset": 2048, 00:24:29.496 "data_size": 63488 00:24:29.496 }, 00:24:29.496 { 00:24:29.496 "name": "BaseBdev2", 00:24:29.496 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:29.496 "is_configured": true, 00:24:29.496 "data_offset": 2048, 00:24:29.496 "data_size": 63488 00:24:29.496 } 00:24:29.496 ] 00:24:29.496 }' 00:24:29.496 11:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:24:29.496 11:35:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:24:30.061 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # verify_raid_bdev_process raid_bdev1 none none 00:24:30.061 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:30.061 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:24:30.061 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=none 00:24:30.061 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:30.061 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:30.061 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:30.319 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:30.319 "name": "raid_bdev1", 00:24:30.319 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:30.319 "strip_size_kb": 0, 00:24:30.319 "state": "online", 00:24:30.319 "raid_level": "raid1", 00:24:30.319 "superblock": true, 00:24:30.319 "num_base_bdevs": 2, 00:24:30.319 "num_base_bdevs_discovered": 1, 00:24:30.319 "num_base_bdevs_operational": 1, 00:24:30.319 "base_bdevs_list": [ 00:24:30.319 { 00:24:30.319 "name": null, 00:24:30.319 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:30.319 "is_configured": false, 00:24:30.319 "data_offset": 2048, 00:24:30.319 "data_size": 63488 00:24:30.319 }, 00:24:30.319 { 00:24:30.319 "name": "BaseBdev2", 00:24:30.319 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:30.319 "is_configured": true, 00:24:30.319 "data_offset": 2048, 00:24:30.319 "data_size": 63488 00:24:30.319 } 00:24:30.319 ] 00:24:30.319 }' 00:24:30.319 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:30.319 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:24:30.319 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:30.319 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:24:30.319 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:24:30.319 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@648 -- # local es=0 00:24:30.319 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:24:30.319 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:24:30.319 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:24:30.319 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:24:30.319 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:24:30.319 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:24:30.319 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:24:30.319 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:24:30.319 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py ]] 00:24:30.319 11:35:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:24:30.577 [2024-07-15 11:35:14.101440] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:24:30.577 [2024-07-15 11:35:14.101572] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:24:30.577 [2024-07-15 11:35:14.101587] bdev_raid.c:3581:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:24:30.577 request: 00:24:30.577 { 00:24:30.577 "base_bdev": "BaseBdev1", 00:24:30.577 "raid_bdev": "raid_bdev1", 00:24:30.577 "method": "bdev_raid_add_base_bdev", 00:24:30.577 "req_id": 1 00:24:30.577 } 00:24:30.577 Got JSON-RPC error response 00:24:30.577 response: 00:24:30.577 { 00:24:30.577 "code": -22, 00:24:30.577 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:24:30.577 } 00:24:30.577 11:35:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@651 -- # es=1 00:24:30.577 11:35:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:24:30.577 11:35:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:24:30.577 11:35:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:24:30.577 11:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # sleep 1 00:24:31.949 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:24:31.949 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:24:31.949 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:24:31.949 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:24:31.949 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:24:31.949 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:24:31.949 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:24:31.949 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:24:31.949 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:24:31.949 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:24:31.949 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:31.949 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:31.949 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:24:31.949 "name": "raid_bdev1", 00:24:31.949 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:31.949 "strip_size_kb": 0, 00:24:31.949 "state": "online", 00:24:31.949 "raid_level": "raid1", 00:24:31.949 "superblock": true, 00:24:31.949 "num_base_bdevs": 2, 00:24:31.949 "num_base_bdevs_discovered": 1, 00:24:31.949 "num_base_bdevs_operational": 1, 00:24:31.949 "base_bdevs_list": [ 00:24:31.949 { 00:24:31.949 "name": null, 00:24:31.949 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:31.949 "is_configured": false, 00:24:31.949 "data_offset": 2048, 00:24:31.949 "data_size": 63488 00:24:31.949 }, 00:24:31.949 { 00:24:31.949 "name": "BaseBdev2", 00:24:31.949 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:31.949 "is_configured": true, 00:24:31.949 "data_offset": 2048, 00:24:31.949 "data_size": 63488 00:24:31.949 } 00:24:31.949 ] 00:24:31.949 }' 00:24:31.949 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:24:31.949 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:24:32.512 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # verify_raid_bdev_process raid_bdev1 none none 00:24:32.512 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:32.512 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:24:32.512 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=none 00:24:32.512 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:32.512 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:32.512 11:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:32.770 11:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:32.770 "name": "raid_bdev1", 00:24:32.770 "uuid": "9a837ef6-b50b-4fd9-8176-92e5f6480842", 00:24:32.770 "strip_size_kb": 0, 00:24:32.770 "state": "online", 00:24:32.770 "raid_level": "raid1", 00:24:32.770 "superblock": true, 00:24:32.770 "num_base_bdevs": 2, 00:24:32.770 "num_base_bdevs_discovered": 1, 00:24:32.770 "num_base_bdevs_operational": 1, 00:24:32.770 "base_bdevs_list": [ 00:24:32.770 { 00:24:32.770 "name": null, 00:24:32.770 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:32.770 "is_configured": false, 00:24:32.770 "data_offset": 2048, 00:24:32.770 "data_size": 63488 00:24:32.770 }, 00:24:32.770 { 00:24:32.770 "name": "BaseBdev2", 00:24:32.770 "uuid": "6259c813-7ee9-57ef-8d52-1a9beb38aa86", 00:24:32.770 "is_configured": true, 00:24:32.770 "data_offset": 2048, 00:24:32.770 "data_size": 63488 00:24:32.770 } 00:24:32.770 ] 00:24:32.770 }' 00:24:32.770 11:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:32.770 11:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:24:32.770 11:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:32.770 11:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:24:32.770 11:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@782 -- # killprocess 979138 00:24:32.770 11:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@948 -- # '[' -z 979138 ']' 00:24:32.770 11:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@952 -- # kill -0 979138 00:24:32.770 11:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@953 -- # uname 00:24:32.770 11:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:24:32.770 11:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 979138 00:24:32.770 11:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:24:32.770 11:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:24:32.770 11:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@966 -- # echo 'killing process with pid 979138' 00:24:32.770 killing process with pid 979138 00:24:32.770 11:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@967 -- # kill 979138 00:24:32.770 Received shutdown signal, test time was about 26.845349 seconds 00:24:32.770 00:24:32.770 Latency(us) 00:24:32.770 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:32.770 =================================================================================================================== 00:24:32.770 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:24:32.770 [2024-07-15 11:35:16.310298] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:24:32.770 [2024-07-15 11:35:16.310395] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:24:32.770 [2024-07-15 11:35:16.310441] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:24:32.770 [2024-07-15 11:35:16.310453] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xc7e2f0 name raid_bdev1, state offline 00:24:32.770 11:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@972 -- # wait 979138 00:24:32.770 [2024-07-15 11:35:16.331070] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:24:33.028 11:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # return 0 00:24:33.028 00:24:33.028 real 0m31.496s 00:24:33.028 user 0m49.127s 00:24:33.028 sys 0m4.592s 00:24:33.028 11:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1124 -- # xtrace_disable 00:24:33.028 11:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:24:33.028 ************************************ 00:24:33.028 END TEST raid_rebuild_test_sb_io 00:24:33.028 ************************************ 00:24:33.028 11:35:16 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:24:33.028 11:35:16 bdev_raid -- bdev/bdev_raid.sh@876 -- # for n in 2 4 00:24:33.028 11:35:16 bdev_raid -- bdev/bdev_raid.sh@877 -- # run_test raid_rebuild_test raid_rebuild_test raid1 4 false false true 00:24:33.028 11:35:16 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:24:33.028 11:35:16 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:24:33.028 11:35:16 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:24:33.285 ************************************ 00:24:33.285 START TEST raid_rebuild_test 00:24:33.285 ************************************ 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1123 -- # raid_rebuild_test raid1 4 false false true 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@568 -- # local raid_level=raid1 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local num_base_bdevs=4 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local superblock=false 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local background_io=false 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local verify=true 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # (( i = 1 )) 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # echo BaseBdev1 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # echo BaseBdev2 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # echo BaseBdev3 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # echo BaseBdev4 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local base_bdevs 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local raid_bdev_name=raid_bdev1 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local strip_size 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local create_arg 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local raid_bdev_size 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local data_offset 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@580 -- # '[' raid1 '!=' raid1 ']' 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@588 -- # strip_size=0 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@591 -- # '[' false = true ']' 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # raid_pid=983787 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # waitforlisten 983787 /var/tmp/spdk-raid.sock 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@595 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@829 -- # '[' -z 983787 ']' 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@834 -- # local max_retries=100 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:24:33.285 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@838 -- # xtrace_disable 00:24:33.285 11:35:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:24:33.285 [2024-07-15 11:35:16.694815] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:24:33.285 [2024-07-15 11:35:16.694880] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid983787 ] 00:24:33.285 I/O size of 3145728 is greater than zero copy threshold (65536). 00:24:33.285 Zero copy mechanism will not be used. 00:24:33.285 [2024-07-15 11:35:16.825812] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:33.543 [2024-07-15 11:35:16.933087] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:24:33.543 [2024-07-15 11:35:17.002806] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:24:33.543 [2024-07-15 11:35:17.002844] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:24:34.105 11:35:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:24:34.105 11:35:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@862 -- # return 0 00:24:34.105 11:35:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:24:34.105 11:35:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:24:34.361 BaseBdev1_malloc 00:24:34.361 11:35:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:24:34.618 [2024-07-15 11:35:18.085156] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:24:34.618 [2024-07-15 11:35:18.085203] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:34.618 [2024-07-15 11:35:18.085227] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2139d40 00:24:34.618 [2024-07-15 11:35:18.085239] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:34.618 [2024-07-15 11:35:18.086960] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:34.618 [2024-07-15 11:35:18.086989] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:24:34.618 BaseBdev1 00:24:34.618 11:35:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:24:34.618 11:35:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:24:34.875 BaseBdev2_malloc 00:24:34.875 11:35:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:24:35.131 [2024-07-15 11:35:18.583341] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:24:35.131 [2024-07-15 11:35:18.583387] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:35.131 [2024-07-15 11:35:18.583414] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x213a860 00:24:35.131 [2024-07-15 11:35:18.583426] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:35.131 [2024-07-15 11:35:18.585018] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:35.131 [2024-07-15 11:35:18.585061] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:24:35.131 BaseBdev2 00:24:35.131 11:35:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:24:35.131 11:35:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:24:35.388 BaseBdev3_malloc 00:24:35.388 11:35:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:24:35.644 [2024-07-15 11:35:19.074518] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:24:35.644 [2024-07-15 11:35:19.074563] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:35.644 [2024-07-15 11:35:19.074586] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x22e78f0 00:24:35.644 [2024-07-15 11:35:19.074598] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:35.644 [2024-07-15 11:35:19.076178] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:35.644 [2024-07-15 11:35:19.076207] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:24:35.644 BaseBdev3 00:24:35.644 11:35:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:24:35.644 11:35:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:24:35.901 BaseBdev4_malloc 00:24:35.901 11:35:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:24:36.158 [2024-07-15 11:35:19.569692] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:24:36.158 [2024-07-15 11:35:19.569737] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:36.158 [2024-07-15 11:35:19.569759] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x22e6ad0 00:24:36.158 [2024-07-15 11:35:19.569772] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:36.158 [2024-07-15 11:35:19.571309] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:36.158 [2024-07-15 11:35:19.571338] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:24:36.158 BaseBdev4 00:24:36.158 11:35:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@606 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b spare_malloc 00:24:36.414 spare_malloc 00:24:36.414 11:35:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:24:36.671 spare_delay 00:24:36.671 11:35:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:24:36.928 [2024-07-15 11:35:20.300157] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:24:36.928 [2024-07-15 11:35:20.300201] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:36.928 [2024-07-15 11:35:20.300223] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x22eb5b0 00:24:36.928 [2024-07-15 11:35:20.300236] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:36.928 [2024-07-15 11:35:20.301799] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:36.928 [2024-07-15 11:35:20.301826] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:24:36.928 spare 00:24:36.928 11:35:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@611 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n raid_bdev1 00:24:37.184 [2024-07-15 11:35:20.536806] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:24:37.184 [2024-07-15 11:35:20.538145] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:24:37.184 [2024-07-15 11:35:20.538199] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:24:37.184 [2024-07-15 11:35:20.538245] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:24:37.184 [2024-07-15 11:35:20.538333] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x226a8a0 00:24:37.184 [2024-07-15 11:35:20.538344] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:24:37.184 [2024-07-15 11:35:20.538560] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x22e4e10 00:24:37.184 [2024-07-15 11:35:20.538712] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x226a8a0 00:24:37.184 [2024-07-15 11:35:20.538722] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x226a8a0 00:24:37.184 [2024-07-15 11:35:20.538837] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:37.184 11:35:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:24:37.184 11:35:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:24:37.184 11:35:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:24:37.184 11:35:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:24:37.184 11:35:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:24:37.184 11:35:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:24:37.184 11:35:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:24:37.184 11:35:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:24:37.184 11:35:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:24:37.184 11:35:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:24:37.184 11:35:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:37.184 11:35:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:37.441 11:35:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:24:37.441 "name": "raid_bdev1", 00:24:37.441 "uuid": "b1d1ae77-789d-45eb-a775-6aa5d7906eb6", 00:24:37.441 "strip_size_kb": 0, 00:24:37.441 "state": "online", 00:24:37.441 "raid_level": "raid1", 00:24:37.441 "superblock": false, 00:24:37.441 "num_base_bdevs": 4, 00:24:37.441 "num_base_bdevs_discovered": 4, 00:24:37.441 "num_base_bdevs_operational": 4, 00:24:37.441 "base_bdevs_list": [ 00:24:37.441 { 00:24:37.441 "name": "BaseBdev1", 00:24:37.441 "uuid": "badd91ed-00b3-5673-97be-a9a32c651c82", 00:24:37.441 "is_configured": true, 00:24:37.441 "data_offset": 0, 00:24:37.441 "data_size": 65536 00:24:37.441 }, 00:24:37.441 { 00:24:37.441 "name": "BaseBdev2", 00:24:37.441 "uuid": "3b4caf0e-90d3-5e18-a179-b444b29f2897", 00:24:37.441 "is_configured": true, 00:24:37.441 "data_offset": 0, 00:24:37.441 "data_size": 65536 00:24:37.441 }, 00:24:37.441 { 00:24:37.441 "name": "BaseBdev3", 00:24:37.441 "uuid": "0dcaf4fc-08c1-52ef-8172-72c75b7f1f5e", 00:24:37.441 "is_configured": true, 00:24:37.441 "data_offset": 0, 00:24:37.441 "data_size": 65536 00:24:37.441 }, 00:24:37.441 { 00:24:37.441 "name": "BaseBdev4", 00:24:37.441 "uuid": "e5516be1-29db-5522-8173-34a697764450", 00:24:37.441 "is_configured": true, 00:24:37.441 "data_offset": 0, 00:24:37.441 "data_size": 65536 00:24:37.441 } 00:24:37.441 ] 00:24:37.441 }' 00:24:37.441 11:35:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:24:37.441 11:35:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:24:38.005 11:35:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@615 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:24:38.005 11:35:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@615 -- # jq -r '.[].num_blocks' 00:24:38.005 [2024-07-15 11:35:21.543745] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:24:38.005 11:35:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@615 -- # raid_bdev_size=65536 00:24:38.005 11:35:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@618 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:24:38.005 11:35:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@618 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:38.263 11:35:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@618 -- # data_offset=0 00:24:38.263 11:35:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@620 -- # '[' false = true ']' 00:24:38.264 11:35:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@623 -- # '[' true = true ']' 00:24:38.264 11:35:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # local write_unit_size 00:24:38.264 11:35:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@627 -- # nbd_start_disks /var/tmp/spdk-raid.sock raid_bdev1 /dev/nbd0 00:24:38.264 11:35:21 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:24:38.264 11:35:21 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:24:38.264 11:35:21 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:24:38.264 11:35:21 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:24:38.264 11:35:21 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:24:38.264 11:35:21 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:24:38.264 11:35:21 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:24:38.264 11:35:21 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:24:38.264 11:35:21 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:24:38.522 [2024-07-15 11:35:22.044817] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x22e4e10 00:24:38.522 /dev/nbd0 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@867 -- # local i 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # break 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:24:38.522 1+0 records in 00:24:38.522 1+0 records out 00:24:38.522 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000187722 s, 21.8 MB/s 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # size=4096 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # return 0 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # '[' raid1 = raid5f ']' 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@632 -- # write_unit_size=1 00:24:38.522 11:35:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@634 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:24:46.630 65536+0 records in 00:24:46.630 65536+0 records out 00:24:46.630 33554432 bytes (34 MB, 32 MiB) copied, 7.39302 s, 4.5 MB/s 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # nbd_stop_disks /var/tmp/spdk-raid.sock /dev/nbd0 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd0 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:24:46.630 [2024-07-15 11:35:29.771321] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@639 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev1 00:24:46.630 [2024-07-15 11:35:29.935809] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@642 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:24:46.630 11:35:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:24:46.631 11:35:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:24:46.631 11:35:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:46.631 11:35:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:46.631 11:35:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:24:46.631 "name": "raid_bdev1", 00:24:46.631 "uuid": "b1d1ae77-789d-45eb-a775-6aa5d7906eb6", 00:24:46.631 "strip_size_kb": 0, 00:24:46.631 "state": "online", 00:24:46.631 "raid_level": "raid1", 00:24:46.631 "superblock": false, 00:24:46.631 "num_base_bdevs": 4, 00:24:46.631 "num_base_bdevs_discovered": 3, 00:24:46.631 "num_base_bdevs_operational": 3, 00:24:46.631 "base_bdevs_list": [ 00:24:46.631 { 00:24:46.631 "name": null, 00:24:46.631 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:46.631 "is_configured": false, 00:24:46.631 "data_offset": 0, 00:24:46.631 "data_size": 65536 00:24:46.631 }, 00:24:46.631 { 00:24:46.631 "name": "BaseBdev2", 00:24:46.631 "uuid": "3b4caf0e-90d3-5e18-a179-b444b29f2897", 00:24:46.631 "is_configured": true, 00:24:46.631 "data_offset": 0, 00:24:46.631 "data_size": 65536 00:24:46.631 }, 00:24:46.631 { 00:24:46.631 "name": "BaseBdev3", 00:24:46.631 "uuid": "0dcaf4fc-08c1-52ef-8172-72c75b7f1f5e", 00:24:46.631 "is_configured": true, 00:24:46.631 "data_offset": 0, 00:24:46.631 "data_size": 65536 00:24:46.631 }, 00:24:46.631 { 00:24:46.631 "name": "BaseBdev4", 00:24:46.631 "uuid": "e5516be1-29db-5522-8173-34a697764450", 00:24:46.631 "is_configured": true, 00:24:46.631 "data_offset": 0, 00:24:46.631 "data_size": 65536 00:24:46.631 } 00:24:46.631 ] 00:24:46.631 }' 00:24:46.631 11:35:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:24:46.631 11:35:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:24:47.197 11:35:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@645 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:24:47.455 [2024-07-15 11:35:30.958540] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:24:47.455 [2024-07-15 11:35:30.962639] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x22706b0 00:24:47.455 [2024-07-15 11:35:30.965003] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:24:47.455 11:35:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # sleep 1 00:24:48.420 11:35:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@649 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:24:48.420 11:35:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:48.420 11:35:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:24:48.420 11:35:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@184 -- # local target=spare 00:24:48.420 11:35:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:48.420 11:35:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:48.420 11:35:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:48.688 11:35:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:48.688 "name": "raid_bdev1", 00:24:48.688 "uuid": "b1d1ae77-789d-45eb-a775-6aa5d7906eb6", 00:24:48.688 "strip_size_kb": 0, 00:24:48.688 "state": "online", 00:24:48.688 "raid_level": "raid1", 00:24:48.688 "superblock": false, 00:24:48.688 "num_base_bdevs": 4, 00:24:48.688 "num_base_bdevs_discovered": 4, 00:24:48.688 "num_base_bdevs_operational": 4, 00:24:48.688 "process": { 00:24:48.688 "type": "rebuild", 00:24:48.688 "target": "spare", 00:24:48.688 "progress": { 00:24:48.688 "blocks": 22528, 00:24:48.688 "percent": 34 00:24:48.688 } 00:24:48.688 }, 00:24:48.688 "base_bdevs_list": [ 00:24:48.688 { 00:24:48.688 "name": "spare", 00:24:48.688 "uuid": "98f9d43e-f4f2-5544-b80d-6cf0b183ba9d", 00:24:48.688 "is_configured": true, 00:24:48.688 "data_offset": 0, 00:24:48.688 "data_size": 65536 00:24:48.688 }, 00:24:48.688 { 00:24:48.688 "name": "BaseBdev2", 00:24:48.688 "uuid": "3b4caf0e-90d3-5e18-a179-b444b29f2897", 00:24:48.688 "is_configured": true, 00:24:48.688 "data_offset": 0, 00:24:48.688 "data_size": 65536 00:24:48.688 }, 00:24:48.688 { 00:24:48.688 "name": "BaseBdev3", 00:24:48.688 "uuid": "0dcaf4fc-08c1-52ef-8172-72c75b7f1f5e", 00:24:48.688 "is_configured": true, 00:24:48.688 "data_offset": 0, 00:24:48.688 "data_size": 65536 00:24:48.688 }, 00:24:48.688 { 00:24:48.688 "name": "BaseBdev4", 00:24:48.688 "uuid": "e5516be1-29db-5522-8173-34a697764450", 00:24:48.688 "is_configured": true, 00:24:48.688 "data_offset": 0, 00:24:48.688 "data_size": 65536 00:24:48.688 } 00:24:48.688 ] 00:24:48.688 }' 00:24:48.688 11:35:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:48.688 11:35:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:24:48.688 11:35:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:48.688 11:35:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:24:48.689 11:35:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@652 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev spare 00:24:48.946 [2024-07-15 11:35:32.475549] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:24:48.947 [2024-07-15 11:35:32.476416] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:24:48.947 [2024-07-15 11:35:32.476458] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:48.947 [2024-07-15 11:35:32.476475] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:24:48.947 [2024-07-15 11:35:32.476484] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:24:48.947 11:35:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@655 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:24:48.947 11:35:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:24:48.947 11:35:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:24:48.947 11:35:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:24:48.947 11:35:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:24:48.947 11:35:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:24:48.947 11:35:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:24:48.947 11:35:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:24:48.947 11:35:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:24:48.947 11:35:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:24:48.947 11:35:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:48.947 11:35:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:49.205 11:35:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:24:49.205 "name": "raid_bdev1", 00:24:49.205 "uuid": "b1d1ae77-789d-45eb-a775-6aa5d7906eb6", 00:24:49.205 "strip_size_kb": 0, 00:24:49.205 "state": "online", 00:24:49.205 "raid_level": "raid1", 00:24:49.205 "superblock": false, 00:24:49.205 "num_base_bdevs": 4, 00:24:49.205 "num_base_bdevs_discovered": 3, 00:24:49.205 "num_base_bdevs_operational": 3, 00:24:49.205 "base_bdevs_list": [ 00:24:49.205 { 00:24:49.205 "name": null, 00:24:49.205 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:49.205 "is_configured": false, 00:24:49.205 "data_offset": 0, 00:24:49.205 "data_size": 65536 00:24:49.205 }, 00:24:49.205 { 00:24:49.205 "name": "BaseBdev2", 00:24:49.205 "uuid": "3b4caf0e-90d3-5e18-a179-b444b29f2897", 00:24:49.205 "is_configured": true, 00:24:49.205 "data_offset": 0, 00:24:49.205 "data_size": 65536 00:24:49.205 }, 00:24:49.205 { 00:24:49.205 "name": "BaseBdev3", 00:24:49.205 "uuid": "0dcaf4fc-08c1-52ef-8172-72c75b7f1f5e", 00:24:49.205 "is_configured": true, 00:24:49.205 "data_offset": 0, 00:24:49.205 "data_size": 65536 00:24:49.205 }, 00:24:49.205 { 00:24:49.205 "name": "BaseBdev4", 00:24:49.205 "uuid": "e5516be1-29db-5522-8173-34a697764450", 00:24:49.205 "is_configured": true, 00:24:49.205 "data_offset": 0, 00:24:49.205 "data_size": 65536 00:24:49.205 } 00:24:49.205 ] 00:24:49.205 }' 00:24:49.205 11:35:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:24:49.205 11:35:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:24:49.771 11:35:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@658 -- # verify_raid_bdev_process raid_bdev1 none none 00:24:49.771 11:35:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:49.771 11:35:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:24:49.771 11:35:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@184 -- # local target=none 00:24:49.771 11:35:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:49.771 11:35:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:49.771 11:35:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:50.028 11:35:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:50.028 "name": "raid_bdev1", 00:24:50.028 "uuid": "b1d1ae77-789d-45eb-a775-6aa5d7906eb6", 00:24:50.028 "strip_size_kb": 0, 00:24:50.028 "state": "online", 00:24:50.028 "raid_level": "raid1", 00:24:50.028 "superblock": false, 00:24:50.028 "num_base_bdevs": 4, 00:24:50.028 "num_base_bdevs_discovered": 3, 00:24:50.028 "num_base_bdevs_operational": 3, 00:24:50.028 "base_bdevs_list": [ 00:24:50.028 { 00:24:50.028 "name": null, 00:24:50.028 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:50.028 "is_configured": false, 00:24:50.028 "data_offset": 0, 00:24:50.028 "data_size": 65536 00:24:50.028 }, 00:24:50.028 { 00:24:50.028 "name": "BaseBdev2", 00:24:50.028 "uuid": "3b4caf0e-90d3-5e18-a179-b444b29f2897", 00:24:50.028 "is_configured": true, 00:24:50.028 "data_offset": 0, 00:24:50.028 "data_size": 65536 00:24:50.028 }, 00:24:50.028 { 00:24:50.028 "name": "BaseBdev3", 00:24:50.028 "uuid": "0dcaf4fc-08c1-52ef-8172-72c75b7f1f5e", 00:24:50.028 "is_configured": true, 00:24:50.028 "data_offset": 0, 00:24:50.028 "data_size": 65536 00:24:50.028 }, 00:24:50.028 { 00:24:50.028 "name": "BaseBdev4", 00:24:50.028 "uuid": "e5516be1-29db-5522-8173-34a697764450", 00:24:50.028 "is_configured": true, 00:24:50.028 "data_offset": 0, 00:24:50.028 "data_size": 65536 00:24:50.028 } 00:24:50.028 ] 00:24:50.028 }' 00:24:50.029 11:35:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:50.029 11:35:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:24:50.029 11:35:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:50.029 11:35:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:24:50.029 11:35:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@661 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:24:50.286 [2024-07-15 11:35:33.775540] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:24:50.286 [2024-07-15 11:35:33.780176] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x22706b0 00:24:50.286 [2024-07-15 11:35:33.781721] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:24:50.286 11:35:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # sleep 1 00:24:51.216 11:35:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:24:51.216 11:35:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:51.216 11:35:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:24:51.216 11:35:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@184 -- # local target=spare 00:24:51.216 11:35:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:51.216 11:35:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:51.216 11:35:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:51.473 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:51.473 "name": "raid_bdev1", 00:24:51.473 "uuid": "b1d1ae77-789d-45eb-a775-6aa5d7906eb6", 00:24:51.473 "strip_size_kb": 0, 00:24:51.473 "state": "online", 00:24:51.473 "raid_level": "raid1", 00:24:51.473 "superblock": false, 00:24:51.473 "num_base_bdevs": 4, 00:24:51.473 "num_base_bdevs_discovered": 4, 00:24:51.473 "num_base_bdevs_operational": 4, 00:24:51.473 "process": { 00:24:51.473 "type": "rebuild", 00:24:51.473 "target": "spare", 00:24:51.473 "progress": { 00:24:51.473 "blocks": 24576, 00:24:51.473 "percent": 37 00:24:51.473 } 00:24:51.473 }, 00:24:51.473 "base_bdevs_list": [ 00:24:51.473 { 00:24:51.473 "name": "spare", 00:24:51.473 "uuid": "98f9d43e-f4f2-5544-b80d-6cf0b183ba9d", 00:24:51.473 "is_configured": true, 00:24:51.473 "data_offset": 0, 00:24:51.473 "data_size": 65536 00:24:51.473 }, 00:24:51.473 { 00:24:51.473 "name": "BaseBdev2", 00:24:51.473 "uuid": "3b4caf0e-90d3-5e18-a179-b444b29f2897", 00:24:51.473 "is_configured": true, 00:24:51.473 "data_offset": 0, 00:24:51.473 "data_size": 65536 00:24:51.473 }, 00:24:51.473 { 00:24:51.473 "name": "BaseBdev3", 00:24:51.473 "uuid": "0dcaf4fc-08c1-52ef-8172-72c75b7f1f5e", 00:24:51.473 "is_configured": true, 00:24:51.473 "data_offset": 0, 00:24:51.473 "data_size": 65536 00:24:51.473 }, 00:24:51.473 { 00:24:51.473 "name": "BaseBdev4", 00:24:51.473 "uuid": "e5516be1-29db-5522-8173-34a697764450", 00:24:51.473 "is_configured": true, 00:24:51.473 "data_offset": 0, 00:24:51.473 "data_size": 65536 00:24:51.473 } 00:24:51.473 ] 00:24:51.473 }' 00:24:51.473 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:51.730 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:24:51.730 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:51.730 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:24:51.730 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@665 -- # '[' false = true ']' 00:24:51.730 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@690 -- # local num_base_bdevs_operational=4 00:24:51.730 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@692 -- # '[' raid1 = raid1 ']' 00:24:51.730 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@692 -- # '[' 4 -gt 2 ']' 00:24:51.730 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@694 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev2 00:24:51.987 [2024-07-15 11:35:35.381294] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:24:51.987 [2024-07-15 11:35:35.394189] bdev_raid.c:1919:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x22706b0 00:24:51.987 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@697 -- # base_bdevs[1]= 00:24:51.987 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@698 -- # (( num_base_bdevs_operational-- )) 00:24:51.987 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@701 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:24:51.987 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:51.987 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:24:51.987 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@184 -- # local target=spare 00:24:51.987 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:51.987 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:51.987 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:52.243 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:52.243 "name": "raid_bdev1", 00:24:52.243 "uuid": "b1d1ae77-789d-45eb-a775-6aa5d7906eb6", 00:24:52.243 "strip_size_kb": 0, 00:24:52.243 "state": "online", 00:24:52.243 "raid_level": "raid1", 00:24:52.243 "superblock": false, 00:24:52.243 "num_base_bdevs": 4, 00:24:52.243 "num_base_bdevs_discovered": 3, 00:24:52.243 "num_base_bdevs_operational": 3, 00:24:52.243 "process": { 00:24:52.243 "type": "rebuild", 00:24:52.243 "target": "spare", 00:24:52.243 "progress": { 00:24:52.243 "blocks": 36864, 00:24:52.243 "percent": 56 00:24:52.243 } 00:24:52.243 }, 00:24:52.243 "base_bdevs_list": [ 00:24:52.243 { 00:24:52.243 "name": "spare", 00:24:52.243 "uuid": "98f9d43e-f4f2-5544-b80d-6cf0b183ba9d", 00:24:52.243 "is_configured": true, 00:24:52.243 "data_offset": 0, 00:24:52.243 "data_size": 65536 00:24:52.243 }, 00:24:52.243 { 00:24:52.243 "name": null, 00:24:52.243 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:52.243 "is_configured": false, 00:24:52.243 "data_offset": 0, 00:24:52.243 "data_size": 65536 00:24:52.243 }, 00:24:52.243 { 00:24:52.243 "name": "BaseBdev3", 00:24:52.243 "uuid": "0dcaf4fc-08c1-52ef-8172-72c75b7f1f5e", 00:24:52.243 "is_configured": true, 00:24:52.243 "data_offset": 0, 00:24:52.243 "data_size": 65536 00:24:52.243 }, 00:24:52.243 { 00:24:52.243 "name": "BaseBdev4", 00:24:52.243 "uuid": "e5516be1-29db-5522-8173-34a697764450", 00:24:52.243 "is_configured": true, 00:24:52.243 "data_offset": 0, 00:24:52.243 "data_size": 65536 00:24:52.243 } 00:24:52.243 ] 00:24:52.243 }' 00:24:52.243 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:52.243 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:24:52.243 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:52.243 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:24:52.243 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@705 -- # local timeout=863 00:24:52.243 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:24:52.243 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:24:52.243 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:52.243 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:24:52.243 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@184 -- # local target=spare 00:24:52.243 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:52.243 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:52.243 11:35:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:52.501 11:35:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:52.501 "name": "raid_bdev1", 00:24:52.501 "uuid": "b1d1ae77-789d-45eb-a775-6aa5d7906eb6", 00:24:52.501 "strip_size_kb": 0, 00:24:52.501 "state": "online", 00:24:52.501 "raid_level": "raid1", 00:24:52.501 "superblock": false, 00:24:52.501 "num_base_bdevs": 4, 00:24:52.501 "num_base_bdevs_discovered": 3, 00:24:52.501 "num_base_bdevs_operational": 3, 00:24:52.501 "process": { 00:24:52.501 "type": "rebuild", 00:24:52.501 "target": "spare", 00:24:52.501 "progress": { 00:24:52.501 "blocks": 43008, 00:24:52.501 "percent": 65 00:24:52.501 } 00:24:52.501 }, 00:24:52.501 "base_bdevs_list": [ 00:24:52.501 { 00:24:52.501 "name": "spare", 00:24:52.501 "uuid": "98f9d43e-f4f2-5544-b80d-6cf0b183ba9d", 00:24:52.501 "is_configured": true, 00:24:52.501 "data_offset": 0, 00:24:52.501 "data_size": 65536 00:24:52.501 }, 00:24:52.501 { 00:24:52.501 "name": null, 00:24:52.501 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:52.501 "is_configured": false, 00:24:52.501 "data_offset": 0, 00:24:52.501 "data_size": 65536 00:24:52.501 }, 00:24:52.501 { 00:24:52.501 "name": "BaseBdev3", 00:24:52.501 "uuid": "0dcaf4fc-08c1-52ef-8172-72c75b7f1f5e", 00:24:52.501 "is_configured": true, 00:24:52.501 "data_offset": 0, 00:24:52.501 "data_size": 65536 00:24:52.501 }, 00:24:52.501 { 00:24:52.501 "name": "BaseBdev4", 00:24:52.501 "uuid": "e5516be1-29db-5522-8173-34a697764450", 00:24:52.501 "is_configured": true, 00:24:52.501 "data_offset": 0, 00:24:52.501 "data_size": 65536 00:24:52.501 } 00:24:52.501 ] 00:24:52.501 }' 00:24:52.501 11:35:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:52.501 11:35:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:24:52.501 11:35:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:52.758 11:35:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:24:52.758 11:35:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@710 -- # sleep 1 00:24:53.688 [2024-07-15 11:35:37.006860] bdev_raid.c:2789:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:24:53.688 [2024-07-15 11:35:37.006924] bdev_raid.c:2504:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:24:53.688 [2024-07-15 11:35:37.006968] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:53.688 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:24:53.688 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:24:53.688 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:53.688 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:24:53.688 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@184 -- # local target=spare 00:24:53.688 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:53.688 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:53.688 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:53.946 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:53.946 "name": "raid_bdev1", 00:24:53.946 "uuid": "b1d1ae77-789d-45eb-a775-6aa5d7906eb6", 00:24:53.946 "strip_size_kb": 0, 00:24:53.946 "state": "online", 00:24:53.946 "raid_level": "raid1", 00:24:53.946 "superblock": false, 00:24:53.946 "num_base_bdevs": 4, 00:24:53.946 "num_base_bdevs_discovered": 3, 00:24:53.946 "num_base_bdevs_operational": 3, 00:24:53.946 "base_bdevs_list": [ 00:24:53.946 { 00:24:53.946 "name": "spare", 00:24:53.946 "uuid": "98f9d43e-f4f2-5544-b80d-6cf0b183ba9d", 00:24:53.946 "is_configured": true, 00:24:53.946 "data_offset": 0, 00:24:53.946 "data_size": 65536 00:24:53.946 }, 00:24:53.946 { 00:24:53.946 "name": null, 00:24:53.946 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:53.946 "is_configured": false, 00:24:53.946 "data_offset": 0, 00:24:53.946 "data_size": 65536 00:24:53.946 }, 00:24:53.946 { 00:24:53.946 "name": "BaseBdev3", 00:24:53.946 "uuid": "0dcaf4fc-08c1-52ef-8172-72c75b7f1f5e", 00:24:53.946 "is_configured": true, 00:24:53.946 "data_offset": 0, 00:24:53.946 "data_size": 65536 00:24:53.946 }, 00:24:53.946 { 00:24:53.946 "name": "BaseBdev4", 00:24:53.946 "uuid": "e5516be1-29db-5522-8173-34a697764450", 00:24:53.946 "is_configured": true, 00:24:53.946 "data_offset": 0, 00:24:53.946 "data_size": 65536 00:24:53.946 } 00:24:53.946 ] 00:24:53.946 }' 00:24:53.946 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:53.946 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # [[ none == \r\e\b\u\i\l\d ]] 00:24:53.946 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:53.947 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # [[ none == \s\p\a\r\e ]] 00:24:53.947 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # break 00:24:53.947 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@714 -- # verify_raid_bdev_process raid_bdev1 none none 00:24:53.947 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:24:53.947 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:24:53.947 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@184 -- # local target=none 00:24:53.947 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:24:53.947 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:53.947 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:54.204 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:54.204 "name": "raid_bdev1", 00:24:54.204 "uuid": "b1d1ae77-789d-45eb-a775-6aa5d7906eb6", 00:24:54.204 "strip_size_kb": 0, 00:24:54.204 "state": "online", 00:24:54.204 "raid_level": "raid1", 00:24:54.204 "superblock": false, 00:24:54.204 "num_base_bdevs": 4, 00:24:54.204 "num_base_bdevs_discovered": 3, 00:24:54.204 "num_base_bdevs_operational": 3, 00:24:54.204 "base_bdevs_list": [ 00:24:54.204 { 00:24:54.204 "name": "spare", 00:24:54.204 "uuid": "98f9d43e-f4f2-5544-b80d-6cf0b183ba9d", 00:24:54.204 "is_configured": true, 00:24:54.204 "data_offset": 0, 00:24:54.204 "data_size": 65536 00:24:54.204 }, 00:24:54.204 { 00:24:54.204 "name": null, 00:24:54.204 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:54.205 "is_configured": false, 00:24:54.205 "data_offset": 0, 00:24:54.205 "data_size": 65536 00:24:54.205 }, 00:24:54.205 { 00:24:54.205 "name": "BaseBdev3", 00:24:54.205 "uuid": "0dcaf4fc-08c1-52ef-8172-72c75b7f1f5e", 00:24:54.205 "is_configured": true, 00:24:54.205 "data_offset": 0, 00:24:54.205 "data_size": 65536 00:24:54.205 }, 00:24:54.205 { 00:24:54.205 "name": "BaseBdev4", 00:24:54.205 "uuid": "e5516be1-29db-5522-8173-34a697764450", 00:24:54.205 "is_configured": true, 00:24:54.205 "data_offset": 0, 00:24:54.205 "data_size": 65536 00:24:54.205 } 00:24:54.205 ] 00:24:54.205 }' 00:24:54.205 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:24:54.205 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:24:54.205 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:24:54.205 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:24:54.205 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:24:54.205 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:24:54.205 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:24:54.205 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:24:54.205 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:24:54.205 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:24:54.205 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:24:54.205 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:24:54.205 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:24:54.205 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@124 -- # local tmp 00:24:54.205 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:54.205 11:35:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:54.463 11:35:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:24:54.463 "name": "raid_bdev1", 00:24:54.463 "uuid": "b1d1ae77-789d-45eb-a775-6aa5d7906eb6", 00:24:54.463 "strip_size_kb": 0, 00:24:54.463 "state": "online", 00:24:54.463 "raid_level": "raid1", 00:24:54.463 "superblock": false, 00:24:54.463 "num_base_bdevs": 4, 00:24:54.463 "num_base_bdevs_discovered": 3, 00:24:54.463 "num_base_bdevs_operational": 3, 00:24:54.463 "base_bdevs_list": [ 00:24:54.463 { 00:24:54.463 "name": "spare", 00:24:54.463 "uuid": "98f9d43e-f4f2-5544-b80d-6cf0b183ba9d", 00:24:54.463 "is_configured": true, 00:24:54.463 "data_offset": 0, 00:24:54.463 "data_size": 65536 00:24:54.463 }, 00:24:54.463 { 00:24:54.463 "name": null, 00:24:54.463 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:54.463 "is_configured": false, 00:24:54.463 "data_offset": 0, 00:24:54.463 "data_size": 65536 00:24:54.463 }, 00:24:54.463 { 00:24:54.463 "name": "BaseBdev3", 00:24:54.463 "uuid": "0dcaf4fc-08c1-52ef-8172-72c75b7f1f5e", 00:24:54.463 "is_configured": true, 00:24:54.463 "data_offset": 0, 00:24:54.463 "data_size": 65536 00:24:54.463 }, 00:24:54.463 { 00:24:54.463 "name": "BaseBdev4", 00:24:54.463 "uuid": "e5516be1-29db-5522-8173-34a697764450", 00:24:54.463 "is_configured": true, 00:24:54.463 "data_offset": 0, 00:24:54.463 "data_size": 65536 00:24:54.463 } 00:24:54.463 ] 00:24:54.463 }' 00:24:54.463 11:35:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:24:54.463 11:35:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:24:55.030 11:35:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@718 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:24:55.288 [2024-07-15 11:35:38.840010] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:24:55.288 [2024-07-15 11:35:38.840037] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:24:55.288 [2024-07-15 11:35:38.840098] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:24:55.288 [2024-07-15 11:35:38.840167] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:24:55.288 [2024-07-15 11:35:38.840178] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x226a8a0 name raid_bdev1, state offline 00:24:55.288 11:35:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:24:55.288 11:35:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # jq length 00:24:55.545 11:35:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # [[ 0 == 0 ]] 00:24:55.545 11:35:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@721 -- # '[' true = true ']' 00:24:55.545 11:35:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' false = true ']' 00:24:55.545 11:35:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@736 -- # nbd_start_disks /var/tmp/spdk-raid.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:24:55.545 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:24:55.545 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:24:55.545 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:24:55.545 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:24:55.545 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:24:55.545 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:24:55.545 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:24:55.545 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:24:55.545 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:24:55.804 /dev/nbd0 00:24:55.804 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:24:55.804 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:24:55.804 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:24:55.804 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@867 -- # local i 00:24:55.804 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:24:55.804 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:24:55.804 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:24:55.804 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # break 00:24:55.804 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:24:55.804 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:24:55.804 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:24:55.804 1+0 records in 00:24:55.804 1+0 records out 00:24:55.804 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00023244 s, 17.6 MB/s 00:24:55.804 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:24:55.804 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # size=4096 00:24:55.804 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:24:55.804 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:24:55.804 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # return 0 00:24:55.804 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:24:55.804 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:24:55.804 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk spare /dev/nbd1 00:24:56.062 /dev/nbd1 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@867 -- # local i 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # break 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:24:56.319 1+0 records in 00:24:56.319 1+0 records out 00:24:56.319 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000457929 s, 8.9 MB/s 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # size=4096 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # return 0 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # nbd_stop_disks /var/tmp/spdk-raid.sock '/dev/nbd0 /dev/nbd1' 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:24:56.319 11:35:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd0 00:24:56.577 11:35:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:24:56.577 11:35:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:24:56.577 11:35:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:24:56.577 11:35:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:24:56.577 11:35:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:24:56.577 11:35:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:24:56.577 11:35:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:24:56.577 11:35:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:24:56.577 11:35:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:24:56.577 11:35:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd1 00:24:56.834 11:35:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:24:56.834 11:35:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:24:56.834 11:35:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:24:56.834 11:35:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:24:56.834 11:35:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:24:56.834 11:35:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:24:56.834 11:35:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:24:56.834 11:35:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:24:56.834 11:35:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@742 -- # '[' false = true ']' 00:24:56.835 11:35:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@782 -- # killprocess 983787 00:24:56.835 11:35:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@948 -- # '[' -z 983787 ']' 00:24:56.835 11:35:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@952 -- # kill -0 983787 00:24:56.835 11:35:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@953 -- # uname 00:24:56.835 11:35:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:24:56.835 11:35:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 983787 00:24:56.835 11:35:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:24:56.835 11:35:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:24:56.835 11:35:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@966 -- # echo 'killing process with pid 983787' 00:24:56.835 killing process with pid 983787 00:24:56.835 11:35:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@967 -- # kill 983787 00:24:56.835 Received shutdown signal, test time was about 60.000000 seconds 00:24:56.835 00:24:56.835 Latency(us) 00:24:56.835 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:56.835 =================================================================================================================== 00:24:56.835 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:24:56.835 [2024-07-15 11:35:40.365852] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:24:56.835 11:35:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@972 -- # wait 983787 00:24:56.835 [2024-07-15 11:35:40.411440] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:24:57.093 11:35:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # return 0 00:24:57.093 00:24:57.093 real 0m23.995s 00:24:57.093 user 0m32.195s 00:24:57.093 sys 0m5.272s 00:24:57.093 11:35:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:24:57.093 11:35:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:24:57.093 ************************************ 00:24:57.093 END TEST raid_rebuild_test 00:24:57.093 ************************************ 00:24:57.093 11:35:40 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:24:57.093 11:35:40 bdev_raid -- bdev/bdev_raid.sh@878 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 4 true false true 00:24:57.093 11:35:40 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:24:57.093 11:35:40 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:24:57.093 11:35:40 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:24:57.352 ************************************ 00:24:57.352 START TEST raid_rebuild_test_sb 00:24:57.352 ************************************ 00:24:57.352 11:35:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1123 -- # raid_rebuild_test raid1 4 true false true 00:24:57.352 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@568 -- # local raid_level=raid1 00:24:57.352 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local num_base_bdevs=4 00:24:57.352 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local superblock=true 00:24:57.352 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local background_io=false 00:24:57.352 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local verify=true 00:24:57.352 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # (( i = 1 )) 00:24:57.352 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:24:57.352 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # echo BaseBdev1 00:24:57.352 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:24:57.352 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:24:57.352 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # echo BaseBdev2 00:24:57.352 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:24:57.352 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # echo BaseBdev3 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # echo BaseBdev4 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local base_bdevs 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local raid_bdev_name=raid_bdev1 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local strip_size 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local create_arg 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local raid_bdev_size 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local data_offset 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@580 -- # '[' raid1 '!=' raid1 ']' 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@588 -- # strip_size=0 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@591 -- # '[' true = true ']' 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # create_arg+=' -s' 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # raid_pid=987019 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # waitforlisten 987019 /var/tmp/spdk-raid.sock 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@595 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@829 -- # '[' -z 987019 ']' 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@834 -- # local max_retries=100 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:24:57.353 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@838 -- # xtrace_disable 00:24:57.353 11:35:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:57.353 [2024-07-15 11:35:40.782396] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:24:57.353 [2024-07-15 11:35:40.782466] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid987019 ] 00:24:57.353 I/O size of 3145728 is greater than zero copy threshold (65536). 00:24:57.353 Zero copy mechanism will not be used. 00:24:57.353 [2024-07-15 11:35:40.914433] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:57.611 [2024-07-15 11:35:41.023268] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:24:57.611 [2024-07-15 11:35:41.088676] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:24:57.611 [2024-07-15 11:35:41.088723] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:24:58.175 11:35:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:24:58.175 11:35:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@862 -- # return 0 00:24:58.175 11:35:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:24:58.175 11:35:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:24:58.432 BaseBdev1_malloc 00:24:58.432 11:35:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:24:58.689 [2024-07-15 11:35:42.194261] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:24:58.689 [2024-07-15 11:35:42.194308] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:58.689 [2024-07-15 11:35:42.194329] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2463d40 00:24:58.689 [2024-07-15 11:35:42.194341] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:58.689 [2024-07-15 11:35:42.195904] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:58.689 [2024-07-15 11:35:42.195939] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:24:58.689 BaseBdev1 00:24:58.689 11:35:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:24:58.689 11:35:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:24:58.947 BaseBdev2_malloc 00:24:58.947 11:35:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:24:59.205 [2024-07-15 11:35:42.628204] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:24:59.205 [2024-07-15 11:35:42.628250] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:59.205 [2024-07-15 11:35:42.628272] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2464860 00:24:59.205 [2024-07-15 11:35:42.628285] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:59.205 [2024-07-15 11:35:42.629645] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:59.205 [2024-07-15 11:35:42.629672] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:24:59.205 BaseBdev2 00:24:59.205 11:35:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:24:59.205 11:35:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:24:59.463 BaseBdev3_malloc 00:24:59.463 11:35:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:24:59.721 [2024-07-15 11:35:43.126006] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:24:59.721 [2024-07-15 11:35:43.126053] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:59.721 [2024-07-15 11:35:43.126073] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x26118f0 00:24:59.721 [2024-07-15 11:35:43.126086] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:59.721 [2024-07-15 11:35:43.127453] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:59.721 [2024-07-15 11:35:43.127479] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:24:59.721 BaseBdev3 00:24:59.721 11:35:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:24:59.721 11:35:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:24:59.977 BaseBdev4_malloc 00:24:59.977 11:35:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:25:00.234 [2024-07-15 11:35:43.615792] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:25:00.234 [2024-07-15 11:35:43.615840] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:00.234 [2024-07-15 11:35:43.615859] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2610ad0 00:25:00.234 [2024-07-15 11:35:43.615871] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:00.234 [2024-07-15 11:35:43.617313] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:00.234 [2024-07-15 11:35:43.617339] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:25:00.234 BaseBdev4 00:25:00.234 11:35:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@606 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b spare_malloc 00:25:00.491 spare_malloc 00:25:00.491 11:35:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:25:00.491 spare_delay 00:25:00.491 11:35:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:25:00.748 [2024-07-15 11:35:44.294387] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:25:00.748 [2024-07-15 11:35:44.294434] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:00.748 [2024-07-15 11:35:44.294455] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x26155b0 00:25:00.748 [2024-07-15 11:35:44.294467] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:00.748 [2024-07-15 11:35:44.295974] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:00.748 [2024-07-15 11:35:44.296001] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:25:00.748 spare 00:25:00.748 11:35:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@611 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n raid_bdev1 00:25:01.005 [2024-07-15 11:35:44.543076] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:25:01.005 [2024-07-15 11:35:44.544320] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:25:01.005 [2024-07-15 11:35:44.544376] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:25:01.005 [2024-07-15 11:35:44.544422] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:25:01.005 [2024-07-15 11:35:44.544619] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x25948a0 00:25:01.005 [2024-07-15 11:35:44.544630] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:25:01.005 [2024-07-15 11:35:44.544826] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x260ee10 00:25:01.005 [2024-07-15 11:35:44.544985] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x25948a0 00:25:01.005 [2024-07-15 11:35:44.544995] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x25948a0 00:25:01.005 [2024-07-15 11:35:44.545090] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:01.005 11:35:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:25:01.005 11:35:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:25:01.005 11:35:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:25:01.005 11:35:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:25:01.005 11:35:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:25:01.005 11:35:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:25:01.005 11:35:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:25:01.005 11:35:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:25:01.005 11:35:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:25:01.005 11:35:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:25:01.005 11:35:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:01.005 11:35:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:01.263 11:35:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:25:01.263 "name": "raid_bdev1", 00:25:01.263 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:01.263 "strip_size_kb": 0, 00:25:01.263 "state": "online", 00:25:01.263 "raid_level": "raid1", 00:25:01.263 "superblock": true, 00:25:01.263 "num_base_bdevs": 4, 00:25:01.263 "num_base_bdevs_discovered": 4, 00:25:01.263 "num_base_bdevs_operational": 4, 00:25:01.263 "base_bdevs_list": [ 00:25:01.263 { 00:25:01.263 "name": "BaseBdev1", 00:25:01.263 "uuid": "b73a1aae-a7b4-5fd3-8f44-77613c6101a4", 00:25:01.263 "is_configured": true, 00:25:01.263 "data_offset": 2048, 00:25:01.263 "data_size": 63488 00:25:01.263 }, 00:25:01.263 { 00:25:01.263 "name": "BaseBdev2", 00:25:01.263 "uuid": "20e59bdc-a542-5da9-9369-0f26db129333", 00:25:01.263 "is_configured": true, 00:25:01.263 "data_offset": 2048, 00:25:01.263 "data_size": 63488 00:25:01.263 }, 00:25:01.263 { 00:25:01.263 "name": "BaseBdev3", 00:25:01.263 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:01.263 "is_configured": true, 00:25:01.263 "data_offset": 2048, 00:25:01.263 "data_size": 63488 00:25:01.263 }, 00:25:01.263 { 00:25:01.263 "name": "BaseBdev4", 00:25:01.263 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:01.263 "is_configured": true, 00:25:01.263 "data_offset": 2048, 00:25:01.263 "data_size": 63488 00:25:01.263 } 00:25:01.263 ] 00:25:01.263 }' 00:25:01.263 11:35:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:25:01.263 11:35:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:02.195 11:35:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@615 -- # jq -r '.[].num_blocks' 00:25:02.195 11:35:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@615 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:25:02.195 [2024-07-15 11:35:45.654293] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:25:02.195 11:35:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@615 -- # raid_bdev_size=63488 00:25:02.195 11:35:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@618 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:02.195 11:35:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@618 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:25:02.453 11:35:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@618 -- # data_offset=2048 00:25:02.453 11:35:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@620 -- # '[' false = true ']' 00:25:02.453 11:35:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@623 -- # '[' true = true ']' 00:25:02.453 11:35:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # local write_unit_size 00:25:02.453 11:35:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@627 -- # nbd_start_disks /var/tmp/spdk-raid.sock raid_bdev1 /dev/nbd0 00:25:02.453 11:35:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:25:02.453 11:35:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:25:02.453 11:35:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:25:02.453 11:35:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:25:02.453 11:35:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:25:02.453 11:35:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:25:02.453 11:35:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:25:02.453 11:35:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:25:02.453 11:35:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:25:02.712 [2024-07-15 11:35:46.163386] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x260ee10 00:25:02.712 /dev/nbd0 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@867 -- # local i 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # break 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:25:02.712 1+0 records in 00:25:02.712 1+0 records out 00:25:02.712 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00027581 s, 14.9 MB/s 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # size=4096 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # return 0 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # '[' raid1 = raid5f ']' 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@632 -- # write_unit_size=1 00:25:02.712 11:35:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@634 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:25:10.833 63488+0 records in 00:25:10.833 63488+0 records out 00:25:10.833 32505856 bytes (33 MB, 31 MiB) copied, 7.7666 s, 4.2 MB/s 00:25:10.833 11:35:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # nbd_stop_disks /var/tmp/spdk-raid.sock /dev/nbd0 00:25:10.833 11:35:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:25:10.833 11:35:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:25:10.833 11:35:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:25:10.833 11:35:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:25:10.833 11:35:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:25:10.833 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd0 00:25:10.833 [2024-07-15 11:35:54.272345] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:10.833 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:25:10.833 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:25:10.833 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:25:10.833 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:25:10.833 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:25:10.833 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:25:10.833 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:25:10.833 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:25:10.833 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@639 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev1 00:25:11.090 [2024-07-15 11:35:54.509030] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:25:11.090 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@642 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:25:11.090 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:25:11.090 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:25:11.090 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:25:11.090 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:25:11.090 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:25:11.090 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:25:11.090 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:25:11.090 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:25:11.090 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:25:11.090 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:11.090 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:11.347 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:25:11.347 "name": "raid_bdev1", 00:25:11.347 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:11.347 "strip_size_kb": 0, 00:25:11.347 "state": "online", 00:25:11.347 "raid_level": "raid1", 00:25:11.347 "superblock": true, 00:25:11.347 "num_base_bdevs": 4, 00:25:11.347 "num_base_bdevs_discovered": 3, 00:25:11.347 "num_base_bdevs_operational": 3, 00:25:11.347 "base_bdevs_list": [ 00:25:11.347 { 00:25:11.347 "name": null, 00:25:11.347 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:11.347 "is_configured": false, 00:25:11.347 "data_offset": 2048, 00:25:11.347 "data_size": 63488 00:25:11.347 }, 00:25:11.347 { 00:25:11.347 "name": "BaseBdev2", 00:25:11.347 "uuid": "20e59bdc-a542-5da9-9369-0f26db129333", 00:25:11.347 "is_configured": true, 00:25:11.347 "data_offset": 2048, 00:25:11.347 "data_size": 63488 00:25:11.347 }, 00:25:11.347 { 00:25:11.347 "name": "BaseBdev3", 00:25:11.347 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:11.347 "is_configured": true, 00:25:11.347 "data_offset": 2048, 00:25:11.347 "data_size": 63488 00:25:11.347 }, 00:25:11.347 { 00:25:11.347 "name": "BaseBdev4", 00:25:11.347 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:11.347 "is_configured": true, 00:25:11.347 "data_offset": 2048, 00:25:11.347 "data_size": 63488 00:25:11.347 } 00:25:11.347 ] 00:25:11.347 }' 00:25:11.347 11:35:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:25:11.347 11:35:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:11.911 11:35:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@645 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:25:12.168 [2024-07-15 11:35:55.603910] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:25:12.168 [2024-07-15 11:35:55.608003] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x260ee10 00:25:12.168 [2024-07-15 11:35:55.610368] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:25:12.168 11:35:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # sleep 1 00:25:13.100 11:35:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@649 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:25:13.100 11:35:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:13.100 11:35:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:25:13.100 11:35:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=spare 00:25:13.100 11:35:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:13.100 11:35:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:13.100 11:35:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:13.357 11:35:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:13.357 "name": "raid_bdev1", 00:25:13.357 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:13.357 "strip_size_kb": 0, 00:25:13.357 "state": "online", 00:25:13.357 "raid_level": "raid1", 00:25:13.357 "superblock": true, 00:25:13.357 "num_base_bdevs": 4, 00:25:13.357 "num_base_bdevs_discovered": 4, 00:25:13.357 "num_base_bdevs_operational": 4, 00:25:13.357 "process": { 00:25:13.357 "type": "rebuild", 00:25:13.357 "target": "spare", 00:25:13.357 "progress": { 00:25:13.357 "blocks": 24576, 00:25:13.357 "percent": 38 00:25:13.357 } 00:25:13.357 }, 00:25:13.357 "base_bdevs_list": [ 00:25:13.357 { 00:25:13.357 "name": "spare", 00:25:13.357 "uuid": "dbc1f6a2-6d51-53bf-ab53-88b1f2633cc4", 00:25:13.357 "is_configured": true, 00:25:13.357 "data_offset": 2048, 00:25:13.357 "data_size": 63488 00:25:13.357 }, 00:25:13.357 { 00:25:13.357 "name": "BaseBdev2", 00:25:13.357 "uuid": "20e59bdc-a542-5da9-9369-0f26db129333", 00:25:13.357 "is_configured": true, 00:25:13.357 "data_offset": 2048, 00:25:13.357 "data_size": 63488 00:25:13.357 }, 00:25:13.357 { 00:25:13.357 "name": "BaseBdev3", 00:25:13.357 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:13.357 "is_configured": true, 00:25:13.357 "data_offset": 2048, 00:25:13.357 "data_size": 63488 00:25:13.357 }, 00:25:13.357 { 00:25:13.357 "name": "BaseBdev4", 00:25:13.357 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:13.357 "is_configured": true, 00:25:13.357 "data_offset": 2048, 00:25:13.357 "data_size": 63488 00:25:13.357 } 00:25:13.357 ] 00:25:13.357 }' 00:25:13.357 11:35:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:13.357 11:35:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:25:13.357 11:35:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:13.614 11:35:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:25:13.614 11:35:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@652 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev spare 00:25:13.614 [2024-07-15 11:35:57.193440] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:25:13.872 [2024-07-15 11:35:57.222945] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:25:13.872 [2024-07-15 11:35:57.222991] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:13.872 [2024-07-15 11:35:57.223008] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:25:13.872 [2024-07-15 11:35:57.223017] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:25:13.873 11:35:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@655 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:25:13.873 11:35:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:25:13.873 11:35:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:25:13.873 11:35:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:25:13.873 11:35:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:25:13.873 11:35:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:25:13.873 11:35:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:25:13.873 11:35:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:25:13.873 11:35:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:25:13.873 11:35:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:25:13.873 11:35:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:13.873 11:35:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:14.130 11:35:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:25:14.130 "name": "raid_bdev1", 00:25:14.130 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:14.130 "strip_size_kb": 0, 00:25:14.131 "state": "online", 00:25:14.131 "raid_level": "raid1", 00:25:14.131 "superblock": true, 00:25:14.131 "num_base_bdevs": 4, 00:25:14.131 "num_base_bdevs_discovered": 3, 00:25:14.131 "num_base_bdevs_operational": 3, 00:25:14.131 "base_bdevs_list": [ 00:25:14.131 { 00:25:14.131 "name": null, 00:25:14.131 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:14.131 "is_configured": false, 00:25:14.131 "data_offset": 2048, 00:25:14.131 "data_size": 63488 00:25:14.131 }, 00:25:14.131 { 00:25:14.131 "name": "BaseBdev2", 00:25:14.131 "uuid": "20e59bdc-a542-5da9-9369-0f26db129333", 00:25:14.131 "is_configured": true, 00:25:14.131 "data_offset": 2048, 00:25:14.131 "data_size": 63488 00:25:14.131 }, 00:25:14.131 { 00:25:14.131 "name": "BaseBdev3", 00:25:14.131 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:14.131 "is_configured": true, 00:25:14.131 "data_offset": 2048, 00:25:14.131 "data_size": 63488 00:25:14.131 }, 00:25:14.131 { 00:25:14.131 "name": "BaseBdev4", 00:25:14.131 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:14.131 "is_configured": true, 00:25:14.131 "data_offset": 2048, 00:25:14.131 "data_size": 63488 00:25:14.131 } 00:25:14.131 ] 00:25:14.131 }' 00:25:14.131 11:35:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:25:14.131 11:35:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:14.697 11:35:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@658 -- # verify_raid_bdev_process raid_bdev1 none none 00:25:14.697 11:35:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:14.697 11:35:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:25:14.697 11:35:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=none 00:25:14.697 11:35:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:14.697 11:35:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:14.697 11:35:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:14.956 11:35:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:14.956 "name": "raid_bdev1", 00:25:14.956 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:14.956 "strip_size_kb": 0, 00:25:14.956 "state": "online", 00:25:14.956 "raid_level": "raid1", 00:25:14.956 "superblock": true, 00:25:14.956 "num_base_bdevs": 4, 00:25:14.956 "num_base_bdevs_discovered": 3, 00:25:14.956 "num_base_bdevs_operational": 3, 00:25:14.956 "base_bdevs_list": [ 00:25:14.956 { 00:25:14.956 "name": null, 00:25:14.956 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:14.956 "is_configured": false, 00:25:14.956 "data_offset": 2048, 00:25:14.956 "data_size": 63488 00:25:14.956 }, 00:25:14.956 { 00:25:14.956 "name": "BaseBdev2", 00:25:14.956 "uuid": "20e59bdc-a542-5da9-9369-0f26db129333", 00:25:14.956 "is_configured": true, 00:25:14.956 "data_offset": 2048, 00:25:14.956 "data_size": 63488 00:25:14.956 }, 00:25:14.956 { 00:25:14.956 "name": "BaseBdev3", 00:25:14.956 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:14.956 "is_configured": true, 00:25:14.956 "data_offset": 2048, 00:25:14.956 "data_size": 63488 00:25:14.956 }, 00:25:14.956 { 00:25:14.956 "name": "BaseBdev4", 00:25:14.956 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:14.956 "is_configured": true, 00:25:14.956 "data_offset": 2048, 00:25:14.956 "data_size": 63488 00:25:14.956 } 00:25:14.956 ] 00:25:14.956 }' 00:25:14.956 11:35:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:14.956 11:35:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:25:14.956 11:35:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:14.956 11:35:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:25:14.956 11:35:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@661 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:25:15.214 [2024-07-15 11:35:58.675329] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:25:15.214 [2024-07-15 11:35:58.679430] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x2594e90 00:25:15.214 [2024-07-15 11:35:58.680936] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:25:15.214 11:35:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # sleep 1 00:25:16.149 11:35:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:25:16.149 11:35:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:16.149 11:35:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:25:16.149 11:35:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=spare 00:25:16.149 11:35:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:16.149 11:35:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:16.149 11:35:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:16.408 11:35:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:16.408 "name": "raid_bdev1", 00:25:16.408 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:16.408 "strip_size_kb": 0, 00:25:16.408 "state": "online", 00:25:16.408 "raid_level": "raid1", 00:25:16.408 "superblock": true, 00:25:16.408 "num_base_bdevs": 4, 00:25:16.408 "num_base_bdevs_discovered": 4, 00:25:16.408 "num_base_bdevs_operational": 4, 00:25:16.408 "process": { 00:25:16.408 "type": "rebuild", 00:25:16.408 "target": "spare", 00:25:16.408 "progress": { 00:25:16.408 "blocks": 24576, 00:25:16.408 "percent": 38 00:25:16.408 } 00:25:16.408 }, 00:25:16.408 "base_bdevs_list": [ 00:25:16.408 { 00:25:16.408 "name": "spare", 00:25:16.408 "uuid": "dbc1f6a2-6d51-53bf-ab53-88b1f2633cc4", 00:25:16.408 "is_configured": true, 00:25:16.408 "data_offset": 2048, 00:25:16.408 "data_size": 63488 00:25:16.408 }, 00:25:16.408 { 00:25:16.408 "name": "BaseBdev2", 00:25:16.408 "uuid": "20e59bdc-a542-5da9-9369-0f26db129333", 00:25:16.408 "is_configured": true, 00:25:16.408 "data_offset": 2048, 00:25:16.408 "data_size": 63488 00:25:16.408 }, 00:25:16.408 { 00:25:16.408 "name": "BaseBdev3", 00:25:16.408 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:16.408 "is_configured": true, 00:25:16.408 "data_offset": 2048, 00:25:16.408 "data_size": 63488 00:25:16.408 }, 00:25:16.408 { 00:25:16.408 "name": "BaseBdev4", 00:25:16.408 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:16.408 "is_configured": true, 00:25:16.408 "data_offset": 2048, 00:25:16.408 "data_size": 63488 00:25:16.408 } 00:25:16.408 ] 00:25:16.408 }' 00:25:16.408 11:35:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:16.408 11:35:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:25:16.667 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:16.667 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:25:16.667 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@665 -- # '[' true = true ']' 00:25:16.667 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@665 -- # '[' = false ']' 00:25:16.667 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev_raid.sh: line 665: [: =: unary operator expected 00:25:16.667 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@690 -- # local num_base_bdevs_operational=4 00:25:16.667 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@692 -- # '[' raid1 = raid1 ']' 00:25:16.667 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@692 -- # '[' 4 -gt 2 ']' 00:25:16.667 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@694 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev2 00:25:16.926 [2024-07-15 11:36:00.276612] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:25:16.926 [2024-07-15 11:36:00.393845] bdev_raid.c:1919:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x2594e90 00:25:16.926 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@697 -- # base_bdevs[1]= 00:25:16.926 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@698 -- # (( num_base_bdevs_operational-- )) 00:25:16.927 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@701 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:25:16.927 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:16.927 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:25:16.927 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=spare 00:25:16.927 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:16.927 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:16.927 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:17.186 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:17.186 "name": "raid_bdev1", 00:25:17.186 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:17.186 "strip_size_kb": 0, 00:25:17.186 "state": "online", 00:25:17.186 "raid_level": "raid1", 00:25:17.186 "superblock": true, 00:25:17.186 "num_base_bdevs": 4, 00:25:17.186 "num_base_bdevs_discovered": 3, 00:25:17.186 "num_base_bdevs_operational": 3, 00:25:17.186 "process": { 00:25:17.186 "type": "rebuild", 00:25:17.186 "target": "spare", 00:25:17.186 "progress": { 00:25:17.186 "blocks": 36864, 00:25:17.186 "percent": 58 00:25:17.186 } 00:25:17.186 }, 00:25:17.186 "base_bdevs_list": [ 00:25:17.186 { 00:25:17.186 "name": "spare", 00:25:17.186 "uuid": "dbc1f6a2-6d51-53bf-ab53-88b1f2633cc4", 00:25:17.186 "is_configured": true, 00:25:17.186 "data_offset": 2048, 00:25:17.186 "data_size": 63488 00:25:17.186 }, 00:25:17.186 { 00:25:17.186 "name": null, 00:25:17.186 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:17.186 "is_configured": false, 00:25:17.186 "data_offset": 2048, 00:25:17.186 "data_size": 63488 00:25:17.186 }, 00:25:17.186 { 00:25:17.186 "name": "BaseBdev3", 00:25:17.186 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:17.186 "is_configured": true, 00:25:17.186 "data_offset": 2048, 00:25:17.186 "data_size": 63488 00:25:17.186 }, 00:25:17.186 { 00:25:17.186 "name": "BaseBdev4", 00:25:17.186 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:17.186 "is_configured": true, 00:25:17.186 "data_offset": 2048, 00:25:17.186 "data_size": 63488 00:25:17.186 } 00:25:17.186 ] 00:25:17.186 }' 00:25:17.186 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:17.186 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:25:17.186 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:17.186 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:25:17.186 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@705 -- # local timeout=888 00:25:17.186 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:25:17.186 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:25:17.186 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:17.186 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:25:17.186 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=spare 00:25:17.186 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:17.186 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:17.186 11:36:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:17.445 11:36:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:17.445 "name": "raid_bdev1", 00:25:17.445 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:17.445 "strip_size_kb": 0, 00:25:17.445 "state": "online", 00:25:17.445 "raid_level": "raid1", 00:25:17.445 "superblock": true, 00:25:17.445 "num_base_bdevs": 4, 00:25:17.445 "num_base_bdevs_discovered": 3, 00:25:17.445 "num_base_bdevs_operational": 3, 00:25:17.445 "process": { 00:25:17.445 "type": "rebuild", 00:25:17.445 "target": "spare", 00:25:17.445 "progress": { 00:25:17.445 "blocks": 43008, 00:25:17.445 "percent": 67 00:25:17.445 } 00:25:17.445 }, 00:25:17.445 "base_bdevs_list": [ 00:25:17.445 { 00:25:17.445 "name": "spare", 00:25:17.445 "uuid": "dbc1f6a2-6d51-53bf-ab53-88b1f2633cc4", 00:25:17.445 "is_configured": true, 00:25:17.445 "data_offset": 2048, 00:25:17.445 "data_size": 63488 00:25:17.445 }, 00:25:17.445 { 00:25:17.445 "name": null, 00:25:17.445 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:17.445 "is_configured": false, 00:25:17.445 "data_offset": 2048, 00:25:17.445 "data_size": 63488 00:25:17.445 }, 00:25:17.445 { 00:25:17.445 "name": "BaseBdev3", 00:25:17.445 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:17.445 "is_configured": true, 00:25:17.445 "data_offset": 2048, 00:25:17.445 "data_size": 63488 00:25:17.445 }, 00:25:17.445 { 00:25:17.445 "name": "BaseBdev4", 00:25:17.445 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:17.445 "is_configured": true, 00:25:17.445 "data_offset": 2048, 00:25:17.445 "data_size": 63488 00:25:17.445 } 00:25:17.445 ] 00:25:17.445 }' 00:25:17.445 11:36:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:17.704 11:36:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:25:17.704 11:36:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:17.704 11:36:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:25:17.704 11:36:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@710 -- # sleep 1 00:25:18.639 [2024-07-15 11:36:01.905655] bdev_raid.c:2789:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:25:18.639 [2024-07-15 11:36:01.905720] bdev_raid.c:2504:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:25:18.639 [2024-07-15 11:36:01.905830] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:18.639 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:25:18.639 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:25:18.639 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:18.639 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:25:18.639 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=spare 00:25:18.639 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:18.639 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:18.639 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:18.897 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:18.897 "name": "raid_bdev1", 00:25:18.897 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:18.897 "strip_size_kb": 0, 00:25:18.897 "state": "online", 00:25:18.897 "raid_level": "raid1", 00:25:18.897 "superblock": true, 00:25:18.897 "num_base_bdevs": 4, 00:25:18.897 "num_base_bdevs_discovered": 3, 00:25:18.897 "num_base_bdevs_operational": 3, 00:25:18.898 "base_bdevs_list": [ 00:25:18.898 { 00:25:18.898 "name": "spare", 00:25:18.898 "uuid": "dbc1f6a2-6d51-53bf-ab53-88b1f2633cc4", 00:25:18.898 "is_configured": true, 00:25:18.898 "data_offset": 2048, 00:25:18.898 "data_size": 63488 00:25:18.898 }, 00:25:18.898 { 00:25:18.898 "name": null, 00:25:18.898 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:18.898 "is_configured": false, 00:25:18.898 "data_offset": 2048, 00:25:18.898 "data_size": 63488 00:25:18.898 }, 00:25:18.898 { 00:25:18.898 "name": "BaseBdev3", 00:25:18.898 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:18.898 "is_configured": true, 00:25:18.898 "data_offset": 2048, 00:25:18.898 "data_size": 63488 00:25:18.898 }, 00:25:18.898 { 00:25:18.898 "name": "BaseBdev4", 00:25:18.898 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:18.898 "is_configured": true, 00:25:18.898 "data_offset": 2048, 00:25:18.898 "data_size": 63488 00:25:18.898 } 00:25:18.898 ] 00:25:18.898 }' 00:25:18.898 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:18.898 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ none == \r\e\b\u\i\l\d ]] 00:25:18.898 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:18.898 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ none == \s\p\a\r\e ]] 00:25:18.898 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # break 00:25:18.898 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@714 -- # verify_raid_bdev_process raid_bdev1 none none 00:25:18.898 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:18.898 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:25:18.898 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=none 00:25:18.898 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:18.898 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:18.898 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:19.156 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:19.156 "name": "raid_bdev1", 00:25:19.156 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:19.156 "strip_size_kb": 0, 00:25:19.156 "state": "online", 00:25:19.156 "raid_level": "raid1", 00:25:19.156 "superblock": true, 00:25:19.156 "num_base_bdevs": 4, 00:25:19.156 "num_base_bdevs_discovered": 3, 00:25:19.156 "num_base_bdevs_operational": 3, 00:25:19.156 "base_bdevs_list": [ 00:25:19.156 { 00:25:19.156 "name": "spare", 00:25:19.156 "uuid": "dbc1f6a2-6d51-53bf-ab53-88b1f2633cc4", 00:25:19.156 "is_configured": true, 00:25:19.156 "data_offset": 2048, 00:25:19.156 "data_size": 63488 00:25:19.156 }, 00:25:19.156 { 00:25:19.156 "name": null, 00:25:19.156 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:19.156 "is_configured": false, 00:25:19.156 "data_offset": 2048, 00:25:19.156 "data_size": 63488 00:25:19.156 }, 00:25:19.156 { 00:25:19.156 "name": "BaseBdev3", 00:25:19.156 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:19.156 "is_configured": true, 00:25:19.156 "data_offset": 2048, 00:25:19.156 "data_size": 63488 00:25:19.156 }, 00:25:19.156 { 00:25:19.156 "name": "BaseBdev4", 00:25:19.156 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:19.156 "is_configured": true, 00:25:19.156 "data_offset": 2048, 00:25:19.156 "data_size": 63488 00:25:19.156 } 00:25:19.156 ] 00:25:19.156 }' 00:25:19.156 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:19.156 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:25:19.156 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:19.414 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:25:19.414 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:25:19.414 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:25:19.414 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:25:19.414 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:25:19.414 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:25:19.414 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:25:19.414 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:25:19.414 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:25:19.414 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:25:19.414 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:25:19.414 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:19.414 11:36:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:19.672 11:36:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:25:19.672 "name": "raid_bdev1", 00:25:19.672 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:19.672 "strip_size_kb": 0, 00:25:19.672 "state": "online", 00:25:19.672 "raid_level": "raid1", 00:25:19.672 "superblock": true, 00:25:19.672 "num_base_bdevs": 4, 00:25:19.672 "num_base_bdevs_discovered": 3, 00:25:19.672 "num_base_bdevs_operational": 3, 00:25:19.672 "base_bdevs_list": [ 00:25:19.672 { 00:25:19.672 "name": "spare", 00:25:19.672 "uuid": "dbc1f6a2-6d51-53bf-ab53-88b1f2633cc4", 00:25:19.672 "is_configured": true, 00:25:19.672 "data_offset": 2048, 00:25:19.672 "data_size": 63488 00:25:19.672 }, 00:25:19.672 { 00:25:19.672 "name": null, 00:25:19.672 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:19.672 "is_configured": false, 00:25:19.672 "data_offset": 2048, 00:25:19.672 "data_size": 63488 00:25:19.672 }, 00:25:19.672 { 00:25:19.672 "name": "BaseBdev3", 00:25:19.672 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:19.672 "is_configured": true, 00:25:19.672 "data_offset": 2048, 00:25:19.672 "data_size": 63488 00:25:19.672 }, 00:25:19.672 { 00:25:19.672 "name": "BaseBdev4", 00:25:19.672 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:19.672 "is_configured": true, 00:25:19.672 "data_offset": 2048, 00:25:19.672 "data_size": 63488 00:25:19.672 } 00:25:19.672 ] 00:25:19.672 }' 00:25:19.672 11:36:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:25:19.672 11:36:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:20.604 11:36:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@718 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:25:20.604 [2024-07-15 11:36:04.055668] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:25:20.604 [2024-07-15 11:36:04.055698] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:25:20.604 [2024-07-15 11:36:04.055759] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:25:20.604 [2024-07-15 11:36:04.055832] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:25:20.604 [2024-07-15 11:36:04.055844] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x25948a0 name raid_bdev1, state offline 00:25:20.604 11:36:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:20.604 11:36:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # jq length 00:25:20.862 11:36:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # [[ 0 == 0 ]] 00:25:20.862 11:36:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@721 -- # '[' true = true ']' 00:25:20.862 11:36:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' false = true ']' 00:25:20.862 11:36:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@736 -- # nbd_start_disks /var/tmp/spdk-raid.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:25:20.862 11:36:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:25:20.862 11:36:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:25:20.862 11:36:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:25:20.862 11:36:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:25:20.862 11:36:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:25:20.862 11:36:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:25:20.862 11:36:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:25:20.862 11:36:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:25:20.862 11:36:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:25:21.428 /dev/nbd0 00:25:21.428 11:36:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:25:21.428 11:36:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:25:21.428 11:36:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:25:21.428 11:36:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@867 -- # local i 00:25:21.428 11:36:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:25:21.428 11:36:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:25:21.428 11:36:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:25:21.428 11:36:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # break 00:25:21.428 11:36:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:25:21.428 11:36:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:25:21.428 11:36:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:25:21.428 1+0 records in 00:25:21.428 1+0 records out 00:25:21.428 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000241172 s, 17.0 MB/s 00:25:21.428 11:36:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:25:21.428 11:36:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # size=4096 00:25:21.428 11:36:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:25:21.428 11:36:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:25:21.428 11:36:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # return 0 00:25:21.428 11:36:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:25:21.428 11:36:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:25:21.428 11:36:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk spare /dev/nbd1 00:25:21.687 /dev/nbd1 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@867 -- # local i 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # break 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:25:21.687 1+0 records in 00:25:21.687 1+0 records out 00:25:21.687 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000330631 s, 12.4 MB/s 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # size=4096 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # return 0 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # nbd_stop_disks /var/tmp/spdk-raid.sock '/dev/nbd0 /dev/nbd1' 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:25:21.687 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd0 00:25:21.945 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:25:21.945 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:25:21.945 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:25:21.945 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:25:21.945 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:25:21.945 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:25:21.945 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:25:21.945 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:25:21.945 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:25:21.945 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd1 00:25:22.203 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:25:22.203 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:25:22.203 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:25:22.203 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:25:22.203 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:25:22.203 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:25:22.203 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:25:22.203 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:25:22.203 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@742 -- # '[' true = true ']' 00:25:22.203 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@744 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:25:22.460 11:36:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:25:22.717 [2024-07-15 11:36:06.135152] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:25:22.717 [2024-07-15 11:36:06.135203] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:22.717 [2024-07-15 11:36:06.135225] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x260eb40 00:25:22.717 [2024-07-15 11:36:06.135238] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:22.717 [2024-07-15 11:36:06.136886] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:22.717 [2024-07-15 11:36:06.136917] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:25:22.718 [2024-07-15 11:36:06.137011] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:25:22.718 [2024-07-15 11:36:06.137040] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:25:22.718 [2024-07-15 11:36:06.137150] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:25:22.718 [2024-07-15 11:36:06.137224] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:25:22.718 spare 00:25:22.718 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:25:22.718 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:25:22.718 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:25:22.718 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:25:22.718 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:25:22.718 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:25:22.718 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:25:22.718 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:25:22.718 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:25:22.718 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:25:22.718 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:22.718 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:22.718 [2024-07-15 11:36:06.237539] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x2598ec0 00:25:22.718 [2024-07-15 11:36:06.237558] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:25:22.718 [2024-07-15 11:36:06.237756] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x24632f0 00:25:22.718 [2024-07-15 11:36:06.237907] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x2598ec0 00:25:22.718 [2024-07-15 11:36:06.237922] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x2598ec0 00:25:22.718 [2024-07-15 11:36:06.238032] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:22.975 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:25:22.975 "name": "raid_bdev1", 00:25:22.975 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:22.975 "strip_size_kb": 0, 00:25:22.975 "state": "online", 00:25:22.975 "raid_level": "raid1", 00:25:22.975 "superblock": true, 00:25:22.975 "num_base_bdevs": 4, 00:25:22.975 "num_base_bdevs_discovered": 3, 00:25:22.975 "num_base_bdevs_operational": 3, 00:25:22.975 "base_bdevs_list": [ 00:25:22.975 { 00:25:22.975 "name": "spare", 00:25:22.975 "uuid": "dbc1f6a2-6d51-53bf-ab53-88b1f2633cc4", 00:25:22.975 "is_configured": true, 00:25:22.975 "data_offset": 2048, 00:25:22.975 "data_size": 63488 00:25:22.975 }, 00:25:22.975 { 00:25:22.975 "name": null, 00:25:22.975 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:22.975 "is_configured": false, 00:25:22.975 "data_offset": 2048, 00:25:22.975 "data_size": 63488 00:25:22.975 }, 00:25:22.975 { 00:25:22.975 "name": "BaseBdev3", 00:25:22.975 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:22.975 "is_configured": true, 00:25:22.975 "data_offset": 2048, 00:25:22.975 "data_size": 63488 00:25:22.975 }, 00:25:22.975 { 00:25:22.975 "name": "BaseBdev4", 00:25:22.975 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:22.975 "is_configured": true, 00:25:22.975 "data_offset": 2048, 00:25:22.975 "data_size": 63488 00:25:22.975 } 00:25:22.975 ] 00:25:22.975 }' 00:25:22.975 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:25:22.975 11:36:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:23.540 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@748 -- # verify_raid_bdev_process raid_bdev1 none none 00:25:23.540 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:23.540 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:25:23.540 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=none 00:25:23.540 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:23.540 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:23.540 11:36:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:23.798 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:23.798 "name": "raid_bdev1", 00:25:23.798 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:23.798 "strip_size_kb": 0, 00:25:23.798 "state": "online", 00:25:23.798 "raid_level": "raid1", 00:25:23.798 "superblock": true, 00:25:23.798 "num_base_bdevs": 4, 00:25:23.798 "num_base_bdevs_discovered": 3, 00:25:23.798 "num_base_bdevs_operational": 3, 00:25:23.798 "base_bdevs_list": [ 00:25:23.798 { 00:25:23.798 "name": "spare", 00:25:23.798 "uuid": "dbc1f6a2-6d51-53bf-ab53-88b1f2633cc4", 00:25:23.798 "is_configured": true, 00:25:23.798 "data_offset": 2048, 00:25:23.798 "data_size": 63488 00:25:23.798 }, 00:25:23.798 { 00:25:23.798 "name": null, 00:25:23.798 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:23.798 "is_configured": false, 00:25:23.798 "data_offset": 2048, 00:25:23.798 "data_size": 63488 00:25:23.798 }, 00:25:23.798 { 00:25:23.798 "name": "BaseBdev3", 00:25:23.798 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:23.798 "is_configured": true, 00:25:23.798 "data_offset": 2048, 00:25:23.798 "data_size": 63488 00:25:23.798 }, 00:25:23.798 { 00:25:23.798 "name": "BaseBdev4", 00:25:23.798 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:23.798 "is_configured": true, 00:25:23.798 "data_offset": 2048, 00:25:23.798 "data_size": 63488 00:25:23.798 } 00:25:23.798 ] 00:25:23.798 }' 00:25:23.798 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:23.798 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:25:23.798 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:23.798 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:25:23.798 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:23.798 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # jq -r '.[].base_bdevs_list[0].name' 00:25:24.056 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # [[ spare == \s\p\a\r\e ]] 00:25:24.056 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@752 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev spare 00:25:24.314 [2024-07-15 11:36:07.799715] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:25:24.314 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@753 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:25:24.314 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:25:24.314 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:25:24.314 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:25:24.314 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:25:24.314 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:25:24.314 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:25:24.314 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:25:24.314 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:25:24.314 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:25:24.314 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:24.314 11:36:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:24.572 11:36:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:25:24.572 "name": "raid_bdev1", 00:25:24.572 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:24.572 "strip_size_kb": 0, 00:25:24.572 "state": "online", 00:25:24.572 "raid_level": "raid1", 00:25:24.572 "superblock": true, 00:25:24.572 "num_base_bdevs": 4, 00:25:24.572 "num_base_bdevs_discovered": 2, 00:25:24.572 "num_base_bdevs_operational": 2, 00:25:24.572 "base_bdevs_list": [ 00:25:24.572 { 00:25:24.572 "name": null, 00:25:24.572 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:24.572 "is_configured": false, 00:25:24.572 "data_offset": 2048, 00:25:24.572 "data_size": 63488 00:25:24.572 }, 00:25:24.572 { 00:25:24.572 "name": null, 00:25:24.572 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:24.572 "is_configured": false, 00:25:24.572 "data_offset": 2048, 00:25:24.572 "data_size": 63488 00:25:24.572 }, 00:25:24.572 { 00:25:24.572 "name": "BaseBdev3", 00:25:24.572 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:24.572 "is_configured": true, 00:25:24.572 "data_offset": 2048, 00:25:24.572 "data_size": 63488 00:25:24.572 }, 00:25:24.572 { 00:25:24.572 "name": "BaseBdev4", 00:25:24.572 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:24.572 "is_configured": true, 00:25:24.572 "data_offset": 2048, 00:25:24.572 "data_size": 63488 00:25:24.572 } 00:25:24.572 ] 00:25:24.572 }' 00:25:24.572 11:36:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:25:24.572 11:36:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:25.137 11:36:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:25:25.394 [2024-07-15 11:36:08.866567] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:25:25.394 [2024-07-15 11:36:08.866735] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:25:25.394 [2024-07-15 11:36:08.866753] bdev_raid.c:3620:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:25:25.394 [2024-07-15 11:36:08.866783] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:25:25.394 [2024-07-15 11:36:08.871322] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x2614fb0 00:25:25.394 [2024-07-15 11:36:08.873776] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:25:25.394 11:36:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # sleep 1 00:25:26.326 11:36:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:25:26.326 11:36:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:26.326 11:36:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:25:26.326 11:36:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=spare 00:25:26.326 11:36:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:26.326 11:36:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:26.327 11:36:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:26.584 11:36:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:26.584 "name": "raid_bdev1", 00:25:26.584 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:26.584 "strip_size_kb": 0, 00:25:26.584 "state": "online", 00:25:26.584 "raid_level": "raid1", 00:25:26.584 "superblock": true, 00:25:26.584 "num_base_bdevs": 4, 00:25:26.584 "num_base_bdevs_discovered": 3, 00:25:26.584 "num_base_bdevs_operational": 3, 00:25:26.584 "process": { 00:25:26.584 "type": "rebuild", 00:25:26.584 "target": "spare", 00:25:26.584 "progress": { 00:25:26.584 "blocks": 24576, 00:25:26.584 "percent": 38 00:25:26.584 } 00:25:26.584 }, 00:25:26.584 "base_bdevs_list": [ 00:25:26.584 { 00:25:26.584 "name": "spare", 00:25:26.584 "uuid": "dbc1f6a2-6d51-53bf-ab53-88b1f2633cc4", 00:25:26.584 "is_configured": true, 00:25:26.584 "data_offset": 2048, 00:25:26.584 "data_size": 63488 00:25:26.584 }, 00:25:26.584 { 00:25:26.584 "name": null, 00:25:26.584 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:26.584 "is_configured": false, 00:25:26.584 "data_offset": 2048, 00:25:26.584 "data_size": 63488 00:25:26.584 }, 00:25:26.584 { 00:25:26.584 "name": "BaseBdev3", 00:25:26.584 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:26.584 "is_configured": true, 00:25:26.584 "data_offset": 2048, 00:25:26.584 "data_size": 63488 00:25:26.584 }, 00:25:26.584 { 00:25:26.584 "name": "BaseBdev4", 00:25:26.584 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:26.584 "is_configured": true, 00:25:26.584 "data_offset": 2048, 00:25:26.584 "data_size": 63488 00:25:26.584 } 00:25:26.584 ] 00:25:26.584 }' 00:25:26.584 11:36:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:26.842 11:36:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:25:26.842 11:36:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:26.842 11:36:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:25:26.842 11:36:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@759 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:25:27.131 [2024-07-15 11:36:10.488004] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:25:27.131 [2024-07-15 11:36:10.587020] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:25:27.131 [2024-07-15 11:36:10.587069] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:27.131 [2024-07-15 11:36:10.587086] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:25:27.131 [2024-07-15 11:36:10.587095] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:25:27.131 11:36:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@760 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:25:27.131 11:36:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:25:27.131 11:36:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:25:27.131 11:36:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:25:27.131 11:36:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:25:27.131 11:36:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:25:27.131 11:36:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:25:27.131 11:36:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:25:27.132 11:36:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:25:27.132 11:36:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:25:27.132 11:36:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:27.132 11:36:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:27.388 11:36:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:25:27.388 "name": "raid_bdev1", 00:25:27.388 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:27.388 "strip_size_kb": 0, 00:25:27.388 "state": "online", 00:25:27.388 "raid_level": "raid1", 00:25:27.388 "superblock": true, 00:25:27.388 "num_base_bdevs": 4, 00:25:27.388 "num_base_bdevs_discovered": 2, 00:25:27.388 "num_base_bdevs_operational": 2, 00:25:27.388 "base_bdevs_list": [ 00:25:27.388 { 00:25:27.388 "name": null, 00:25:27.388 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:27.388 "is_configured": false, 00:25:27.388 "data_offset": 2048, 00:25:27.388 "data_size": 63488 00:25:27.388 }, 00:25:27.388 { 00:25:27.388 "name": null, 00:25:27.388 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:27.388 "is_configured": false, 00:25:27.388 "data_offset": 2048, 00:25:27.388 "data_size": 63488 00:25:27.388 }, 00:25:27.388 { 00:25:27.388 "name": "BaseBdev3", 00:25:27.388 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:27.388 "is_configured": true, 00:25:27.388 "data_offset": 2048, 00:25:27.388 "data_size": 63488 00:25:27.388 }, 00:25:27.388 { 00:25:27.388 "name": "BaseBdev4", 00:25:27.388 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:27.388 "is_configured": true, 00:25:27.388 "data_offset": 2048, 00:25:27.388 "data_size": 63488 00:25:27.388 } 00:25:27.388 ] 00:25:27.388 }' 00:25:27.388 11:36:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:25:27.389 11:36:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:27.952 11:36:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:25:28.209 [2024-07-15 11:36:11.682487] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:25:28.209 [2024-07-15 11:36:11.682541] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:28.209 [2024-07-15 11:36:11.682562] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2599930 00:25:28.209 [2024-07-15 11:36:11.682575] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:28.209 [2024-07-15 11:36:11.682963] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:28.209 [2024-07-15 11:36:11.682983] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:25:28.209 [2024-07-15 11:36:11.683069] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:25:28.209 [2024-07-15 11:36:11.683083] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:25:28.209 [2024-07-15 11:36:11.683095] bdev_raid.c:3620:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:25:28.209 [2024-07-15 11:36:11.683115] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:25:28.209 [2024-07-15 11:36:11.687093] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x2614f70 00:25:28.209 spare 00:25:28.209 [2024-07-15 11:36:11.688488] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:25:28.209 11:36:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # sleep 1 00:25:29.139 11:36:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:25:29.139 11:36:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:29.139 11:36:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:25:29.139 11:36:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=spare 00:25:29.139 11:36:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:29.139 11:36:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:29.139 11:36:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:29.395 11:36:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:29.395 "name": "raid_bdev1", 00:25:29.395 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:29.395 "strip_size_kb": 0, 00:25:29.395 "state": "online", 00:25:29.395 "raid_level": "raid1", 00:25:29.395 "superblock": true, 00:25:29.395 "num_base_bdevs": 4, 00:25:29.395 "num_base_bdevs_discovered": 3, 00:25:29.395 "num_base_bdevs_operational": 3, 00:25:29.395 "process": { 00:25:29.395 "type": "rebuild", 00:25:29.395 "target": "spare", 00:25:29.395 "progress": { 00:25:29.395 "blocks": 24576, 00:25:29.395 "percent": 38 00:25:29.395 } 00:25:29.395 }, 00:25:29.395 "base_bdevs_list": [ 00:25:29.395 { 00:25:29.395 "name": "spare", 00:25:29.395 "uuid": "dbc1f6a2-6d51-53bf-ab53-88b1f2633cc4", 00:25:29.395 "is_configured": true, 00:25:29.395 "data_offset": 2048, 00:25:29.395 "data_size": 63488 00:25:29.395 }, 00:25:29.395 { 00:25:29.395 "name": null, 00:25:29.395 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:29.395 "is_configured": false, 00:25:29.395 "data_offset": 2048, 00:25:29.395 "data_size": 63488 00:25:29.395 }, 00:25:29.395 { 00:25:29.395 "name": "BaseBdev3", 00:25:29.395 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:29.395 "is_configured": true, 00:25:29.395 "data_offset": 2048, 00:25:29.395 "data_size": 63488 00:25:29.395 }, 00:25:29.395 { 00:25:29.395 "name": "BaseBdev4", 00:25:29.395 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:29.395 "is_configured": true, 00:25:29.395 "data_offset": 2048, 00:25:29.395 "data_size": 63488 00:25:29.395 } 00:25:29.396 ] 00:25:29.396 }' 00:25:29.396 11:36:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:29.672 11:36:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:25:29.672 11:36:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:29.672 11:36:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:25:29.672 11:36:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@766 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:25:29.672 [2024-07-15 11:36:13.264567] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:25:29.930 [2024-07-15 11:36:13.301121] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:25:29.930 [2024-07-15 11:36:13.301165] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:29.930 [2024-07-15 11:36:13.301182] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:25:29.930 [2024-07-15 11:36:13.301190] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:25:29.930 11:36:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@767 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:25:29.930 11:36:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:25:29.930 11:36:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:25:29.930 11:36:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:25:29.930 11:36:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:25:29.930 11:36:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:25:29.930 11:36:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:25:29.930 11:36:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:25:29.930 11:36:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:25:29.930 11:36:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:25:29.930 11:36:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:29.930 11:36:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:30.188 11:36:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:25:30.188 "name": "raid_bdev1", 00:25:30.188 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:30.188 "strip_size_kb": 0, 00:25:30.188 "state": "online", 00:25:30.188 "raid_level": "raid1", 00:25:30.188 "superblock": true, 00:25:30.188 "num_base_bdevs": 4, 00:25:30.188 "num_base_bdevs_discovered": 2, 00:25:30.188 "num_base_bdevs_operational": 2, 00:25:30.188 "base_bdevs_list": [ 00:25:30.188 { 00:25:30.188 "name": null, 00:25:30.188 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:30.188 "is_configured": false, 00:25:30.188 "data_offset": 2048, 00:25:30.188 "data_size": 63488 00:25:30.188 }, 00:25:30.188 { 00:25:30.188 "name": null, 00:25:30.188 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:30.188 "is_configured": false, 00:25:30.188 "data_offset": 2048, 00:25:30.188 "data_size": 63488 00:25:30.188 }, 00:25:30.188 { 00:25:30.188 "name": "BaseBdev3", 00:25:30.188 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:30.188 "is_configured": true, 00:25:30.188 "data_offset": 2048, 00:25:30.188 "data_size": 63488 00:25:30.188 }, 00:25:30.188 { 00:25:30.188 "name": "BaseBdev4", 00:25:30.188 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:30.188 "is_configured": true, 00:25:30.188 "data_offset": 2048, 00:25:30.188 "data_size": 63488 00:25:30.188 } 00:25:30.188 ] 00:25:30.188 }' 00:25:30.188 11:36:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:25:30.188 11:36:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:30.754 11:36:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # verify_raid_bdev_process raid_bdev1 none none 00:25:30.754 11:36:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:30.754 11:36:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:25:30.754 11:36:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=none 00:25:30.754 11:36:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:30.754 11:36:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:30.754 11:36:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:31.012 11:36:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:31.012 "name": "raid_bdev1", 00:25:31.012 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:31.012 "strip_size_kb": 0, 00:25:31.012 "state": "online", 00:25:31.012 "raid_level": "raid1", 00:25:31.012 "superblock": true, 00:25:31.012 "num_base_bdevs": 4, 00:25:31.012 "num_base_bdevs_discovered": 2, 00:25:31.012 "num_base_bdevs_operational": 2, 00:25:31.012 "base_bdevs_list": [ 00:25:31.012 { 00:25:31.012 "name": null, 00:25:31.012 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:31.012 "is_configured": false, 00:25:31.012 "data_offset": 2048, 00:25:31.012 "data_size": 63488 00:25:31.012 }, 00:25:31.012 { 00:25:31.012 "name": null, 00:25:31.012 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:31.012 "is_configured": false, 00:25:31.012 "data_offset": 2048, 00:25:31.012 "data_size": 63488 00:25:31.012 }, 00:25:31.012 { 00:25:31.012 "name": "BaseBdev3", 00:25:31.012 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:31.012 "is_configured": true, 00:25:31.012 "data_offset": 2048, 00:25:31.012 "data_size": 63488 00:25:31.012 }, 00:25:31.012 { 00:25:31.012 "name": "BaseBdev4", 00:25:31.012 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:31.012 "is_configured": true, 00:25:31.012 "data_offset": 2048, 00:25:31.012 "data_size": 63488 00:25:31.012 } 00:25:31.012 ] 00:25:31.012 }' 00:25:31.012 11:36:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:31.012 11:36:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:25:31.012 11:36:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:31.012 11:36:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:25:31.012 11:36:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@771 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete BaseBdev1 00:25:31.270 11:36:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@772 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:25:31.528 [2024-07-15 11:36:14.990181] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:25:31.528 [2024-07-15 11:36:14.990224] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:31.528 [2024-07-15 11:36:14.990244] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2595650 00:25:31.528 [2024-07-15 11:36:14.990256] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:31.528 [2024-07-15 11:36:14.990604] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:31.528 [2024-07-15 11:36:14.990623] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:25:31.528 [2024-07-15 11:36:14.990687] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:25:31.528 [2024-07-15 11:36:14.990701] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:25:31.528 [2024-07-15 11:36:14.990717] bdev_raid.c:3581:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:25:31.528 BaseBdev1 00:25:31.528 11:36:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # sleep 1 00:25:32.461 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:25:32.461 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:25:32.461 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:25:32.461 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:25:32.461 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:25:32.461 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:25:32.461 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:25:32.461 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:25:32.461 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:25:32.461 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:25:32.461 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:32.461 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:32.719 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:25:32.719 "name": "raid_bdev1", 00:25:32.719 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:32.719 "strip_size_kb": 0, 00:25:32.719 "state": "online", 00:25:32.719 "raid_level": "raid1", 00:25:32.719 "superblock": true, 00:25:32.719 "num_base_bdevs": 4, 00:25:32.719 "num_base_bdevs_discovered": 2, 00:25:32.719 "num_base_bdevs_operational": 2, 00:25:32.719 "base_bdevs_list": [ 00:25:32.719 { 00:25:32.719 "name": null, 00:25:32.719 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:32.719 "is_configured": false, 00:25:32.719 "data_offset": 2048, 00:25:32.719 "data_size": 63488 00:25:32.719 }, 00:25:32.719 { 00:25:32.719 "name": null, 00:25:32.719 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:32.719 "is_configured": false, 00:25:32.719 "data_offset": 2048, 00:25:32.719 "data_size": 63488 00:25:32.719 }, 00:25:32.719 { 00:25:32.719 "name": "BaseBdev3", 00:25:32.719 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:32.719 "is_configured": true, 00:25:32.719 "data_offset": 2048, 00:25:32.719 "data_size": 63488 00:25:32.719 }, 00:25:32.719 { 00:25:32.719 "name": "BaseBdev4", 00:25:32.719 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:32.719 "is_configured": true, 00:25:32.719 "data_offset": 2048, 00:25:32.719 "data_size": 63488 00:25:32.719 } 00:25:32.719 ] 00:25:32.719 }' 00:25:32.719 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:25:32.719 11:36:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:33.285 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # verify_raid_bdev_process raid_bdev1 none none 00:25:33.285 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:33.285 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:25:33.285 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=none 00:25:33.285 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:33.285 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:33.285 11:36:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:33.542 11:36:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:33.542 "name": "raid_bdev1", 00:25:33.542 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:33.542 "strip_size_kb": 0, 00:25:33.542 "state": "online", 00:25:33.542 "raid_level": "raid1", 00:25:33.542 "superblock": true, 00:25:33.542 "num_base_bdevs": 4, 00:25:33.542 "num_base_bdevs_discovered": 2, 00:25:33.542 "num_base_bdevs_operational": 2, 00:25:33.542 "base_bdevs_list": [ 00:25:33.542 { 00:25:33.542 "name": null, 00:25:33.542 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:33.542 "is_configured": false, 00:25:33.542 "data_offset": 2048, 00:25:33.542 "data_size": 63488 00:25:33.542 }, 00:25:33.542 { 00:25:33.543 "name": null, 00:25:33.543 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:33.543 "is_configured": false, 00:25:33.543 "data_offset": 2048, 00:25:33.543 "data_size": 63488 00:25:33.543 }, 00:25:33.543 { 00:25:33.543 "name": "BaseBdev3", 00:25:33.543 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:33.543 "is_configured": true, 00:25:33.543 "data_offset": 2048, 00:25:33.543 "data_size": 63488 00:25:33.543 }, 00:25:33.543 { 00:25:33.543 "name": "BaseBdev4", 00:25:33.543 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:33.543 "is_configured": true, 00:25:33.543 "data_offset": 2048, 00:25:33.543 "data_size": 63488 00:25:33.543 } 00:25:33.543 ] 00:25:33.543 }' 00:25:33.543 11:36:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:33.801 11:36:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:25:33.801 11:36:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:33.801 11:36:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:25:33.801 11:36:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:25:33.801 11:36:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@648 -- # local es=0 00:25:33.801 11:36:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:25:33.801 11:36:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:25:33.801 11:36:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:25:33.801 11:36:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:25:33.801 11:36:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:25:33.801 11:36:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:25:33.801 11:36:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:25:33.801 11:36:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:25:33.801 11:36:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py ]] 00:25:33.801 11:36:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:25:33.801 [2024-07-15 11:36:17.372525] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:25:33.801 [2024-07-15 11:36:17.372650] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:25:33.801 [2024-07-15 11:36:17.372666] bdev_raid.c:3581:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:25:33.801 request: 00:25:33.801 { 00:25:33.801 "base_bdev": "BaseBdev1", 00:25:33.801 "raid_bdev": "raid_bdev1", 00:25:33.801 "method": "bdev_raid_add_base_bdev", 00:25:33.801 "req_id": 1 00:25:33.801 } 00:25:33.801 Got JSON-RPC error response 00:25:33.801 response: 00:25:33.801 { 00:25:33.801 "code": -22, 00:25:33.801 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:25:33.801 } 00:25:33.801 11:36:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@651 -- # es=1 00:25:33.801 11:36:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:25:33.801 11:36:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:25:33.801 11:36:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:25:33.801 11:36:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # sleep 1 00:25:35.174 11:36:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:25:35.174 11:36:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:25:35.174 11:36:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:25:35.174 11:36:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:25:35.174 11:36:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:25:35.174 11:36:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:25:35.174 11:36:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:25:35.174 11:36:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:25:35.174 11:36:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:25:35.174 11:36:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@124 -- # local tmp 00:25:35.174 11:36:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:35.174 11:36:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:35.174 11:36:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:25:35.174 "name": "raid_bdev1", 00:25:35.174 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:35.174 "strip_size_kb": 0, 00:25:35.174 "state": "online", 00:25:35.174 "raid_level": "raid1", 00:25:35.174 "superblock": true, 00:25:35.174 "num_base_bdevs": 4, 00:25:35.174 "num_base_bdevs_discovered": 2, 00:25:35.174 "num_base_bdevs_operational": 2, 00:25:35.174 "base_bdevs_list": [ 00:25:35.174 { 00:25:35.174 "name": null, 00:25:35.174 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:35.174 "is_configured": false, 00:25:35.174 "data_offset": 2048, 00:25:35.174 "data_size": 63488 00:25:35.174 }, 00:25:35.174 { 00:25:35.174 "name": null, 00:25:35.174 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:35.174 "is_configured": false, 00:25:35.174 "data_offset": 2048, 00:25:35.174 "data_size": 63488 00:25:35.174 }, 00:25:35.174 { 00:25:35.174 "name": "BaseBdev3", 00:25:35.174 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:35.174 "is_configured": true, 00:25:35.174 "data_offset": 2048, 00:25:35.174 "data_size": 63488 00:25:35.174 }, 00:25:35.174 { 00:25:35.174 "name": "BaseBdev4", 00:25:35.174 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:35.174 "is_configured": true, 00:25:35.174 "data_offset": 2048, 00:25:35.174 "data_size": 63488 00:25:35.174 } 00:25:35.174 ] 00:25:35.174 }' 00:25:35.174 11:36:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:25:35.174 11:36:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:35.738 11:36:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # verify_raid_bdev_process raid_bdev1 none none 00:25:35.738 11:36:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:35.738 11:36:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:25:35.738 11:36:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@184 -- # local target=none 00:25:35.738 11:36:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:35.738 11:36:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:35.738 11:36:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:35.995 11:36:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:35.995 "name": "raid_bdev1", 00:25:35.995 "uuid": "6e930040-fe33-45ac-ae46-fc73dd715111", 00:25:35.995 "strip_size_kb": 0, 00:25:35.995 "state": "online", 00:25:35.995 "raid_level": "raid1", 00:25:35.995 "superblock": true, 00:25:35.995 "num_base_bdevs": 4, 00:25:35.995 "num_base_bdevs_discovered": 2, 00:25:35.995 "num_base_bdevs_operational": 2, 00:25:35.995 "base_bdevs_list": [ 00:25:35.995 { 00:25:35.996 "name": null, 00:25:35.996 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:35.996 "is_configured": false, 00:25:35.996 "data_offset": 2048, 00:25:35.996 "data_size": 63488 00:25:35.996 }, 00:25:35.996 { 00:25:35.996 "name": null, 00:25:35.996 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:35.996 "is_configured": false, 00:25:35.996 "data_offset": 2048, 00:25:35.996 "data_size": 63488 00:25:35.996 }, 00:25:35.996 { 00:25:35.996 "name": "BaseBdev3", 00:25:35.996 "uuid": "73d7a6b4-7562-598f-84b6-929151420491", 00:25:35.996 "is_configured": true, 00:25:35.996 "data_offset": 2048, 00:25:35.996 "data_size": 63488 00:25:35.996 }, 00:25:35.996 { 00:25:35.996 "name": "BaseBdev4", 00:25:35.996 "uuid": "29546299-f528-5354-a1bc-89a9d871b0b5", 00:25:35.996 "is_configured": true, 00:25:35.996 "data_offset": 2048, 00:25:35.996 "data_size": 63488 00:25:35.996 } 00:25:35.996 ] 00:25:35.996 }' 00:25:35.996 11:36:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:35.996 11:36:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:25:35.996 11:36:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:35.996 11:36:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:25:35.996 11:36:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@782 -- # killprocess 987019 00:25:35.996 11:36:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@948 -- # '[' -z 987019 ']' 00:25:35.996 11:36:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@952 -- # kill -0 987019 00:25:35.996 11:36:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@953 -- # uname 00:25:35.996 11:36:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:25:35.996 11:36:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 987019 00:25:35.996 11:36:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:25:35.996 11:36:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:25:35.996 11:36:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@966 -- # echo 'killing process with pid 987019' 00:25:35.996 killing process with pid 987019 00:25:35.996 11:36:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@967 -- # kill 987019 00:25:35.996 Received shutdown signal, test time was about 60.000000 seconds 00:25:35.996 00:25:35.996 Latency(us) 00:25:35.996 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:25:35.996 =================================================================================================================== 00:25:35.996 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:25:35.996 [2024-07-15 11:36:19.568063] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:25:35.996 [2024-07-15 11:36:19.568166] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:25:35.996 11:36:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@972 -- # wait 987019 00:25:35.996 [2024-07-15 11:36:19.568222] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:25:35.996 [2024-07-15 11:36:19.568236] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x2598ec0 name raid_bdev1, state offline 00:25:36.253 [2024-07-15 11:36:19.622209] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # return 0 00:25:36.511 00:25:36.511 real 0m39.139s 00:25:36.511 user 0m56.215s 00:25:36.511 sys 0m7.379s 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1124 -- # xtrace_disable 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:36.511 ************************************ 00:25:36.511 END TEST raid_rebuild_test_sb 00:25:36.511 ************************************ 00:25:36.511 11:36:19 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:25:36.511 11:36:19 bdev_raid -- bdev/bdev_raid.sh@879 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 4 false true true 00:25:36.511 11:36:19 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:25:36.511 11:36:19 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:25:36.511 11:36:19 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:25:36.511 ************************************ 00:25:36.511 START TEST raid_rebuild_test_io 00:25:36.511 ************************************ 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1123 -- # raid_rebuild_test raid1 4 false true true 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@568 -- # local raid_level=raid1 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local num_base_bdevs=4 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local superblock=false 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local background_io=true 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local verify=true 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # (( i = 1 )) 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # echo BaseBdev1 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # echo BaseBdev2 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # echo BaseBdev3 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # echo BaseBdev4 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local base_bdevs 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local raid_bdev_name=raid_bdev1 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local strip_size 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local create_arg 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local raid_bdev_size 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local data_offset 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@580 -- # '[' raid1 '!=' raid1 ']' 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@588 -- # strip_size=0 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@591 -- # '[' false = true ']' 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # raid_pid=993073 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # waitforlisten 993073 /var/tmp/spdk-raid.sock 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@595 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@829 -- # '[' -z 993073 ']' 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@834 -- # local max_retries=100 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:25:36.511 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@838 -- # xtrace_disable 00:25:36.511 11:36:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:25:36.511 [2024-07-15 11:36:20.013168] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:25:36.511 [2024-07-15 11:36:20.013243] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid993073 ] 00:25:36.511 I/O size of 3145728 is greater than zero copy threshold (65536). 00:25:36.511 Zero copy mechanism will not be used. 00:25:36.768 [2024-07-15 11:36:20.144443] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:36.769 [2024-07-15 11:36:20.241402] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:25:36.769 [2024-07-15 11:36:20.299736] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:25:36.769 [2024-07-15 11:36:20.299769] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:25:37.701 11:36:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:25:37.701 11:36:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@862 -- # return 0 00:25:37.701 11:36:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:25:37.701 11:36:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:25:37.701 BaseBdev1_malloc 00:25:37.701 11:36:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:25:37.701 [2024-07-15 11:36:21.291393] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:25:37.701 [2024-07-15 11:36:21.291442] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:37.701 [2024-07-15 11:36:21.291468] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1428d40 00:25:37.701 [2024-07-15 11:36:21.291480] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:37.701 [2024-07-15 11:36:21.293126] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:37.701 [2024-07-15 11:36:21.293156] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:25:37.959 BaseBdev1 00:25:37.959 11:36:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:25:37.959 11:36:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:25:37.959 BaseBdev2_malloc 00:25:38.217 11:36:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:25:38.217 [2024-07-15 11:36:21.777623] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:25:38.217 [2024-07-15 11:36:21.777667] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:38.217 [2024-07-15 11:36:21.777694] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1429860 00:25:38.217 [2024-07-15 11:36:21.777706] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:38.217 [2024-07-15 11:36:21.779094] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:38.217 [2024-07-15 11:36:21.779124] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:25:38.217 BaseBdev2 00:25:38.217 11:36:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:25:38.217 11:36:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:25:38.474 BaseBdev3_malloc 00:25:38.474 11:36:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:25:38.731 [2024-07-15 11:36:22.207375] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:25:38.731 [2024-07-15 11:36:22.207425] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:38.731 [2024-07-15 11:36:22.207450] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x15d68f0 00:25:38.731 [2024-07-15 11:36:22.207463] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:38.731 [2024-07-15 11:36:22.208960] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:38.731 [2024-07-15 11:36:22.208990] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:25:38.731 BaseBdev3 00:25:38.731 11:36:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:25:38.731 11:36:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:25:38.988 BaseBdev4_malloc 00:25:38.988 11:36:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:25:38.988 [2024-07-15 11:36:22.560983] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:25:38.988 [2024-07-15 11:36:22.561031] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:38.988 [2024-07-15 11:36:22.561052] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x15d5ad0 00:25:38.988 [2024-07-15 11:36:22.561065] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:38.988 [2024-07-15 11:36:22.562422] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:38.988 [2024-07-15 11:36:22.562451] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:25:38.988 BaseBdev4 00:25:39.246 11:36:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@606 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b spare_malloc 00:25:39.246 spare_malloc 00:25:39.246 11:36:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:25:39.504 spare_delay 00:25:39.504 11:36:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:25:39.504 [2024-07-15 11:36:23.090911] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:25:39.504 [2024-07-15 11:36:23.090962] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:39.504 [2024-07-15 11:36:23.090983] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x15da5b0 00:25:39.504 [2024-07-15 11:36:23.090996] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:39.504 [2024-07-15 11:36:23.092403] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:39.504 [2024-07-15 11:36:23.092431] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:25:39.504 spare 00:25:39.761 11:36:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@611 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n raid_bdev1 00:25:39.761 [2024-07-15 11:36:23.331571] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:25:39.761 [2024-07-15 11:36:23.332854] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:25:39.761 [2024-07-15 11:36:23.332910] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:25:39.761 [2024-07-15 11:36:23.332963] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:25:39.761 [2024-07-15 11:36:23.333045] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x15598a0 00:25:39.761 [2024-07-15 11:36:23.333056] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:25:39.761 [2024-07-15 11:36:23.333271] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x15d3e10 00:25:39.761 [2024-07-15 11:36:23.333421] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x15598a0 00:25:39.761 [2024-07-15 11:36:23.333432] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x15598a0 00:25:39.761 [2024-07-15 11:36:23.333544] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:39.761 11:36:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:25:39.761 11:36:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:25:39.761 11:36:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:25:39.761 11:36:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:25:39.761 11:36:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:25:39.761 11:36:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:25:39.761 11:36:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:25:39.761 11:36:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:25:40.019 11:36:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:25:40.019 11:36:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:25:40.019 11:36:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:40.019 11:36:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:40.019 11:36:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:25:40.019 "name": "raid_bdev1", 00:25:40.019 "uuid": "f28c3119-d110-48e4-8d17-d2f0a0aabecf", 00:25:40.019 "strip_size_kb": 0, 00:25:40.019 "state": "online", 00:25:40.019 "raid_level": "raid1", 00:25:40.019 "superblock": false, 00:25:40.019 "num_base_bdevs": 4, 00:25:40.019 "num_base_bdevs_discovered": 4, 00:25:40.019 "num_base_bdevs_operational": 4, 00:25:40.019 "base_bdevs_list": [ 00:25:40.019 { 00:25:40.019 "name": "BaseBdev1", 00:25:40.019 "uuid": "342f28d8-8a02-573e-a7e6-11a0732f0b52", 00:25:40.019 "is_configured": true, 00:25:40.019 "data_offset": 0, 00:25:40.019 "data_size": 65536 00:25:40.019 }, 00:25:40.019 { 00:25:40.019 "name": "BaseBdev2", 00:25:40.019 "uuid": "766d35e9-0460-5384-b711-97d474366077", 00:25:40.019 "is_configured": true, 00:25:40.019 "data_offset": 0, 00:25:40.019 "data_size": 65536 00:25:40.019 }, 00:25:40.019 { 00:25:40.019 "name": "BaseBdev3", 00:25:40.019 "uuid": "5596fee9-60e6-5363-94c2-166425c3531d", 00:25:40.019 "is_configured": true, 00:25:40.019 "data_offset": 0, 00:25:40.019 "data_size": 65536 00:25:40.019 }, 00:25:40.019 { 00:25:40.019 "name": "BaseBdev4", 00:25:40.019 "uuid": "28274275-afa4-5f66-88c5-3c71427ba318", 00:25:40.019 "is_configured": true, 00:25:40.019 "data_offset": 0, 00:25:40.019 "data_size": 65536 00:25:40.019 } 00:25:40.019 ] 00:25:40.019 }' 00:25:40.019 11:36:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:25:40.019 11:36:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:25:40.585 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@615 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:25:40.585 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@615 -- # jq -r '.[].num_blocks' 00:25:40.842 [2024-07-15 11:36:24.358570] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:25:40.842 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@615 -- # raid_bdev_size=65536 00:25:40.842 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@618 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:40.842 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@618 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:25:41.112 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@618 -- # data_offset=0 00:25:41.112 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@620 -- # '[' true = true ']' 00:25:41.112 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@639 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev1 00:25:41.112 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@622 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:25:41.369 [2024-07-15 11:36:24.745701] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x155f970 00:25:41.369 I/O size of 3145728 is greater than zero copy threshold (65536). 00:25:41.369 Zero copy mechanism will not be used. 00:25:41.369 Running I/O for 60 seconds... 00:25:41.369 [2024-07-15 11:36:24.867938] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:25:41.369 [2024-07-15 11:36:24.876164] bdev_raid.c:1919:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x155f970 00:25:41.369 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@642 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:25:41.369 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:25:41.369 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:25:41.369 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:25:41.369 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:25:41.369 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:25:41.369 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:25:41.369 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:25:41.369 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:25:41.369 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:25:41.370 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:41.370 11:36:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:41.627 11:36:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:25:41.627 "name": "raid_bdev1", 00:25:41.627 "uuid": "f28c3119-d110-48e4-8d17-d2f0a0aabecf", 00:25:41.627 "strip_size_kb": 0, 00:25:41.627 "state": "online", 00:25:41.627 "raid_level": "raid1", 00:25:41.627 "superblock": false, 00:25:41.627 "num_base_bdevs": 4, 00:25:41.627 "num_base_bdevs_discovered": 3, 00:25:41.627 "num_base_bdevs_operational": 3, 00:25:41.627 "base_bdevs_list": [ 00:25:41.627 { 00:25:41.627 "name": null, 00:25:41.627 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:41.627 "is_configured": false, 00:25:41.627 "data_offset": 0, 00:25:41.627 "data_size": 65536 00:25:41.627 }, 00:25:41.627 { 00:25:41.627 "name": "BaseBdev2", 00:25:41.627 "uuid": "766d35e9-0460-5384-b711-97d474366077", 00:25:41.627 "is_configured": true, 00:25:41.627 "data_offset": 0, 00:25:41.627 "data_size": 65536 00:25:41.627 }, 00:25:41.627 { 00:25:41.627 "name": "BaseBdev3", 00:25:41.627 "uuid": "5596fee9-60e6-5363-94c2-166425c3531d", 00:25:41.627 "is_configured": true, 00:25:41.627 "data_offset": 0, 00:25:41.627 "data_size": 65536 00:25:41.627 }, 00:25:41.627 { 00:25:41.627 "name": "BaseBdev4", 00:25:41.627 "uuid": "28274275-afa4-5f66-88c5-3c71427ba318", 00:25:41.627 "is_configured": true, 00:25:41.627 "data_offset": 0, 00:25:41.627 "data_size": 65536 00:25:41.627 } 00:25:41.627 ] 00:25:41.627 }' 00:25:41.627 11:36:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:25:41.627 11:36:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:25:42.558 11:36:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@645 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:25:42.558 [2024-07-15 11:36:26.019371] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:25:42.558 11:36:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # sleep 1 00:25:42.558 [2024-07-15 11:36:26.103338] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x112ffa0 00:25:42.558 [2024-07-15 11:36:26.105733] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:25:42.814 [2024-07-15 11:36:26.252092] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:25:43.071 [2024-07-15 11:36:26.490429] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:25:43.071 [2024-07-15 11:36:26.491085] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:25:43.328 [2024-07-15 11:36:26.856423] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:25:43.585 [2024-07-15 11:36:26.958862] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:25:43.585 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@649 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:25:43.585 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:43.585 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:25:43.585 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:25:43.585 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:43.585 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:43.585 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:43.843 [2024-07-15 11:36:27.281363] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:25:43.843 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:43.843 "name": "raid_bdev1", 00:25:43.843 "uuid": "f28c3119-d110-48e4-8d17-d2f0a0aabecf", 00:25:43.843 "strip_size_kb": 0, 00:25:43.843 "state": "online", 00:25:43.843 "raid_level": "raid1", 00:25:43.843 "superblock": false, 00:25:43.843 "num_base_bdevs": 4, 00:25:43.843 "num_base_bdevs_discovered": 4, 00:25:43.843 "num_base_bdevs_operational": 4, 00:25:43.843 "process": { 00:25:43.843 "type": "rebuild", 00:25:43.843 "target": "spare", 00:25:43.843 "progress": { 00:25:43.843 "blocks": 16384, 00:25:43.843 "percent": 25 00:25:43.843 } 00:25:43.843 }, 00:25:43.843 "base_bdevs_list": [ 00:25:43.843 { 00:25:43.843 "name": "spare", 00:25:43.843 "uuid": "06d5a07d-dd35-5980-a254-0327d69f9ce8", 00:25:43.843 "is_configured": true, 00:25:43.843 "data_offset": 0, 00:25:43.843 "data_size": 65536 00:25:43.843 }, 00:25:43.843 { 00:25:43.843 "name": "BaseBdev2", 00:25:43.843 "uuid": "766d35e9-0460-5384-b711-97d474366077", 00:25:43.843 "is_configured": true, 00:25:43.843 "data_offset": 0, 00:25:43.843 "data_size": 65536 00:25:43.843 }, 00:25:43.843 { 00:25:43.843 "name": "BaseBdev3", 00:25:43.843 "uuid": "5596fee9-60e6-5363-94c2-166425c3531d", 00:25:43.843 "is_configured": true, 00:25:43.843 "data_offset": 0, 00:25:43.843 "data_size": 65536 00:25:43.843 }, 00:25:43.843 { 00:25:43.843 "name": "BaseBdev4", 00:25:43.843 "uuid": "28274275-afa4-5f66-88c5-3c71427ba318", 00:25:43.843 "is_configured": true, 00:25:43.843 "data_offset": 0, 00:25:43.843 "data_size": 65536 00:25:43.843 } 00:25:43.843 ] 00:25:43.843 }' 00:25:43.843 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:43.843 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:25:43.843 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:43.843 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:25:43.843 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@652 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev spare 00:25:44.102 [2024-07-15 11:36:27.648862] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:25:44.392 [2024-07-15 11:36:27.748889] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:25:44.392 [2024-07-15 11:36:27.851706] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:25:44.392 [2024-07-15 11:36:27.862981] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:44.392 [2024-07-15 11:36:27.863016] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:25:44.392 [2024-07-15 11:36:27.863027] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:25:44.392 [2024-07-15 11:36:27.895169] bdev_raid.c:1919:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x155f970 00:25:44.392 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@655 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:25:44.392 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:25:44.392 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:25:44.392 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:25:44.392 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:25:44.392 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:25:44.392 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:25:44.392 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:25:44.392 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:25:44.392 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:25:44.392 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:44.392 11:36:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:44.653 11:36:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:25:44.653 "name": "raid_bdev1", 00:25:44.653 "uuid": "f28c3119-d110-48e4-8d17-d2f0a0aabecf", 00:25:44.653 "strip_size_kb": 0, 00:25:44.653 "state": "online", 00:25:44.653 "raid_level": "raid1", 00:25:44.653 "superblock": false, 00:25:44.653 "num_base_bdevs": 4, 00:25:44.653 "num_base_bdevs_discovered": 3, 00:25:44.653 "num_base_bdevs_operational": 3, 00:25:44.653 "base_bdevs_list": [ 00:25:44.653 { 00:25:44.653 "name": null, 00:25:44.653 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:44.653 "is_configured": false, 00:25:44.653 "data_offset": 0, 00:25:44.653 "data_size": 65536 00:25:44.653 }, 00:25:44.653 { 00:25:44.653 "name": "BaseBdev2", 00:25:44.653 "uuid": "766d35e9-0460-5384-b711-97d474366077", 00:25:44.653 "is_configured": true, 00:25:44.653 "data_offset": 0, 00:25:44.653 "data_size": 65536 00:25:44.653 }, 00:25:44.653 { 00:25:44.653 "name": "BaseBdev3", 00:25:44.653 "uuid": "5596fee9-60e6-5363-94c2-166425c3531d", 00:25:44.653 "is_configured": true, 00:25:44.653 "data_offset": 0, 00:25:44.653 "data_size": 65536 00:25:44.653 }, 00:25:44.653 { 00:25:44.653 "name": "BaseBdev4", 00:25:44.653 "uuid": "28274275-afa4-5f66-88c5-3c71427ba318", 00:25:44.653 "is_configured": true, 00:25:44.653 "data_offset": 0, 00:25:44.653 "data_size": 65536 00:25:44.653 } 00:25:44.653 ] 00:25:44.653 }' 00:25:44.653 11:36:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:25:44.653 11:36:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:25:45.587 11:36:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@658 -- # verify_raid_bdev_process raid_bdev1 none none 00:25:45.587 11:36:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:45.587 11:36:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:25:45.587 11:36:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@184 -- # local target=none 00:25:45.587 11:36:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:45.587 11:36:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:45.587 11:36:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:45.587 11:36:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:45.587 "name": "raid_bdev1", 00:25:45.587 "uuid": "f28c3119-d110-48e4-8d17-d2f0a0aabecf", 00:25:45.587 "strip_size_kb": 0, 00:25:45.587 "state": "online", 00:25:45.587 "raid_level": "raid1", 00:25:45.587 "superblock": false, 00:25:45.587 "num_base_bdevs": 4, 00:25:45.587 "num_base_bdevs_discovered": 3, 00:25:45.587 "num_base_bdevs_operational": 3, 00:25:45.587 "base_bdevs_list": [ 00:25:45.587 { 00:25:45.587 "name": null, 00:25:45.587 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:45.587 "is_configured": false, 00:25:45.587 "data_offset": 0, 00:25:45.587 "data_size": 65536 00:25:45.587 }, 00:25:45.587 { 00:25:45.587 "name": "BaseBdev2", 00:25:45.587 "uuid": "766d35e9-0460-5384-b711-97d474366077", 00:25:45.587 "is_configured": true, 00:25:45.587 "data_offset": 0, 00:25:45.587 "data_size": 65536 00:25:45.587 }, 00:25:45.587 { 00:25:45.587 "name": "BaseBdev3", 00:25:45.587 "uuid": "5596fee9-60e6-5363-94c2-166425c3531d", 00:25:45.587 "is_configured": true, 00:25:45.587 "data_offset": 0, 00:25:45.587 "data_size": 65536 00:25:45.587 }, 00:25:45.587 { 00:25:45.587 "name": "BaseBdev4", 00:25:45.587 "uuid": "28274275-afa4-5f66-88c5-3c71427ba318", 00:25:45.587 "is_configured": true, 00:25:45.587 "data_offset": 0, 00:25:45.587 "data_size": 65536 00:25:45.587 } 00:25:45.587 ] 00:25:45.587 }' 00:25:45.587 11:36:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:45.587 11:36:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:25:45.587 11:36:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:45.845 11:36:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:25:45.845 11:36:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@661 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:25:45.845 [2024-07-15 11:36:29.407517] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:25:46.102 11:36:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # sleep 1 00:25:46.102 [2024-07-15 11:36:29.473094] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1560150 00:25:46.102 [2024-07-15 11:36:29.474628] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:25:46.102 [2024-07-15 11:36:29.610310] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:25:46.102 [2024-07-15 11:36:29.611771] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:25:46.358 [2024-07-15 11:36:29.843330] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:25:46.358 [2024-07-15 11:36:29.844016] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:25:46.614 [2024-07-15 11:36:30.202833] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:25:46.872 [2024-07-15 11:36:30.324951] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:25:47.129 11:36:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:25:47.129 11:36:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:47.129 11:36:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:25:47.129 11:36:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:25:47.129 11:36:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:47.129 11:36:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:47.129 11:36:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:47.129 [2024-07-15 11:36:30.578824] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:25:47.129 [2024-07-15 11:36:30.580002] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:25:47.386 11:36:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:47.386 "name": "raid_bdev1", 00:25:47.386 "uuid": "f28c3119-d110-48e4-8d17-d2f0a0aabecf", 00:25:47.386 "strip_size_kb": 0, 00:25:47.386 "state": "online", 00:25:47.386 "raid_level": "raid1", 00:25:47.386 "superblock": false, 00:25:47.386 "num_base_bdevs": 4, 00:25:47.386 "num_base_bdevs_discovered": 4, 00:25:47.386 "num_base_bdevs_operational": 4, 00:25:47.386 "process": { 00:25:47.386 "type": "rebuild", 00:25:47.386 "target": "spare", 00:25:47.386 "progress": { 00:25:47.386 "blocks": 14336, 00:25:47.386 "percent": 21 00:25:47.386 } 00:25:47.386 }, 00:25:47.386 "base_bdevs_list": [ 00:25:47.386 { 00:25:47.386 "name": "spare", 00:25:47.386 "uuid": "06d5a07d-dd35-5980-a254-0327d69f9ce8", 00:25:47.386 "is_configured": true, 00:25:47.386 "data_offset": 0, 00:25:47.386 "data_size": 65536 00:25:47.386 }, 00:25:47.386 { 00:25:47.386 "name": "BaseBdev2", 00:25:47.386 "uuid": "766d35e9-0460-5384-b711-97d474366077", 00:25:47.386 "is_configured": true, 00:25:47.386 "data_offset": 0, 00:25:47.386 "data_size": 65536 00:25:47.386 }, 00:25:47.386 { 00:25:47.386 "name": "BaseBdev3", 00:25:47.386 "uuid": "5596fee9-60e6-5363-94c2-166425c3531d", 00:25:47.386 "is_configured": true, 00:25:47.386 "data_offset": 0, 00:25:47.386 "data_size": 65536 00:25:47.386 }, 00:25:47.386 { 00:25:47.386 "name": "BaseBdev4", 00:25:47.386 "uuid": "28274275-afa4-5f66-88c5-3c71427ba318", 00:25:47.386 "is_configured": true, 00:25:47.387 "data_offset": 0, 00:25:47.387 "data_size": 65536 00:25:47.387 } 00:25:47.387 ] 00:25:47.387 }' 00:25:47.387 11:36:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:47.387 11:36:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:25:47.387 11:36:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:47.387 [2024-07-15 11:36:30.810220] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:25:47.387 11:36:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:25:47.387 11:36:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@665 -- # '[' false = true ']' 00:25:47.387 11:36:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@690 -- # local num_base_bdevs_operational=4 00:25:47.387 11:36:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@692 -- # '[' raid1 = raid1 ']' 00:25:47.387 11:36:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@692 -- # '[' 4 -gt 2 ']' 00:25:47.387 11:36:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@694 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev2 00:25:47.643 [2024-07-15 11:36:31.056558] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:25:47.643 [2024-07-15 11:36:31.196791] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:25:47.900 [2024-07-15 11:36:31.299207] bdev_raid.c:1919:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x155f970 00:25:47.900 [2024-07-15 11:36:31.299236] bdev_raid.c:1919:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x1560150 00:25:47.900 [2024-07-15 11:36:31.299280] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:25:47.900 [2024-07-15 11:36:31.317605] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:25:47.900 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@697 -- # base_bdevs[1]= 00:25:47.900 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@698 -- # (( num_base_bdevs_operational-- )) 00:25:47.900 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@701 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:25:47.900 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:47.900 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:25:47.900 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:25:47.900 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:47.900 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:47.900 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:48.157 [2024-07-15 11:36:31.548247] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:25:48.157 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:48.157 "name": "raid_bdev1", 00:25:48.157 "uuid": "f28c3119-d110-48e4-8d17-d2f0a0aabecf", 00:25:48.157 "strip_size_kb": 0, 00:25:48.157 "state": "online", 00:25:48.157 "raid_level": "raid1", 00:25:48.157 "superblock": false, 00:25:48.157 "num_base_bdevs": 4, 00:25:48.157 "num_base_bdevs_discovered": 3, 00:25:48.157 "num_base_bdevs_operational": 3, 00:25:48.157 "process": { 00:25:48.157 "type": "rebuild", 00:25:48.157 "target": "spare", 00:25:48.157 "progress": { 00:25:48.157 "blocks": 22528, 00:25:48.157 "percent": 34 00:25:48.157 } 00:25:48.157 }, 00:25:48.157 "base_bdevs_list": [ 00:25:48.157 { 00:25:48.157 "name": "spare", 00:25:48.157 "uuid": "06d5a07d-dd35-5980-a254-0327d69f9ce8", 00:25:48.157 "is_configured": true, 00:25:48.157 "data_offset": 0, 00:25:48.157 "data_size": 65536 00:25:48.157 }, 00:25:48.157 { 00:25:48.157 "name": null, 00:25:48.157 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:48.157 "is_configured": false, 00:25:48.157 "data_offset": 0, 00:25:48.157 "data_size": 65536 00:25:48.157 }, 00:25:48.157 { 00:25:48.157 "name": "BaseBdev3", 00:25:48.157 "uuid": "5596fee9-60e6-5363-94c2-166425c3531d", 00:25:48.157 "is_configured": true, 00:25:48.157 "data_offset": 0, 00:25:48.157 "data_size": 65536 00:25:48.157 }, 00:25:48.157 { 00:25:48.157 "name": "BaseBdev4", 00:25:48.157 "uuid": "28274275-afa4-5f66-88c5-3c71427ba318", 00:25:48.157 "is_configured": true, 00:25:48.157 "data_offset": 0, 00:25:48.157 "data_size": 65536 00:25:48.157 } 00:25:48.157 ] 00:25:48.157 }' 00:25:48.157 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:48.157 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:25:48.157 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:48.157 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:25:48.157 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@705 -- # local timeout=919 00:25:48.157 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:25:48.157 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:25:48.157 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:48.157 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:25:48.157 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:25:48.157 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:48.157 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:48.157 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:48.413 [2024-07-15 11:36:31.799566] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:25:48.413 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:48.413 "name": "raid_bdev1", 00:25:48.413 "uuid": "f28c3119-d110-48e4-8d17-d2f0a0aabecf", 00:25:48.413 "strip_size_kb": 0, 00:25:48.413 "state": "online", 00:25:48.413 "raid_level": "raid1", 00:25:48.413 "superblock": false, 00:25:48.413 "num_base_bdevs": 4, 00:25:48.413 "num_base_bdevs_discovered": 3, 00:25:48.413 "num_base_bdevs_operational": 3, 00:25:48.413 "process": { 00:25:48.413 "type": "rebuild", 00:25:48.413 "target": "spare", 00:25:48.413 "progress": { 00:25:48.413 "blocks": 26624, 00:25:48.413 "percent": 40 00:25:48.413 } 00:25:48.413 }, 00:25:48.413 "base_bdevs_list": [ 00:25:48.413 { 00:25:48.413 "name": "spare", 00:25:48.413 "uuid": "06d5a07d-dd35-5980-a254-0327d69f9ce8", 00:25:48.413 "is_configured": true, 00:25:48.413 "data_offset": 0, 00:25:48.413 "data_size": 65536 00:25:48.413 }, 00:25:48.413 { 00:25:48.413 "name": null, 00:25:48.413 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:48.413 "is_configured": false, 00:25:48.413 "data_offset": 0, 00:25:48.413 "data_size": 65536 00:25:48.413 }, 00:25:48.413 { 00:25:48.413 "name": "BaseBdev3", 00:25:48.413 "uuid": "5596fee9-60e6-5363-94c2-166425c3531d", 00:25:48.413 "is_configured": true, 00:25:48.413 "data_offset": 0, 00:25:48.413 "data_size": 65536 00:25:48.413 }, 00:25:48.413 { 00:25:48.413 "name": "BaseBdev4", 00:25:48.413 "uuid": "28274275-afa4-5f66-88c5-3c71427ba318", 00:25:48.413 "is_configured": true, 00:25:48.413 "data_offset": 0, 00:25:48.413 "data_size": 65536 00:25:48.414 } 00:25:48.414 ] 00:25:48.414 }' 00:25:48.414 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:48.414 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:25:48.414 11:36:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:48.670 11:36:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:25:48.670 11:36:32 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@710 -- # sleep 1 00:25:48.670 [2024-07-15 11:36:32.040049] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:25:48.926 [2024-07-15 11:36:32.378382] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:25:49.487 [2024-07-15 11:36:32.802951] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:25:49.487 [2024-07-15 11:36:32.803284] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:25:49.487 [2024-07-15 11:36:32.913645] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:25:49.487 [2024-07-15 11:36:32.914085] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:25:49.487 11:36:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:25:49.487 11:36:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:25:49.487 11:36:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:49.487 11:36:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:25:49.487 11:36:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:25:49.487 11:36:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:49.487 11:36:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:49.487 11:36:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:49.744 [2024-07-15 11:36:33.267551] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 45056 offset_begin: 43008 offset_end: 49152 00:25:49.999 [2024-07-15 11:36:33.487668] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:25:49.999 11:36:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:49.999 "name": "raid_bdev1", 00:25:49.999 "uuid": "f28c3119-d110-48e4-8d17-d2f0a0aabecf", 00:25:49.999 "strip_size_kb": 0, 00:25:49.999 "state": "online", 00:25:49.999 "raid_level": "raid1", 00:25:49.999 "superblock": false, 00:25:49.999 "num_base_bdevs": 4, 00:25:49.999 "num_base_bdevs_discovered": 3, 00:25:50.000 "num_base_bdevs_operational": 3, 00:25:50.000 "process": { 00:25:50.000 "type": "rebuild", 00:25:50.000 "target": "spare", 00:25:50.000 "progress": { 00:25:50.000 "blocks": 47104, 00:25:50.000 "percent": 71 00:25:50.000 } 00:25:50.000 }, 00:25:50.000 "base_bdevs_list": [ 00:25:50.000 { 00:25:50.000 "name": "spare", 00:25:50.000 "uuid": "06d5a07d-dd35-5980-a254-0327d69f9ce8", 00:25:50.000 "is_configured": true, 00:25:50.000 "data_offset": 0, 00:25:50.000 "data_size": 65536 00:25:50.000 }, 00:25:50.000 { 00:25:50.000 "name": null, 00:25:50.000 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:50.000 "is_configured": false, 00:25:50.000 "data_offset": 0, 00:25:50.000 "data_size": 65536 00:25:50.000 }, 00:25:50.000 { 00:25:50.000 "name": "BaseBdev3", 00:25:50.000 "uuid": "5596fee9-60e6-5363-94c2-166425c3531d", 00:25:50.000 "is_configured": true, 00:25:50.000 "data_offset": 0, 00:25:50.000 "data_size": 65536 00:25:50.000 }, 00:25:50.000 { 00:25:50.000 "name": "BaseBdev4", 00:25:50.000 "uuid": "28274275-afa4-5f66-88c5-3c71427ba318", 00:25:50.000 "is_configured": true, 00:25:50.000 "data_offset": 0, 00:25:50.000 "data_size": 65536 00:25:50.000 } 00:25:50.000 ] 00:25:50.000 }' 00:25:50.000 11:36:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:50.000 11:36:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:25:50.256 11:36:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:50.256 11:36:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:25:50.256 11:36:33 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@710 -- # sleep 1 00:25:50.514 [2024-07-15 11:36:33.949035] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:25:51.078 [2024-07-15 11:36:34.623068] bdev_raid.c:2789:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:25:51.078 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:25:51.078 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:25:51.078 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:51.078 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:25:51.078 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:25:51.078 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:51.078 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:51.078 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:51.335 [2024-07-15 11:36:34.723347] bdev_raid.c:2504:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:25:51.335 [2024-07-15 11:36:34.726050] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:51.335 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:51.335 "name": "raid_bdev1", 00:25:51.335 "uuid": "f28c3119-d110-48e4-8d17-d2f0a0aabecf", 00:25:51.335 "strip_size_kb": 0, 00:25:51.335 "state": "online", 00:25:51.335 "raid_level": "raid1", 00:25:51.335 "superblock": false, 00:25:51.335 "num_base_bdevs": 4, 00:25:51.335 "num_base_bdevs_discovered": 3, 00:25:51.335 "num_base_bdevs_operational": 3, 00:25:51.335 "base_bdevs_list": [ 00:25:51.335 { 00:25:51.335 "name": "spare", 00:25:51.335 "uuid": "06d5a07d-dd35-5980-a254-0327d69f9ce8", 00:25:51.335 "is_configured": true, 00:25:51.335 "data_offset": 0, 00:25:51.335 "data_size": 65536 00:25:51.335 }, 00:25:51.335 { 00:25:51.335 "name": null, 00:25:51.335 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:51.335 "is_configured": false, 00:25:51.335 "data_offset": 0, 00:25:51.335 "data_size": 65536 00:25:51.335 }, 00:25:51.335 { 00:25:51.335 "name": "BaseBdev3", 00:25:51.335 "uuid": "5596fee9-60e6-5363-94c2-166425c3531d", 00:25:51.335 "is_configured": true, 00:25:51.335 "data_offset": 0, 00:25:51.335 "data_size": 65536 00:25:51.336 }, 00:25:51.336 { 00:25:51.336 "name": "BaseBdev4", 00:25:51.336 "uuid": "28274275-afa4-5f66-88c5-3c71427ba318", 00:25:51.336 "is_configured": true, 00:25:51.336 "data_offset": 0, 00:25:51.336 "data_size": 65536 00:25:51.336 } 00:25:51.336 ] 00:25:51.336 }' 00:25:51.336 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:51.593 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # [[ none == \r\e\b\u\i\l\d ]] 00:25:51.593 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:51.593 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # [[ none == \s\p\a\r\e ]] 00:25:51.593 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # break 00:25:51.593 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@714 -- # verify_raid_bdev_process raid_bdev1 none none 00:25:51.594 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:25:51.594 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:25:51.594 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@184 -- # local target=none 00:25:51.594 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:25:51.594 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:51.594 11:36:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:51.852 11:36:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:51.852 "name": "raid_bdev1", 00:25:51.852 "uuid": "f28c3119-d110-48e4-8d17-d2f0a0aabecf", 00:25:51.852 "strip_size_kb": 0, 00:25:51.852 "state": "online", 00:25:51.852 "raid_level": "raid1", 00:25:51.852 "superblock": false, 00:25:51.852 "num_base_bdevs": 4, 00:25:51.852 "num_base_bdevs_discovered": 3, 00:25:51.852 "num_base_bdevs_operational": 3, 00:25:51.852 "base_bdevs_list": [ 00:25:51.852 { 00:25:51.852 "name": "spare", 00:25:51.852 "uuid": "06d5a07d-dd35-5980-a254-0327d69f9ce8", 00:25:51.852 "is_configured": true, 00:25:51.852 "data_offset": 0, 00:25:51.852 "data_size": 65536 00:25:51.852 }, 00:25:51.852 { 00:25:51.852 "name": null, 00:25:51.852 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:51.852 "is_configured": false, 00:25:51.852 "data_offset": 0, 00:25:51.852 "data_size": 65536 00:25:51.852 }, 00:25:51.852 { 00:25:51.852 "name": "BaseBdev3", 00:25:51.852 "uuid": "5596fee9-60e6-5363-94c2-166425c3531d", 00:25:51.852 "is_configured": true, 00:25:51.852 "data_offset": 0, 00:25:51.852 "data_size": 65536 00:25:51.852 }, 00:25:51.852 { 00:25:51.852 "name": "BaseBdev4", 00:25:51.852 "uuid": "28274275-afa4-5f66-88c5-3c71427ba318", 00:25:51.852 "is_configured": true, 00:25:51.852 "data_offset": 0, 00:25:51.852 "data_size": 65536 00:25:51.852 } 00:25:51.852 ] 00:25:51.852 }' 00:25:51.852 11:36:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:25:51.852 11:36:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:25:51.852 11:36:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:25:51.852 11:36:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:25:51.852 11:36:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:25:51.852 11:36:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:25:51.852 11:36:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:25:51.852 11:36:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:25:51.852 11:36:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:25:51.852 11:36:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:25:51.852 11:36:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:25:51.852 11:36:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:25:51.852 11:36:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:25:51.852 11:36:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:25:51.852 11:36:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:51.852 11:36:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:52.109 11:36:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:25:52.109 "name": "raid_bdev1", 00:25:52.109 "uuid": "f28c3119-d110-48e4-8d17-d2f0a0aabecf", 00:25:52.109 "strip_size_kb": 0, 00:25:52.109 "state": "online", 00:25:52.109 "raid_level": "raid1", 00:25:52.109 "superblock": false, 00:25:52.109 "num_base_bdevs": 4, 00:25:52.109 "num_base_bdevs_discovered": 3, 00:25:52.109 "num_base_bdevs_operational": 3, 00:25:52.109 "base_bdevs_list": [ 00:25:52.109 { 00:25:52.109 "name": "spare", 00:25:52.109 "uuid": "06d5a07d-dd35-5980-a254-0327d69f9ce8", 00:25:52.109 "is_configured": true, 00:25:52.109 "data_offset": 0, 00:25:52.110 "data_size": 65536 00:25:52.110 }, 00:25:52.110 { 00:25:52.110 "name": null, 00:25:52.110 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:52.110 "is_configured": false, 00:25:52.110 "data_offset": 0, 00:25:52.110 "data_size": 65536 00:25:52.110 }, 00:25:52.110 { 00:25:52.110 "name": "BaseBdev3", 00:25:52.110 "uuid": "5596fee9-60e6-5363-94c2-166425c3531d", 00:25:52.110 "is_configured": true, 00:25:52.110 "data_offset": 0, 00:25:52.110 "data_size": 65536 00:25:52.110 }, 00:25:52.110 { 00:25:52.110 "name": "BaseBdev4", 00:25:52.110 "uuid": "28274275-afa4-5f66-88c5-3c71427ba318", 00:25:52.110 "is_configured": true, 00:25:52.110 "data_offset": 0, 00:25:52.110 "data_size": 65536 00:25:52.110 } 00:25:52.110 ] 00:25:52.110 }' 00:25:52.110 11:36:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:25:52.110 11:36:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:25:52.674 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@718 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:25:52.931 [2024-07-15 11:36:36.371947] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:25:52.931 [2024-07-15 11:36:36.371985] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:25:52.931 00:25:52.931 Latency(us) 00:25:52.931 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:25:52.931 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:25:52.931 raid_bdev1 : 11.61 91.81 275.42 0.00 0.00 14684.70 293.84 120358.29 00:25:52.931 =================================================================================================================== 00:25:52.931 Total : 91.81 275.42 0.00 0.00 14684.70 293.84 120358.29 00:25:52.931 [2024-07-15 11:36:36.391900] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:52.931 [2024-07-15 11:36:36.391935] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:25:52.931 [2024-07-15 11:36:36.392034] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:25:52.931 [2024-07-15 11:36:36.392046] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x15598a0 name raid_bdev1, state offline 00:25:52.931 0 00:25:52.931 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:52.931 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # jq length 00:25:53.189 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # [[ 0 == 0 ]] 00:25:53.189 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@721 -- # '[' true = true ']' 00:25:53.189 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:25:53.189 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@724 -- # nbd_start_disks /var/tmp/spdk-raid.sock spare /dev/nbd0 00:25:53.189 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:25:53.189 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:25:53.189 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:25:53.189 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:25:53.189 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:25:53.189 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:25:53.189 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:25:53.189 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:25:53.189 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk spare /dev/nbd0 00:25:53.447 /dev/nbd0 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@867 -- # local i 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # break 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:25:53.447 1+0 records in 00:25:53.447 1+0 records out 00:25:53.447 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000267665 s, 15.3 MB/s 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # size=4096 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # return 0 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # for bdev in "${base_bdevs[@]:1}" 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # '[' -z '' ']' 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # continue 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # for bdev in "${base_bdevs[@]:1}" 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # '[' -z BaseBdev3 ']' 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@729 -- # nbd_start_disks /var/tmp/spdk-raid.sock BaseBdev3 /dev/nbd1 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:25:53.447 11:36:36 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:25:53.705 /dev/nbd1 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@867 -- # local i 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # break 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:25:53.705 1+0 records in 00:25:53.705 1+0 records out 00:25:53.705 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000297616 s, 13.8 MB/s 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # size=4096 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # return 0 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # nbd_stop_disks /var/tmp/spdk-raid.sock /dev/nbd1 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:25:53.705 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd1 00:25:53.962 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:25:53.962 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:25:53.962 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:25:53.962 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:25:53.962 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:25:53.962 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:25:53.962 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:25:53.962 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:25:53.962 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # for bdev in "${base_bdevs[@]:1}" 00:25:53.962 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # '[' -z BaseBdev4 ']' 00:25:53.962 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@729 -- # nbd_start_disks /var/tmp/spdk-raid.sock BaseBdev4 /dev/nbd1 00:25:53.962 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:25:53.962 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:25:53.962 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:25:53.962 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:25:53.962 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:25:53.962 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:25:53.962 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:25:53.962 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:25:53.963 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:25:54.220 /dev/nbd1 00:25:54.220 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:25:54.220 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:25:54.220 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:25:54.220 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@867 -- # local i 00:25:54.220 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:25:54.220 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:25:54.220 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:25:54.221 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # break 00:25:54.221 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:25:54.221 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:25:54.221 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:25:54.221 1+0 records in 00:25:54.221 1+0 records out 00:25:54.221 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000250902 s, 16.3 MB/s 00:25:54.221 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:25:54.221 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # size=4096 00:25:54.221 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:25:54.221 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:25:54.221 11:36:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # return 0 00:25:54.221 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:25:54.221 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:25:54.221 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:25:54.478 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # nbd_stop_disks /var/tmp/spdk-raid.sock /dev/nbd1 00:25:54.478 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:25:54.478 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:25:54.478 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:25:54.478 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:25:54.478 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:25:54.478 11:36:37 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd1 00:25:54.736 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:25:54.736 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:25:54.736 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:25:54.736 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:25:54.736 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:25:54.736 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:25:54.736 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:25:54.736 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:25:54.736 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@733 -- # nbd_stop_disks /var/tmp/spdk-raid.sock /dev/nbd0 00:25:54.736 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:25:54.736 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:25:54.736 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:25:54.736 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:25:54.736 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:25:54.736 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd0 00:25:54.994 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:25:54.994 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:25:54.994 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:25:54.994 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:25:54.994 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:25:54.994 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:25:54.994 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:25:54.994 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:25:54.994 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@742 -- # '[' false = true ']' 00:25:54.994 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@782 -- # killprocess 993073 00:25:54.994 11:36:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@948 -- # '[' -z 993073 ']' 00:25:54.994 11:36:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@952 -- # kill -0 993073 00:25:54.994 11:36:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@953 -- # uname 00:25:54.994 11:36:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:25:54.994 11:36:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 993073 00:25:54.994 11:36:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:25:54.994 11:36:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:25:54.994 11:36:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@966 -- # echo 'killing process with pid 993073' 00:25:54.994 killing process with pid 993073 00:25:54.994 11:36:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@967 -- # kill 993073 00:25:54.994 Received shutdown signal, test time was about 13.698183 seconds 00:25:54.994 00:25:54.994 Latency(us) 00:25:54.994 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:25:54.994 =================================================================================================================== 00:25:54.994 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:25:54.994 [2024-07-15 11:36:38.479815] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:25:54.994 11:36:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@972 -- # wait 993073 00:25:54.994 [2024-07-15 11:36:38.520385] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # return 0 00:25:55.252 00:25:55.252 real 0m18.791s 00:25:55.252 user 0m28.969s 00:25:55.252 sys 0m3.422s 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1124 -- # xtrace_disable 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:25:55.252 ************************************ 00:25:55.252 END TEST raid_rebuild_test_io 00:25:55.252 ************************************ 00:25:55.252 11:36:38 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:25:55.252 11:36:38 bdev_raid -- bdev/bdev_raid.sh@880 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 4 true true true 00:25:55.252 11:36:38 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:25:55.252 11:36:38 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:25:55.252 11:36:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:25:55.252 ************************************ 00:25:55.252 START TEST raid_rebuild_test_sb_io 00:25:55.252 ************************************ 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1123 -- # raid_rebuild_test raid1 4 true true true 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@568 -- # local raid_level=raid1 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local num_base_bdevs=4 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local superblock=true 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local background_io=true 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local verify=true 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # (( i = 1 )) 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # echo BaseBdev1 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # echo BaseBdev2 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # echo BaseBdev3 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # echo BaseBdev4 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:25:55.252 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:25:55.253 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:25:55.253 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local base_bdevs 00:25:55.253 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local raid_bdev_name=raid_bdev1 00:25:55.253 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local strip_size 00:25:55.253 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local create_arg 00:25:55.253 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local raid_bdev_size 00:25:55.253 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local data_offset 00:25:55.253 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@580 -- # '[' raid1 '!=' raid1 ']' 00:25:55.253 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@588 -- # strip_size=0 00:25:55.253 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@591 -- # '[' true = true ']' 00:25:55.253 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # create_arg+=' -s' 00:25:55.253 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # raid_pid=995796 00:25:55.253 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # waitforlisten 995796 /var/tmp/spdk-raid.sock 00:25:55.253 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@595 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:25:55.253 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@829 -- # '[' -z 995796 ']' 00:25:55.253 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:25:55.253 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@834 -- # local max_retries=100 00:25:55.253 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:25:55.253 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:25:55.253 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@838 -- # xtrace_disable 00:25:55.253 11:36:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:25:55.511 [2024-07-15 11:36:38.871842] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:25:55.511 [2024-07-15 11:36:38.871907] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid995796 ] 00:25:55.511 I/O size of 3145728 is greater than zero copy threshold (65536). 00:25:55.511 Zero copy mechanism will not be used. 00:25:55.511 [2024-07-15 11:36:38.998716] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:55.511 [2024-07-15 11:36:39.100493] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:25:55.769 [2024-07-15 11:36:39.171826] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:25:55.769 [2024-07-15 11:36:39.171864] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:25:56.335 11:36:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:25:56.335 11:36:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@862 -- # return 0 00:25:56.335 11:36:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:25:56.335 11:36:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:25:56.593 BaseBdev1_malloc 00:25:56.593 11:36:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:25:56.850 [2024-07-15 11:36:40.225887] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:25:56.850 [2024-07-15 11:36:40.225937] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:56.850 [2024-07-15 11:36:40.225962] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1be9d40 00:25:56.850 [2024-07-15 11:36:40.225975] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:56.850 [2024-07-15 11:36:40.227621] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:56.850 [2024-07-15 11:36:40.227649] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:25:56.850 BaseBdev1 00:25:56.850 11:36:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:25:56.850 11:36:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:25:57.108 BaseBdev2_malloc 00:25:57.108 11:36:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:25:57.396 [2024-07-15 11:36:40.732103] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:25:57.396 [2024-07-15 11:36:40.732154] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:57.396 [2024-07-15 11:36:40.732179] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1bea860 00:25:57.396 [2024-07-15 11:36:40.732192] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:57.396 [2024-07-15 11:36:40.733731] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:57.396 [2024-07-15 11:36:40.733758] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:25:57.396 BaseBdev2 00:25:57.396 11:36:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:25:57.397 11:36:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:25:57.700 BaseBdev3_malloc 00:25:57.700 11:36:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:25:57.700 [2024-07-15 11:36:41.218494] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:25:57.700 [2024-07-15 11:36:41.218541] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:57.700 [2024-07-15 11:36:41.218564] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1d978f0 00:25:57.700 [2024-07-15 11:36:41.218576] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:57.700 [2024-07-15 11:36:41.220175] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:57.700 [2024-07-15 11:36:41.220204] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:25:57.700 BaseBdev3 00:25:57.700 11:36:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:25:57.700 11:36:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:25:57.958 BaseBdev4_malloc 00:25:57.958 11:36:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:25:58.524 [2024-07-15 11:36:41.970332] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:25:58.524 [2024-07-15 11:36:41.970381] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:58.524 [2024-07-15 11:36:41.970403] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1d96ad0 00:25:58.524 [2024-07-15 11:36:41.970416] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:58.524 [2024-07-15 11:36:41.972005] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:58.524 [2024-07-15 11:36:41.972033] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:25:58.524 BaseBdev4 00:25:58.524 11:36:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@606 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 512 -b spare_malloc 00:25:58.782 spare_malloc 00:25:58.782 11:36:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:25:59.348 spare_delay 00:25:59.348 11:36:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:25:59.606 [2024-07-15 11:36:42.989572] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:25:59.606 [2024-07-15 11:36:42.989622] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:59.606 [2024-07-15 11:36:42.989644] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1d9b5b0 00:25:59.606 [2024-07-15 11:36:42.989657] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:59.606 [2024-07-15 11:36:42.991281] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:59.606 [2024-07-15 11:36:42.991311] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:25:59.606 spare 00:25:59.606 11:36:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@611 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4' -n raid_bdev1 00:25:59.864 [2024-07-15 11:36:43.234252] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:25:59.864 [2024-07-15 11:36:43.235606] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:25:59.864 [2024-07-15 11:36:43.235664] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:25:59.864 [2024-07-15 11:36:43.235711] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:25:59.864 [2024-07-15 11:36:43.235916] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1d1a8a0 00:25:59.864 [2024-07-15 11:36:43.235938] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:25:59.864 [2024-07-15 11:36:43.236142] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1d94e10 00:25:59.864 [2024-07-15 11:36:43.236295] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1d1a8a0 00:25:59.864 [2024-07-15 11:36:43.236305] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x1d1a8a0 00:25:59.864 [2024-07-15 11:36:43.236404] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:59.864 11:36:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:25:59.864 11:36:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:25:59.864 11:36:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:25:59.864 11:36:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:25:59.864 11:36:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:25:59.864 11:36:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=4 00:25:59.864 11:36:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:25:59.864 11:36:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:25:59.864 11:36:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:25:59.864 11:36:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:25:59.864 11:36:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:25:59.864 11:36:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:00.121 11:36:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:00.121 "name": "raid_bdev1", 00:26:00.121 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:00.121 "strip_size_kb": 0, 00:26:00.121 "state": "online", 00:26:00.121 "raid_level": "raid1", 00:26:00.121 "superblock": true, 00:26:00.121 "num_base_bdevs": 4, 00:26:00.121 "num_base_bdevs_discovered": 4, 00:26:00.121 "num_base_bdevs_operational": 4, 00:26:00.121 "base_bdevs_list": [ 00:26:00.121 { 00:26:00.121 "name": "BaseBdev1", 00:26:00.121 "uuid": "1b55e0a9-e124-5d2d-a97d-61ac4e8cd5fa", 00:26:00.121 "is_configured": true, 00:26:00.121 "data_offset": 2048, 00:26:00.121 "data_size": 63488 00:26:00.121 }, 00:26:00.121 { 00:26:00.121 "name": "BaseBdev2", 00:26:00.121 "uuid": "658ef952-38e6-505a-96bc-82364f281b42", 00:26:00.121 "is_configured": true, 00:26:00.121 "data_offset": 2048, 00:26:00.121 "data_size": 63488 00:26:00.121 }, 00:26:00.121 { 00:26:00.121 "name": "BaseBdev3", 00:26:00.121 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:00.121 "is_configured": true, 00:26:00.121 "data_offset": 2048, 00:26:00.121 "data_size": 63488 00:26:00.121 }, 00:26:00.121 { 00:26:00.121 "name": "BaseBdev4", 00:26:00.121 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:00.121 "is_configured": true, 00:26:00.121 "data_offset": 2048, 00:26:00.121 "data_size": 63488 00:26:00.121 } 00:26:00.121 ] 00:26:00.121 }' 00:26:00.121 11:36:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:00.121 11:36:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:26:01.054 11:36:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@615 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:26:01.054 11:36:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@615 -- # jq -r '.[].num_blocks' 00:26:01.054 [2024-07-15 11:36:44.574072] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:01.054 11:36:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@615 -- # raid_bdev_size=63488 00:26:01.055 11:36:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@618 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:01.055 11:36:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@618 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:26:01.320 11:36:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@618 -- # data_offset=2048 00:26:01.320 11:36:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@620 -- # '[' true = true ']' 00:26:01.320 11:36:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@639 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev1 00:26:01.320 11:36:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@622 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/spdk-raid.sock perform_tests 00:26:01.578 [2024-07-15 11:36:44.948865] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1be9670 00:26:01.578 I/O size of 3145728 is greater than zero copy threshold (65536). 00:26:01.578 Zero copy mechanism will not be used. 00:26:01.578 Running I/O for 60 seconds... 00:26:01.578 [2024-07-15 11:36:45.070501] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:26:01.578 [2024-07-15 11:36:45.070731] bdev_raid.c:1919:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x1be9670 00:26:01.578 11:36:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@642 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:26:01.578 11:36:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:26:01.578 11:36:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:26:01.578 11:36:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:01.578 11:36:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:01.578 11:36:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:26:01.578 11:36:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:01.578 11:36:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:01.578 11:36:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:01.578 11:36:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:01.578 11:36:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:01.578 11:36:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:01.837 11:36:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:01.837 "name": "raid_bdev1", 00:26:01.837 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:01.837 "strip_size_kb": 0, 00:26:01.837 "state": "online", 00:26:01.837 "raid_level": "raid1", 00:26:01.837 "superblock": true, 00:26:01.837 "num_base_bdevs": 4, 00:26:01.837 "num_base_bdevs_discovered": 3, 00:26:01.837 "num_base_bdevs_operational": 3, 00:26:01.837 "base_bdevs_list": [ 00:26:01.837 { 00:26:01.837 "name": null, 00:26:01.837 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:01.837 "is_configured": false, 00:26:01.837 "data_offset": 2048, 00:26:01.837 "data_size": 63488 00:26:01.837 }, 00:26:01.837 { 00:26:01.837 "name": "BaseBdev2", 00:26:01.837 "uuid": "658ef952-38e6-505a-96bc-82364f281b42", 00:26:01.837 "is_configured": true, 00:26:01.837 "data_offset": 2048, 00:26:01.837 "data_size": 63488 00:26:01.837 }, 00:26:01.837 { 00:26:01.837 "name": "BaseBdev3", 00:26:01.837 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:01.837 "is_configured": true, 00:26:01.837 "data_offset": 2048, 00:26:01.837 "data_size": 63488 00:26:01.837 }, 00:26:01.837 { 00:26:01.837 "name": "BaseBdev4", 00:26:01.837 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:01.837 "is_configured": true, 00:26:01.837 "data_offset": 2048, 00:26:01.837 "data_size": 63488 00:26:01.837 } 00:26:01.837 ] 00:26:01.837 }' 00:26:01.837 11:36:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:01.837 11:36:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:26:02.402 11:36:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@645 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:26:02.659 [2024-07-15 11:36:46.210545] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:26:02.917 11:36:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # sleep 1 00:26:02.917 [2024-07-15 11:36:46.293495] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1d1cb40 00:26:02.917 [2024-07-15 11:36:46.295876] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:26:02.917 [2024-07-15 11:36:46.436822] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:26:03.176 [2024-07-15 11:36:46.688604] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:26:03.176 [2024-07-15 11:36:46.689251] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:26:03.742 [2024-07-15 11:36:47.038280] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:26:03.742 [2024-07-15 11:36:47.038598] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:26:03.742 [2024-07-15 11:36:47.269071] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:26:03.742 [2024-07-15 11:36:47.269760] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:26:03.742 11:36:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@649 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:03.742 11:36:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:26:03.742 11:36:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:26:03.742 11:36:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:26:03.742 11:36:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:26:03.742 11:36:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:03.742 11:36:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:04.000 11:36:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:26:04.000 "name": "raid_bdev1", 00:26:04.000 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:04.000 "strip_size_kb": 0, 00:26:04.000 "state": "online", 00:26:04.000 "raid_level": "raid1", 00:26:04.000 "superblock": true, 00:26:04.000 "num_base_bdevs": 4, 00:26:04.000 "num_base_bdevs_discovered": 4, 00:26:04.000 "num_base_bdevs_operational": 4, 00:26:04.000 "process": { 00:26:04.000 "type": "rebuild", 00:26:04.000 "target": "spare", 00:26:04.000 "progress": { 00:26:04.000 "blocks": 12288, 00:26:04.000 "percent": 19 00:26:04.000 } 00:26:04.000 }, 00:26:04.000 "base_bdevs_list": [ 00:26:04.000 { 00:26:04.000 "name": "spare", 00:26:04.000 "uuid": "30d0a2d4-dd37-5305-a3bc-53b7aed38ff6", 00:26:04.000 "is_configured": true, 00:26:04.000 "data_offset": 2048, 00:26:04.000 "data_size": 63488 00:26:04.000 }, 00:26:04.000 { 00:26:04.000 "name": "BaseBdev2", 00:26:04.000 "uuid": "658ef952-38e6-505a-96bc-82364f281b42", 00:26:04.000 "is_configured": true, 00:26:04.000 "data_offset": 2048, 00:26:04.000 "data_size": 63488 00:26:04.000 }, 00:26:04.000 { 00:26:04.000 "name": "BaseBdev3", 00:26:04.000 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:04.000 "is_configured": true, 00:26:04.000 "data_offset": 2048, 00:26:04.000 "data_size": 63488 00:26:04.000 }, 00:26:04.000 { 00:26:04.000 "name": "BaseBdev4", 00:26:04.000 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:04.000 "is_configured": true, 00:26:04.000 "data_offset": 2048, 00:26:04.000 "data_size": 63488 00:26:04.000 } 00:26:04.000 ] 00:26:04.000 }' 00:26:04.000 11:36:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:26:04.000 11:36:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:04.000 11:36:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:26:04.258 11:36:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:26:04.258 11:36:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@652 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev spare 00:26:04.258 [2024-07-15 11:36:47.613872] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:26:04.258 [2024-07-15 11:36:47.614399] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:26:04.258 [2024-07-15 11:36:47.829731] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:26:04.258 [2024-07-15 11:36:47.830022] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:26:04.258 [2024-07-15 11:36:47.840660] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:26:04.517 [2024-07-15 11:36:47.878453] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:26:04.517 [2024-07-15 11:36:47.989003] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:26:04.517 [2024-07-15 11:36:48.009649] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:04.517 [2024-07-15 11:36:48.009681] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:26:04.517 [2024-07-15 11:36:48.009693] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:26:04.517 [2024-07-15 11:36:48.032343] bdev_raid.c:1919:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x1be9670 00:26:04.517 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@655 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:26:04.517 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:26:04.517 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:26:04.517 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:04.517 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:04.517 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:26:04.517 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:04.517 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:04.517 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:04.517 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:04.517 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:04.517 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:04.774 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:04.774 "name": "raid_bdev1", 00:26:04.774 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:04.774 "strip_size_kb": 0, 00:26:04.774 "state": "online", 00:26:04.774 "raid_level": "raid1", 00:26:04.774 "superblock": true, 00:26:04.774 "num_base_bdevs": 4, 00:26:04.774 "num_base_bdevs_discovered": 3, 00:26:04.774 "num_base_bdevs_operational": 3, 00:26:04.774 "base_bdevs_list": [ 00:26:04.774 { 00:26:04.774 "name": null, 00:26:04.774 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:04.774 "is_configured": false, 00:26:04.774 "data_offset": 2048, 00:26:04.774 "data_size": 63488 00:26:04.774 }, 00:26:04.774 { 00:26:04.774 "name": "BaseBdev2", 00:26:04.774 "uuid": "658ef952-38e6-505a-96bc-82364f281b42", 00:26:04.774 "is_configured": true, 00:26:04.774 "data_offset": 2048, 00:26:04.774 "data_size": 63488 00:26:04.774 }, 00:26:04.774 { 00:26:04.774 "name": "BaseBdev3", 00:26:04.774 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:04.774 "is_configured": true, 00:26:04.774 "data_offset": 2048, 00:26:04.774 "data_size": 63488 00:26:04.774 }, 00:26:04.774 { 00:26:04.774 "name": "BaseBdev4", 00:26:04.774 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:04.774 "is_configured": true, 00:26:04.774 "data_offset": 2048, 00:26:04.774 "data_size": 63488 00:26:04.774 } 00:26:04.774 ] 00:26:04.774 }' 00:26:04.774 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:04.774 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:26:05.708 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@658 -- # verify_raid_bdev_process raid_bdev1 none none 00:26:05.708 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:26:05.708 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:26:05.708 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=none 00:26:05.708 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:26:05.708 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:05.708 11:36:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:05.708 11:36:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:26:05.708 "name": "raid_bdev1", 00:26:05.708 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:05.708 "strip_size_kb": 0, 00:26:05.708 "state": "online", 00:26:05.708 "raid_level": "raid1", 00:26:05.708 "superblock": true, 00:26:05.708 "num_base_bdevs": 4, 00:26:05.708 "num_base_bdevs_discovered": 3, 00:26:05.708 "num_base_bdevs_operational": 3, 00:26:05.708 "base_bdevs_list": [ 00:26:05.708 { 00:26:05.708 "name": null, 00:26:05.708 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:05.708 "is_configured": false, 00:26:05.708 "data_offset": 2048, 00:26:05.708 "data_size": 63488 00:26:05.708 }, 00:26:05.708 { 00:26:05.708 "name": "BaseBdev2", 00:26:05.708 "uuid": "658ef952-38e6-505a-96bc-82364f281b42", 00:26:05.708 "is_configured": true, 00:26:05.708 "data_offset": 2048, 00:26:05.708 "data_size": 63488 00:26:05.708 }, 00:26:05.708 { 00:26:05.708 "name": "BaseBdev3", 00:26:05.708 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:05.708 "is_configured": true, 00:26:05.708 "data_offset": 2048, 00:26:05.708 "data_size": 63488 00:26:05.708 }, 00:26:05.708 { 00:26:05.708 "name": "BaseBdev4", 00:26:05.708 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:05.708 "is_configured": true, 00:26:05.708 "data_offset": 2048, 00:26:05.708 "data_size": 63488 00:26:05.708 } 00:26:05.708 ] 00:26:05.708 }' 00:26:05.708 11:36:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:26:05.708 11:36:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:26:05.708 11:36:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:26:05.966 11:36:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:26:05.966 11:36:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@661 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:26:05.966 [2024-07-15 11:36:49.549072] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:26:06.224 11:36:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # sleep 1 00:26:06.224 [2024-07-15 11:36:49.624389] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1d1e8f0 00:26:06.224 [2024-07-15 11:36:49.625898] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:26:06.224 [2024-07-15 11:36:49.764872] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:26:06.224 [2024-07-15 11:36:49.765210] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:26:06.483 [2024-07-15 11:36:49.934612] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:26:06.740 [2024-07-15 11:36:50.291855] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:26:06.998 [2024-07-15 11:36:50.524223] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:26:07.257 11:36:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:07.257 11:36:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:26:07.257 11:36:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:26:07.257 11:36:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:26:07.257 11:36:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:26:07.257 11:36:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:07.257 11:36:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:07.516 11:36:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:26:07.516 "name": "raid_bdev1", 00:26:07.516 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:07.516 "strip_size_kb": 0, 00:26:07.516 "state": "online", 00:26:07.516 "raid_level": "raid1", 00:26:07.516 "superblock": true, 00:26:07.516 "num_base_bdevs": 4, 00:26:07.516 "num_base_bdevs_discovered": 4, 00:26:07.516 "num_base_bdevs_operational": 4, 00:26:07.516 "process": { 00:26:07.516 "type": "rebuild", 00:26:07.516 "target": "spare", 00:26:07.516 "progress": { 00:26:07.516 "blocks": 12288, 00:26:07.516 "percent": 19 00:26:07.516 } 00:26:07.516 }, 00:26:07.516 "base_bdevs_list": [ 00:26:07.516 { 00:26:07.516 "name": "spare", 00:26:07.516 "uuid": "30d0a2d4-dd37-5305-a3bc-53b7aed38ff6", 00:26:07.516 "is_configured": true, 00:26:07.516 "data_offset": 2048, 00:26:07.516 "data_size": 63488 00:26:07.516 }, 00:26:07.516 { 00:26:07.516 "name": "BaseBdev2", 00:26:07.516 "uuid": "658ef952-38e6-505a-96bc-82364f281b42", 00:26:07.516 "is_configured": true, 00:26:07.516 "data_offset": 2048, 00:26:07.516 "data_size": 63488 00:26:07.516 }, 00:26:07.516 { 00:26:07.516 "name": "BaseBdev3", 00:26:07.516 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:07.516 "is_configured": true, 00:26:07.516 "data_offset": 2048, 00:26:07.516 "data_size": 63488 00:26:07.516 }, 00:26:07.516 { 00:26:07.516 "name": "BaseBdev4", 00:26:07.516 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:07.516 "is_configured": true, 00:26:07.516 "data_offset": 2048, 00:26:07.516 "data_size": 63488 00:26:07.516 } 00:26:07.516 ] 00:26:07.516 }' 00:26:07.516 11:36:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:26:07.516 [2024-07-15 11:36:50.885434] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:26:07.516 11:36:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:07.516 11:36:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:26:07.516 11:36:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:26:07.516 11:36:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@665 -- # '[' true = true ']' 00:26:07.516 11:36:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@665 -- # '[' = false ']' 00:26:07.516 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev_raid.sh: line 665: [: =: unary operator expected 00:26:07.516 11:36:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@690 -- # local num_base_bdevs_operational=4 00:26:07.516 11:36:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@692 -- # '[' raid1 = raid1 ']' 00:26:07.516 11:36:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@692 -- # '[' 4 -gt 2 ']' 00:26:07.516 11:36:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@694 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev2 00:26:07.773 [2024-07-15 11:36:51.118828] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:26:07.773 [2024-07-15 11:36:51.207618] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:26:08.030 [2024-07-15 11:36:51.571277] bdev_raid.c:1919:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x1be9670 00:26:08.030 [2024-07-15 11:36:51.571312] bdev_raid.c:1919:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x1d1e8f0 00:26:08.030 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@697 -- # base_bdevs[1]= 00:26:08.030 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@698 -- # (( num_base_bdevs_operational-- )) 00:26:08.030 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@701 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:08.030 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:26:08.030 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:26:08.030 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:26:08.030 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:26:08.030 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:08.030 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:08.287 [2024-07-15 11:36:51.710759] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:26:08.287 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:26:08.287 "name": "raid_bdev1", 00:26:08.287 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:08.287 "strip_size_kb": 0, 00:26:08.287 "state": "online", 00:26:08.287 "raid_level": "raid1", 00:26:08.287 "superblock": true, 00:26:08.287 "num_base_bdevs": 4, 00:26:08.287 "num_base_bdevs_discovered": 3, 00:26:08.287 "num_base_bdevs_operational": 3, 00:26:08.287 "process": { 00:26:08.287 "type": "rebuild", 00:26:08.287 "target": "spare", 00:26:08.287 "progress": { 00:26:08.287 "blocks": 20480, 00:26:08.287 "percent": 32 00:26:08.287 } 00:26:08.287 }, 00:26:08.287 "base_bdevs_list": [ 00:26:08.287 { 00:26:08.287 "name": "spare", 00:26:08.287 "uuid": "30d0a2d4-dd37-5305-a3bc-53b7aed38ff6", 00:26:08.287 "is_configured": true, 00:26:08.287 "data_offset": 2048, 00:26:08.287 "data_size": 63488 00:26:08.287 }, 00:26:08.287 { 00:26:08.287 "name": null, 00:26:08.287 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:08.287 "is_configured": false, 00:26:08.287 "data_offset": 2048, 00:26:08.287 "data_size": 63488 00:26:08.287 }, 00:26:08.287 { 00:26:08.287 "name": "BaseBdev3", 00:26:08.287 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:08.287 "is_configured": true, 00:26:08.287 "data_offset": 2048, 00:26:08.287 "data_size": 63488 00:26:08.287 }, 00:26:08.287 { 00:26:08.287 "name": "BaseBdev4", 00:26:08.287 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:08.287 "is_configured": true, 00:26:08.287 "data_offset": 2048, 00:26:08.287 "data_size": 63488 00:26:08.287 } 00:26:08.287 ] 00:26:08.287 }' 00:26:08.287 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:26:08.287 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:08.287 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:26:08.544 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:26:08.544 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@705 -- # local timeout=939 00:26:08.544 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:26:08.544 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:08.544 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:26:08.544 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:26:08.544 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:26:08.544 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:26:08.544 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:08.544 11:36:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:08.544 [2024-07-15 11:36:51.949994] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:26:08.801 [2024-07-15 11:36:52.179572] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:26:08.801 11:36:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:26:08.801 "name": "raid_bdev1", 00:26:08.801 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:08.801 "strip_size_kb": 0, 00:26:08.801 "state": "online", 00:26:08.801 "raid_level": "raid1", 00:26:08.801 "superblock": true, 00:26:08.801 "num_base_bdevs": 4, 00:26:08.801 "num_base_bdevs_discovered": 3, 00:26:08.801 "num_base_bdevs_operational": 3, 00:26:08.801 "process": { 00:26:08.801 "type": "rebuild", 00:26:08.801 "target": "spare", 00:26:08.801 "progress": { 00:26:08.801 "blocks": 24576, 00:26:08.801 "percent": 38 00:26:08.801 } 00:26:08.801 }, 00:26:08.801 "base_bdevs_list": [ 00:26:08.801 { 00:26:08.801 "name": "spare", 00:26:08.801 "uuid": "30d0a2d4-dd37-5305-a3bc-53b7aed38ff6", 00:26:08.801 "is_configured": true, 00:26:08.801 "data_offset": 2048, 00:26:08.801 "data_size": 63488 00:26:08.801 }, 00:26:08.801 { 00:26:08.801 "name": null, 00:26:08.801 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:08.801 "is_configured": false, 00:26:08.801 "data_offset": 2048, 00:26:08.801 "data_size": 63488 00:26:08.801 }, 00:26:08.801 { 00:26:08.801 "name": "BaseBdev3", 00:26:08.801 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:08.801 "is_configured": true, 00:26:08.801 "data_offset": 2048, 00:26:08.801 "data_size": 63488 00:26:08.801 }, 00:26:08.801 { 00:26:08.801 "name": "BaseBdev4", 00:26:08.801 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:08.801 "is_configured": true, 00:26:08.801 "data_offset": 2048, 00:26:08.801 "data_size": 63488 00:26:08.801 } 00:26:08.801 ] 00:26:08.801 }' 00:26:08.801 11:36:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:26:08.801 11:36:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:08.801 11:36:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:26:08.801 11:36:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:26:08.801 11:36:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@710 -- # sleep 1 00:26:09.058 [2024-07-15 11:36:52.420913] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:26:09.375 [2024-07-15 11:36:52.754981] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:26:09.633 [2024-07-15 11:36:52.977237] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:26:09.889 [2024-07-15 11:36:53.246523] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:26:09.889 11:36:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:26:09.889 11:36:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:09.889 11:36:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:26:09.889 11:36:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:26:09.889 11:36:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:26:09.889 11:36:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:26:09.889 11:36:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:09.889 11:36:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:10.147 11:36:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:26:10.147 "name": "raid_bdev1", 00:26:10.147 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:10.147 "strip_size_kb": 0, 00:26:10.147 "state": "online", 00:26:10.147 "raid_level": "raid1", 00:26:10.147 "superblock": true, 00:26:10.147 "num_base_bdevs": 4, 00:26:10.147 "num_base_bdevs_discovered": 3, 00:26:10.147 "num_base_bdevs_operational": 3, 00:26:10.147 "process": { 00:26:10.147 "type": "rebuild", 00:26:10.147 "target": "spare", 00:26:10.147 "progress": { 00:26:10.147 "blocks": 43008, 00:26:10.147 "percent": 67 00:26:10.147 } 00:26:10.147 }, 00:26:10.147 "base_bdevs_list": [ 00:26:10.147 { 00:26:10.147 "name": "spare", 00:26:10.147 "uuid": "30d0a2d4-dd37-5305-a3bc-53b7aed38ff6", 00:26:10.147 "is_configured": true, 00:26:10.147 "data_offset": 2048, 00:26:10.147 "data_size": 63488 00:26:10.147 }, 00:26:10.147 { 00:26:10.147 "name": null, 00:26:10.147 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:10.147 "is_configured": false, 00:26:10.147 "data_offset": 2048, 00:26:10.147 "data_size": 63488 00:26:10.147 }, 00:26:10.147 { 00:26:10.147 "name": "BaseBdev3", 00:26:10.147 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:10.147 "is_configured": true, 00:26:10.147 "data_offset": 2048, 00:26:10.147 "data_size": 63488 00:26:10.147 }, 00:26:10.147 { 00:26:10.147 "name": "BaseBdev4", 00:26:10.147 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:10.147 "is_configured": true, 00:26:10.147 "data_offset": 2048, 00:26:10.147 "data_size": 63488 00:26:10.147 } 00:26:10.147 ] 00:26:10.147 }' 00:26:10.147 11:36:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:26:10.147 11:36:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:10.147 11:36:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:26:10.147 11:36:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:26:10.147 11:36:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@710 -- # sleep 1 00:26:11.078 [2024-07-15 11:36:54.398338] bdev_raid.c: 839:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:26:11.078 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:26:11.078 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:11.078 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:26:11.078 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:26:11.078 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:26:11.078 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:26:11.078 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:11.078 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:11.336 [2024-07-15 11:36:54.729779] bdev_raid.c:2789:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:26:11.336 [2024-07-15 11:36:54.830037] bdev_raid.c:2504:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:26:11.336 [2024-07-15 11:36:54.840897] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:11.336 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:26:11.336 "name": "raid_bdev1", 00:26:11.336 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:11.336 "strip_size_kb": 0, 00:26:11.336 "state": "online", 00:26:11.336 "raid_level": "raid1", 00:26:11.336 "superblock": true, 00:26:11.336 "num_base_bdevs": 4, 00:26:11.336 "num_base_bdevs_discovered": 3, 00:26:11.336 "num_base_bdevs_operational": 3, 00:26:11.336 "base_bdevs_list": [ 00:26:11.336 { 00:26:11.336 "name": "spare", 00:26:11.336 "uuid": "30d0a2d4-dd37-5305-a3bc-53b7aed38ff6", 00:26:11.336 "is_configured": true, 00:26:11.336 "data_offset": 2048, 00:26:11.336 "data_size": 63488 00:26:11.336 }, 00:26:11.336 { 00:26:11.336 "name": null, 00:26:11.336 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:11.336 "is_configured": false, 00:26:11.336 "data_offset": 2048, 00:26:11.336 "data_size": 63488 00:26:11.336 }, 00:26:11.336 { 00:26:11.336 "name": "BaseBdev3", 00:26:11.336 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:11.336 "is_configured": true, 00:26:11.336 "data_offset": 2048, 00:26:11.336 "data_size": 63488 00:26:11.336 }, 00:26:11.336 { 00:26:11.336 "name": "BaseBdev4", 00:26:11.337 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:11.337 "is_configured": true, 00:26:11.337 "data_offset": 2048, 00:26:11.337 "data_size": 63488 00:26:11.337 } 00:26:11.337 ] 00:26:11.337 }' 00:26:11.337 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:26:11.337 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ none == \r\e\b\u\i\l\d ]] 00:26:11.337 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:26:11.595 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ none == \s\p\a\r\e ]] 00:26:11.595 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # break 00:26:11.595 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@714 -- # verify_raid_bdev_process raid_bdev1 none none 00:26:11.595 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:26:11.595 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:26:11.595 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=none 00:26:11.595 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:26:11.595 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:11.595 11:36:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:11.853 11:36:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:26:11.853 "name": "raid_bdev1", 00:26:11.853 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:11.853 "strip_size_kb": 0, 00:26:11.853 "state": "online", 00:26:11.853 "raid_level": "raid1", 00:26:11.853 "superblock": true, 00:26:11.853 "num_base_bdevs": 4, 00:26:11.853 "num_base_bdevs_discovered": 3, 00:26:11.853 "num_base_bdevs_operational": 3, 00:26:11.853 "base_bdevs_list": [ 00:26:11.853 { 00:26:11.853 "name": "spare", 00:26:11.853 "uuid": "30d0a2d4-dd37-5305-a3bc-53b7aed38ff6", 00:26:11.853 "is_configured": true, 00:26:11.853 "data_offset": 2048, 00:26:11.853 "data_size": 63488 00:26:11.853 }, 00:26:11.853 { 00:26:11.853 "name": null, 00:26:11.853 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:11.853 "is_configured": false, 00:26:11.853 "data_offset": 2048, 00:26:11.853 "data_size": 63488 00:26:11.853 }, 00:26:11.853 { 00:26:11.853 "name": "BaseBdev3", 00:26:11.853 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:11.853 "is_configured": true, 00:26:11.853 "data_offset": 2048, 00:26:11.853 "data_size": 63488 00:26:11.853 }, 00:26:11.853 { 00:26:11.853 "name": "BaseBdev4", 00:26:11.853 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:11.853 "is_configured": true, 00:26:11.853 "data_offset": 2048, 00:26:11.853 "data_size": 63488 00:26:11.853 } 00:26:11.853 ] 00:26:11.853 }' 00:26:11.853 11:36:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:26:11.853 11:36:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:26:11.853 11:36:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:26:11.853 11:36:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:26:11.853 11:36:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:26:11.853 11:36:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:26:11.853 11:36:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:26:11.853 11:36:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:11.853 11:36:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:11.853 11:36:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:26:11.853 11:36:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:11.853 11:36:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:11.853 11:36:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:11.853 11:36:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:11.853 11:36:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:11.853 11:36:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:12.111 11:36:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:12.111 "name": "raid_bdev1", 00:26:12.111 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:12.111 "strip_size_kb": 0, 00:26:12.111 "state": "online", 00:26:12.111 "raid_level": "raid1", 00:26:12.111 "superblock": true, 00:26:12.111 "num_base_bdevs": 4, 00:26:12.111 "num_base_bdevs_discovered": 3, 00:26:12.111 "num_base_bdevs_operational": 3, 00:26:12.111 "base_bdevs_list": [ 00:26:12.111 { 00:26:12.111 "name": "spare", 00:26:12.111 "uuid": "30d0a2d4-dd37-5305-a3bc-53b7aed38ff6", 00:26:12.111 "is_configured": true, 00:26:12.111 "data_offset": 2048, 00:26:12.111 "data_size": 63488 00:26:12.111 }, 00:26:12.111 { 00:26:12.111 "name": null, 00:26:12.111 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:12.111 "is_configured": false, 00:26:12.111 "data_offset": 2048, 00:26:12.111 "data_size": 63488 00:26:12.111 }, 00:26:12.111 { 00:26:12.111 "name": "BaseBdev3", 00:26:12.111 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:12.111 "is_configured": true, 00:26:12.111 "data_offset": 2048, 00:26:12.111 "data_size": 63488 00:26:12.111 }, 00:26:12.111 { 00:26:12.111 "name": "BaseBdev4", 00:26:12.111 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:12.111 "is_configured": true, 00:26:12.111 "data_offset": 2048, 00:26:12.111 "data_size": 63488 00:26:12.111 } 00:26:12.111 ] 00:26:12.111 }' 00:26:12.111 11:36:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:12.111 11:36:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:26:12.716 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@718 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:26:12.974 [2024-07-15 11:36:56.361522] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:26:12.974 [2024-07-15 11:36:56.361555] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:26:12.974 00:26:12.974 Latency(us) 00:26:12.974 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:12.974 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:26:12.974 raid_bdev1 : 11.42 93.94 281.82 0.00 0.00 14248.54 297.41 122181.90 00:26:12.974 =================================================================================================================== 00:26:12.974 Total : 93.94 281.82 0.00 0.00 14248.54 297.41 122181.90 00:26:12.974 [2024-07-15 11:36:56.405573] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:12.974 [2024-07-15 11:36:56.405602] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:12.974 [2024-07-15 11:36:56.405695] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:12.974 [2024-07-15 11:36:56.405707] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1d1a8a0 name raid_bdev1, state offline 00:26:12.974 0 00:26:12.975 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:12.975 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # jq length 00:26:13.233 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # [[ 0 == 0 ]] 00:26:13.233 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@721 -- # '[' true = true ']' 00:26:13.233 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:26:13.233 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@724 -- # nbd_start_disks /var/tmp/spdk-raid.sock spare /dev/nbd0 00:26:13.233 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:26:13.233 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:26:13.233 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:26:13.233 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:26:13.233 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:26:13.233 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:26:13.233 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:26:13.233 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:26:13.233 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk spare /dev/nbd0 00:26:13.491 /dev/nbd0 00:26:13.491 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:26:13.491 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:26:13.491 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:26:13.491 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@867 -- # local i 00:26:13.491 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:26:13.491 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:26:13.491 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:26:13.491 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # break 00:26:13.491 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:26:13.491 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:26:13.491 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:26:13.491 1+0 records in 00:26:13.491 1+0 records out 00:26:13.491 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000263147 s, 15.6 MB/s 00:26:13.491 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:26:13.491 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # size=4096 00:26:13.491 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:26:13.491 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:26:13.491 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # return 0 00:26:13.491 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:26:13.491 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:26:13.491 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # for bdev in "${base_bdevs[@]:1}" 00:26:13.492 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # '[' -z '' ']' 00:26:13.492 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # continue 00:26:13.492 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # for bdev in "${base_bdevs[@]:1}" 00:26:13.492 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # '[' -z BaseBdev3 ']' 00:26:13.492 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@729 -- # nbd_start_disks /var/tmp/spdk-raid.sock BaseBdev3 /dev/nbd1 00:26:13.492 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:26:13.492 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:26:13.492 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:26:13.492 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:26:13.492 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:26:13.492 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:26:13.492 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:26:13.492 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:26:13.492 11:36:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:26:13.750 /dev/nbd1 00:26:13.750 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:26:13.750 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:26:13.750 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:26:13.750 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@867 -- # local i 00:26:13.750 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:26:13.750 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:26:13.750 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:26:13.750 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # break 00:26:13.750 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:26:13.750 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:26:13.750 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:26:13.750 1+0 records in 00:26:13.750 1+0 records out 00:26:13.750 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000267852 s, 15.3 MB/s 00:26:13.750 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:26:13.750 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # size=4096 00:26:13.750 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:26:13.750 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:26:13.750 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # return 0 00:26:13.751 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:26:13.751 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:26:13.751 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:26:13.751 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # nbd_stop_disks /var/tmp/spdk-raid.sock /dev/nbd1 00:26:13.751 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:26:13.751 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:26:13.751 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:26:13.751 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:26:13.751 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:26:13.751 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd1 00:26:14.034 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:26:14.034 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:26:14.034 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:26:14.034 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:26:14.034 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:26:14.034 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:26:14.034 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@39 -- # sleep 0.1 00:26:14.293 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i++ )) 00:26:14.293 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:26:14.293 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:26:14.293 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:26:14.293 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:26:14.293 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # for bdev in "${base_bdevs[@]:1}" 00:26:14.293 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # '[' -z BaseBdev4 ']' 00:26:14.293 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@729 -- # nbd_start_disks /var/tmp/spdk-raid.sock BaseBdev4 /dev/nbd1 00:26:14.293 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:26:14.293 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:26:14.293 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:26:14.293 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:26:14.293 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:26:14.293 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:26:14.293 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:26:14.293 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:26:14.293 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:26:14.551 /dev/nbd1 00:26:14.551 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:26:14.551 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:26:14.551 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:26:14.551 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@867 -- # local i 00:26:14.551 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:26:14.551 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:26:14.551 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:26:14.551 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # break 00:26:14.551 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:26:14.551 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:26:14.551 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:26:14.551 1+0 records in 00:26:14.551 1+0 records out 00:26:14.551 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000212182 s, 19.3 MB/s 00:26:14.551 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:26:14.551 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # size=4096 00:26:14.551 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:26:14.551 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:26:14.551 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # return 0 00:26:14.551 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:26:14.551 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:26:14.551 11:36:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:26:14.551 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # nbd_stop_disks /var/tmp/spdk-raid.sock /dev/nbd1 00:26:14.551 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:26:14.551 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:26:14.551 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:26:14.551 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:26:14.551 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:26:14.551 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd1 00:26:14.810 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:26:14.810 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:26:14.810 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:26:14.810 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:26:14.810 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:26:14.810 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:26:14.810 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:26:14.810 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:26:14.810 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@733 -- # nbd_stop_disks /var/tmp/spdk-raid.sock /dev/nbd0 00:26:14.810 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:26:14.810 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:26:14.810 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:26:14.810 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:26:14.810 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:26:14.810 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd0 00:26:15.068 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:26:15.068 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:26:15.068 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:26:15.068 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:26:15.068 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:26:15.068 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:26:15.068 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:26:15.068 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:26:15.068 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@742 -- # '[' true = true ']' 00:26:15.068 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@744 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:26:15.326 11:36:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:26:15.584 [2024-07-15 11:36:59.053414] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:26:15.584 [2024-07-15 11:36:59.053462] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:15.584 [2024-07-15 11:36:59.053485] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1d1b210 00:26:15.584 [2024-07-15 11:36:59.053499] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:15.584 [2024-07-15 11:36:59.055122] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:15.584 [2024-07-15 11:36:59.055151] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:26:15.584 [2024-07-15 11:36:59.055230] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:26:15.584 [2024-07-15 11:36:59.055257] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:26:15.584 [2024-07-15 11:36:59.055361] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:26:15.585 [2024-07-15 11:36:59.055435] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:26:15.585 spare 00:26:15.585 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:26:15.585 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:26:15.585 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:26:15.585 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:15.585 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:15.585 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=3 00:26:15.585 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:15.585 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:15.585 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:15.585 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:15.585 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:15.585 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:15.585 [2024-07-15 11:36:59.155753] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1d1bf20 00:26:15.585 [2024-07-15 11:36:59.155773] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:26:15.585 [2024-07-15 11:36:59.155983] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1d1ba80 00:26:15.585 [2024-07-15 11:36:59.156137] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1d1bf20 00:26:15.585 [2024-07-15 11:36:59.156147] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x1d1bf20 00:26:15.585 [2024-07-15 11:36:59.156257] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:15.843 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:15.843 "name": "raid_bdev1", 00:26:15.843 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:15.843 "strip_size_kb": 0, 00:26:15.843 "state": "online", 00:26:15.843 "raid_level": "raid1", 00:26:15.843 "superblock": true, 00:26:15.843 "num_base_bdevs": 4, 00:26:15.843 "num_base_bdevs_discovered": 3, 00:26:15.843 "num_base_bdevs_operational": 3, 00:26:15.843 "base_bdevs_list": [ 00:26:15.843 { 00:26:15.843 "name": "spare", 00:26:15.843 "uuid": "30d0a2d4-dd37-5305-a3bc-53b7aed38ff6", 00:26:15.843 "is_configured": true, 00:26:15.843 "data_offset": 2048, 00:26:15.843 "data_size": 63488 00:26:15.843 }, 00:26:15.843 { 00:26:15.843 "name": null, 00:26:15.843 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:15.843 "is_configured": false, 00:26:15.843 "data_offset": 2048, 00:26:15.843 "data_size": 63488 00:26:15.843 }, 00:26:15.843 { 00:26:15.843 "name": "BaseBdev3", 00:26:15.843 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:15.843 "is_configured": true, 00:26:15.843 "data_offset": 2048, 00:26:15.843 "data_size": 63488 00:26:15.843 }, 00:26:15.843 { 00:26:15.843 "name": "BaseBdev4", 00:26:15.843 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:15.843 "is_configured": true, 00:26:15.843 "data_offset": 2048, 00:26:15.843 "data_size": 63488 00:26:15.843 } 00:26:15.843 ] 00:26:15.843 }' 00:26:15.843 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:15.843 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:26:16.409 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@748 -- # verify_raid_bdev_process raid_bdev1 none none 00:26:16.409 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:26:16.409 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:26:16.409 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=none 00:26:16.409 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:26:16.409 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:16.409 11:36:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:16.667 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:26:16.667 "name": "raid_bdev1", 00:26:16.667 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:16.667 "strip_size_kb": 0, 00:26:16.667 "state": "online", 00:26:16.667 "raid_level": "raid1", 00:26:16.667 "superblock": true, 00:26:16.667 "num_base_bdevs": 4, 00:26:16.667 "num_base_bdevs_discovered": 3, 00:26:16.667 "num_base_bdevs_operational": 3, 00:26:16.667 "base_bdevs_list": [ 00:26:16.667 { 00:26:16.667 "name": "spare", 00:26:16.667 "uuid": "30d0a2d4-dd37-5305-a3bc-53b7aed38ff6", 00:26:16.667 "is_configured": true, 00:26:16.667 "data_offset": 2048, 00:26:16.667 "data_size": 63488 00:26:16.667 }, 00:26:16.667 { 00:26:16.667 "name": null, 00:26:16.667 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:16.667 "is_configured": false, 00:26:16.667 "data_offset": 2048, 00:26:16.667 "data_size": 63488 00:26:16.667 }, 00:26:16.667 { 00:26:16.667 "name": "BaseBdev3", 00:26:16.667 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:16.667 "is_configured": true, 00:26:16.667 "data_offset": 2048, 00:26:16.667 "data_size": 63488 00:26:16.667 }, 00:26:16.667 { 00:26:16.667 "name": "BaseBdev4", 00:26:16.667 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:16.667 "is_configured": true, 00:26:16.667 "data_offset": 2048, 00:26:16.667 "data_size": 63488 00:26:16.667 } 00:26:16.667 ] 00:26:16.667 }' 00:26:16.667 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:26:16.667 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:26:16.667 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:26:16.925 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:26:16.925 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:16.925 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # jq -r '.[].base_bdevs_list[0].name' 00:26:16.925 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # [[ spare == \s\p\a\r\e ]] 00:26:16.925 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@752 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev spare 00:26:17.183 [2024-07-15 11:37:00.682052] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:26:17.183 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@753 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:26:17.183 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:26:17.183 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:26:17.183 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:17.183 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:17.183 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:26:17.183 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:17.183 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:17.183 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:17.183 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:17.183 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:17.183 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:17.441 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:17.441 "name": "raid_bdev1", 00:26:17.441 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:17.441 "strip_size_kb": 0, 00:26:17.441 "state": "online", 00:26:17.441 "raid_level": "raid1", 00:26:17.441 "superblock": true, 00:26:17.441 "num_base_bdevs": 4, 00:26:17.441 "num_base_bdevs_discovered": 2, 00:26:17.441 "num_base_bdevs_operational": 2, 00:26:17.441 "base_bdevs_list": [ 00:26:17.441 { 00:26:17.441 "name": null, 00:26:17.441 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:17.441 "is_configured": false, 00:26:17.441 "data_offset": 2048, 00:26:17.441 "data_size": 63488 00:26:17.441 }, 00:26:17.441 { 00:26:17.441 "name": null, 00:26:17.441 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:17.441 "is_configured": false, 00:26:17.441 "data_offset": 2048, 00:26:17.441 "data_size": 63488 00:26:17.441 }, 00:26:17.441 { 00:26:17.441 "name": "BaseBdev3", 00:26:17.441 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:17.441 "is_configured": true, 00:26:17.441 "data_offset": 2048, 00:26:17.441 "data_size": 63488 00:26:17.441 }, 00:26:17.441 { 00:26:17.441 "name": "BaseBdev4", 00:26:17.441 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:17.441 "is_configured": true, 00:26:17.441 "data_offset": 2048, 00:26:17.441 "data_size": 63488 00:26:17.441 } 00:26:17.441 ] 00:26:17.441 }' 00:26:17.441 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:17.441 11:37:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:26:18.008 11:37:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:26:18.008 [2024-07-15 11:37:01.520408] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:26:18.008 [2024-07-15 11:37:01.520557] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:26:18.008 [2024-07-15 11:37:01.520574] bdev_raid.c:3620:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:26:18.008 [2024-07-15 11:37:01.520602] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:26:18.008 [2024-07-15 11:37:01.525094] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1be8d50 00:26:18.008 [2024-07-15 11:37:01.527441] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:26:18.008 11:37:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # sleep 1 00:26:19.382 11:37:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:19.382 11:37:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:26:19.382 11:37:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:26:19.382 11:37:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:26:19.382 11:37:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:26:19.382 11:37:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:19.382 11:37:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:19.382 11:37:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:26:19.382 "name": "raid_bdev1", 00:26:19.382 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:19.382 "strip_size_kb": 0, 00:26:19.382 "state": "online", 00:26:19.382 "raid_level": "raid1", 00:26:19.382 "superblock": true, 00:26:19.382 "num_base_bdevs": 4, 00:26:19.382 "num_base_bdevs_discovered": 3, 00:26:19.382 "num_base_bdevs_operational": 3, 00:26:19.382 "process": { 00:26:19.382 "type": "rebuild", 00:26:19.382 "target": "spare", 00:26:19.382 "progress": { 00:26:19.382 "blocks": 22528, 00:26:19.382 "percent": 35 00:26:19.382 } 00:26:19.382 }, 00:26:19.382 "base_bdevs_list": [ 00:26:19.382 { 00:26:19.382 "name": "spare", 00:26:19.382 "uuid": "30d0a2d4-dd37-5305-a3bc-53b7aed38ff6", 00:26:19.382 "is_configured": true, 00:26:19.382 "data_offset": 2048, 00:26:19.382 "data_size": 63488 00:26:19.382 }, 00:26:19.382 { 00:26:19.382 "name": null, 00:26:19.382 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:19.382 "is_configured": false, 00:26:19.382 "data_offset": 2048, 00:26:19.382 "data_size": 63488 00:26:19.382 }, 00:26:19.382 { 00:26:19.382 "name": "BaseBdev3", 00:26:19.382 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:19.382 "is_configured": true, 00:26:19.382 "data_offset": 2048, 00:26:19.382 "data_size": 63488 00:26:19.382 }, 00:26:19.382 { 00:26:19.382 "name": "BaseBdev4", 00:26:19.382 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:19.382 "is_configured": true, 00:26:19.382 "data_offset": 2048, 00:26:19.382 "data_size": 63488 00:26:19.382 } 00:26:19.382 ] 00:26:19.382 }' 00:26:19.382 11:37:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:26:19.382 11:37:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:19.382 11:37:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:26:19.382 11:37:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:26:19.382 11:37:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@759 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:26:19.641 [2024-07-15 11:37:03.049699] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:26:19.641 [2024-07-15 11:37:03.140120] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:26:19.641 [2024-07-15 11:37:03.140165] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:19.641 [2024-07-15 11:37:03.140182] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:26:19.641 [2024-07-15 11:37:03.140190] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:26:19.641 11:37:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@760 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:26:19.641 11:37:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:26:19.641 11:37:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:26:19.641 11:37:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:19.641 11:37:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:19.641 11:37:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:26:19.641 11:37:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:19.641 11:37:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:19.641 11:37:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:19.641 11:37:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:19.641 11:37:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:19.641 11:37:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:19.899 11:37:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:19.899 "name": "raid_bdev1", 00:26:19.899 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:19.899 "strip_size_kb": 0, 00:26:19.899 "state": "online", 00:26:19.899 "raid_level": "raid1", 00:26:19.899 "superblock": true, 00:26:19.899 "num_base_bdevs": 4, 00:26:19.899 "num_base_bdevs_discovered": 2, 00:26:19.899 "num_base_bdevs_operational": 2, 00:26:19.899 "base_bdevs_list": [ 00:26:19.899 { 00:26:19.899 "name": null, 00:26:19.899 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:19.899 "is_configured": false, 00:26:19.899 "data_offset": 2048, 00:26:19.899 "data_size": 63488 00:26:19.899 }, 00:26:19.899 { 00:26:19.899 "name": null, 00:26:19.899 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:19.899 "is_configured": false, 00:26:19.899 "data_offset": 2048, 00:26:19.899 "data_size": 63488 00:26:19.899 }, 00:26:19.899 { 00:26:19.899 "name": "BaseBdev3", 00:26:19.899 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:19.899 "is_configured": true, 00:26:19.899 "data_offset": 2048, 00:26:19.899 "data_size": 63488 00:26:19.899 }, 00:26:19.899 { 00:26:19.899 "name": "BaseBdev4", 00:26:19.899 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:19.899 "is_configured": true, 00:26:19.899 "data_offset": 2048, 00:26:19.899 "data_size": 63488 00:26:19.899 } 00:26:19.899 ] 00:26:19.899 }' 00:26:19.899 11:37:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:19.899 11:37:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:26:20.465 11:37:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:26:20.722 [2024-07-15 11:37:04.143101] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:26:20.722 [2024-07-15 11:37:04.143149] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:20.722 [2024-07-15 11:37:04.143172] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1d94170 00:26:20.722 [2024-07-15 11:37:04.143184] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:20.722 [2024-07-15 11:37:04.143553] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:20.722 [2024-07-15 11:37:04.143571] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:26:20.722 [2024-07-15 11:37:04.143649] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:26:20.722 [2024-07-15 11:37:04.143661] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:26:20.722 [2024-07-15 11:37:04.143672] bdev_raid.c:3620:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:26:20.722 [2024-07-15 11:37:04.143691] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:26:20.722 [2024-07-15 11:37:04.148142] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1be8ce0 00:26:20.722 spare 00:26:20.722 [2024-07-15 11:37:04.149534] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:26:20.722 11:37:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # sleep 1 00:26:21.655 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:21.655 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:26:21.655 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:26:21.655 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=spare 00:26:21.655 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:26:21.655 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:21.655 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:21.914 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:26:21.914 "name": "raid_bdev1", 00:26:21.914 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:21.914 "strip_size_kb": 0, 00:26:21.914 "state": "online", 00:26:21.914 "raid_level": "raid1", 00:26:21.914 "superblock": true, 00:26:21.914 "num_base_bdevs": 4, 00:26:21.914 "num_base_bdevs_discovered": 3, 00:26:21.914 "num_base_bdevs_operational": 3, 00:26:21.914 "process": { 00:26:21.914 "type": "rebuild", 00:26:21.914 "target": "spare", 00:26:21.914 "progress": { 00:26:21.914 "blocks": 22528, 00:26:21.914 "percent": 35 00:26:21.914 } 00:26:21.914 }, 00:26:21.914 "base_bdevs_list": [ 00:26:21.914 { 00:26:21.914 "name": "spare", 00:26:21.914 "uuid": "30d0a2d4-dd37-5305-a3bc-53b7aed38ff6", 00:26:21.914 "is_configured": true, 00:26:21.914 "data_offset": 2048, 00:26:21.914 "data_size": 63488 00:26:21.914 }, 00:26:21.914 { 00:26:21.914 "name": null, 00:26:21.914 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:21.914 "is_configured": false, 00:26:21.914 "data_offset": 2048, 00:26:21.914 "data_size": 63488 00:26:21.914 }, 00:26:21.914 { 00:26:21.914 "name": "BaseBdev3", 00:26:21.914 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:21.914 "is_configured": true, 00:26:21.914 "data_offset": 2048, 00:26:21.914 "data_size": 63488 00:26:21.914 }, 00:26:21.914 { 00:26:21.914 "name": "BaseBdev4", 00:26:21.914 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:21.914 "is_configured": true, 00:26:21.914 "data_offset": 2048, 00:26:21.914 "data_size": 63488 00:26:21.914 } 00:26:21.914 ] 00:26:21.914 }' 00:26:21.914 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:26:21.914 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:21.914 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:26:21.914 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:26:21.914 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@766 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:26:22.173 [2024-07-15 11:37:05.669901] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:26:22.173 [2024-07-15 11:37:05.762046] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:26:22.173 [2024-07-15 11:37:05.762103] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:22.173 [2024-07-15 11:37:05.762120] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:26:22.173 [2024-07-15 11:37:05.762129] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:26:22.431 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@767 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:26:22.431 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:26:22.431 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:26:22.431 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:22.431 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:22.431 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:26:22.431 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:22.431 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:22.431 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:22.431 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:22.431 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:22.431 11:37:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:22.689 11:37:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:22.689 "name": "raid_bdev1", 00:26:22.689 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:22.689 "strip_size_kb": 0, 00:26:22.689 "state": "online", 00:26:22.689 "raid_level": "raid1", 00:26:22.689 "superblock": true, 00:26:22.689 "num_base_bdevs": 4, 00:26:22.689 "num_base_bdevs_discovered": 2, 00:26:22.689 "num_base_bdevs_operational": 2, 00:26:22.689 "base_bdevs_list": [ 00:26:22.689 { 00:26:22.689 "name": null, 00:26:22.689 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:22.689 "is_configured": false, 00:26:22.689 "data_offset": 2048, 00:26:22.689 "data_size": 63488 00:26:22.689 }, 00:26:22.689 { 00:26:22.689 "name": null, 00:26:22.689 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:22.689 "is_configured": false, 00:26:22.689 "data_offset": 2048, 00:26:22.689 "data_size": 63488 00:26:22.689 }, 00:26:22.689 { 00:26:22.689 "name": "BaseBdev3", 00:26:22.689 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:22.689 "is_configured": true, 00:26:22.689 "data_offset": 2048, 00:26:22.689 "data_size": 63488 00:26:22.689 }, 00:26:22.689 { 00:26:22.689 "name": "BaseBdev4", 00:26:22.689 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:22.689 "is_configured": true, 00:26:22.689 "data_offset": 2048, 00:26:22.689 "data_size": 63488 00:26:22.689 } 00:26:22.689 ] 00:26:22.689 }' 00:26:22.689 11:37:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:22.689 11:37:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:26:23.268 11:37:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # verify_raid_bdev_process raid_bdev1 none none 00:26:23.268 11:37:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:26:23.268 11:37:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:26:23.268 11:37:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=none 00:26:23.268 11:37:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:26:23.268 11:37:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:23.269 11:37:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:23.527 11:37:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:26:23.527 "name": "raid_bdev1", 00:26:23.527 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:23.527 "strip_size_kb": 0, 00:26:23.527 "state": "online", 00:26:23.527 "raid_level": "raid1", 00:26:23.527 "superblock": true, 00:26:23.527 "num_base_bdevs": 4, 00:26:23.527 "num_base_bdevs_discovered": 2, 00:26:23.527 "num_base_bdevs_operational": 2, 00:26:23.527 "base_bdevs_list": [ 00:26:23.527 { 00:26:23.527 "name": null, 00:26:23.527 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:23.527 "is_configured": false, 00:26:23.527 "data_offset": 2048, 00:26:23.527 "data_size": 63488 00:26:23.527 }, 00:26:23.527 { 00:26:23.527 "name": null, 00:26:23.527 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:23.527 "is_configured": false, 00:26:23.527 "data_offset": 2048, 00:26:23.527 "data_size": 63488 00:26:23.527 }, 00:26:23.527 { 00:26:23.527 "name": "BaseBdev3", 00:26:23.527 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:23.527 "is_configured": true, 00:26:23.527 "data_offset": 2048, 00:26:23.527 "data_size": 63488 00:26:23.527 }, 00:26:23.527 { 00:26:23.527 "name": "BaseBdev4", 00:26:23.527 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:23.527 "is_configured": true, 00:26:23.527 "data_offset": 2048, 00:26:23.527 "data_size": 63488 00:26:23.527 } 00:26:23.527 ] 00:26:23.527 }' 00:26:23.527 11:37:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:26:23.527 11:37:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:26:23.527 11:37:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:26:23.527 11:37:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:26:23.527 11:37:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@771 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete BaseBdev1 00:26:23.786 11:37:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@772 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:26:24.044 [2024-07-15 11:37:07.463668] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:26:24.044 [2024-07-15 11:37:07.463720] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:24.044 [2024-07-15 11:37:07.463741] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1d20210 00:26:24.044 [2024-07-15 11:37:07.463753] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:24.044 [2024-07-15 11:37:07.464106] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:24.044 [2024-07-15 11:37:07.464124] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:26:24.044 [2024-07-15 11:37:07.464183] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:26:24.044 [2024-07-15 11:37:07.464196] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:26:24.044 [2024-07-15 11:37:07.464206] bdev_raid.c:3581:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:26:24.044 BaseBdev1 00:26:24.044 11:37:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # sleep 1 00:26:24.981 11:37:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:26:24.981 11:37:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:26:24.981 11:37:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:26:24.981 11:37:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:24.981 11:37:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:24.981 11:37:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:26:24.981 11:37:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:24.981 11:37:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:24.981 11:37:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:24.981 11:37:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:24.981 11:37:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:24.981 11:37:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:25.239 11:37:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:25.239 "name": "raid_bdev1", 00:26:25.239 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:25.239 "strip_size_kb": 0, 00:26:25.239 "state": "online", 00:26:25.239 "raid_level": "raid1", 00:26:25.239 "superblock": true, 00:26:25.239 "num_base_bdevs": 4, 00:26:25.239 "num_base_bdevs_discovered": 2, 00:26:25.239 "num_base_bdevs_operational": 2, 00:26:25.239 "base_bdevs_list": [ 00:26:25.239 { 00:26:25.239 "name": null, 00:26:25.239 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:25.239 "is_configured": false, 00:26:25.239 "data_offset": 2048, 00:26:25.239 "data_size": 63488 00:26:25.239 }, 00:26:25.239 { 00:26:25.239 "name": null, 00:26:25.239 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:25.239 "is_configured": false, 00:26:25.239 "data_offset": 2048, 00:26:25.239 "data_size": 63488 00:26:25.239 }, 00:26:25.239 { 00:26:25.239 "name": "BaseBdev3", 00:26:25.239 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:25.239 "is_configured": true, 00:26:25.239 "data_offset": 2048, 00:26:25.239 "data_size": 63488 00:26:25.239 }, 00:26:25.239 { 00:26:25.239 "name": "BaseBdev4", 00:26:25.239 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:25.239 "is_configured": true, 00:26:25.239 "data_offset": 2048, 00:26:25.239 "data_size": 63488 00:26:25.239 } 00:26:25.239 ] 00:26:25.239 }' 00:26:25.239 11:37:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:25.239 11:37:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:26:25.807 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # verify_raid_bdev_process raid_bdev1 none none 00:26:25.807 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:26:25.807 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:26:25.807 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=none 00:26:25.807 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:26:25.807 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:25.807 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:26.065 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:26:26.065 "name": "raid_bdev1", 00:26:26.065 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:26.065 "strip_size_kb": 0, 00:26:26.065 "state": "online", 00:26:26.065 "raid_level": "raid1", 00:26:26.065 "superblock": true, 00:26:26.065 "num_base_bdevs": 4, 00:26:26.065 "num_base_bdevs_discovered": 2, 00:26:26.065 "num_base_bdevs_operational": 2, 00:26:26.065 "base_bdevs_list": [ 00:26:26.065 { 00:26:26.065 "name": null, 00:26:26.065 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:26.065 "is_configured": false, 00:26:26.065 "data_offset": 2048, 00:26:26.065 "data_size": 63488 00:26:26.065 }, 00:26:26.065 { 00:26:26.065 "name": null, 00:26:26.065 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:26.065 "is_configured": false, 00:26:26.065 "data_offset": 2048, 00:26:26.065 "data_size": 63488 00:26:26.065 }, 00:26:26.065 { 00:26:26.065 "name": "BaseBdev3", 00:26:26.065 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:26.066 "is_configured": true, 00:26:26.066 "data_offset": 2048, 00:26:26.066 "data_size": 63488 00:26:26.066 }, 00:26:26.066 { 00:26:26.066 "name": "BaseBdev4", 00:26:26.066 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:26.066 "is_configured": true, 00:26:26.066 "data_offset": 2048, 00:26:26.066 "data_size": 63488 00:26:26.066 } 00:26:26.066 ] 00:26:26.066 }' 00:26:26.066 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:26:26.066 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:26:26.066 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:26:26.325 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:26:26.325 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:26:26.325 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@648 -- # local es=0 00:26:26.325 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:26:26.325 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:26:26.325 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:26:26.325 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:26:26.325 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:26:26.325 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:26:26.325 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:26:26.325 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:26:26.325 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py ]] 00:26:26.325 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:26:26.325 [2024-07-15 11:37:09.894419] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:26:26.325 [2024-07-15 11:37:09.894544] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:26:26.325 [2024-07-15 11:37:09.894560] bdev_raid.c:3581:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:26:26.325 request: 00:26:26.325 { 00:26:26.325 "base_bdev": "BaseBdev1", 00:26:26.325 "raid_bdev": "raid_bdev1", 00:26:26.325 "method": "bdev_raid_add_base_bdev", 00:26:26.325 "req_id": 1 00:26:26.325 } 00:26:26.325 Got JSON-RPC error response 00:26:26.325 response: 00:26:26.325 { 00:26:26.325 "code": -22, 00:26:26.325 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:26:26.325 } 00:26:26.325 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@651 -- # es=1 00:26:26.325 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:26:26.325 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:26:26.325 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:26:26.325 11:37:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # sleep 1 00:26:27.740 11:37:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:26:27.740 11:37:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:26:27.740 11:37:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:26:27.740 11:37:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:27.740 11:37:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:27.740 11:37:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:26:27.740 11:37:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:27.740 11:37:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:27.740 11:37:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:27.740 11:37:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:27.740 11:37:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:27.740 11:37:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:27.740 11:37:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:27.740 "name": "raid_bdev1", 00:26:27.740 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:27.740 "strip_size_kb": 0, 00:26:27.740 "state": "online", 00:26:27.740 "raid_level": "raid1", 00:26:27.740 "superblock": true, 00:26:27.740 "num_base_bdevs": 4, 00:26:27.740 "num_base_bdevs_discovered": 2, 00:26:27.740 "num_base_bdevs_operational": 2, 00:26:27.740 "base_bdevs_list": [ 00:26:27.740 { 00:26:27.740 "name": null, 00:26:27.740 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:27.740 "is_configured": false, 00:26:27.740 "data_offset": 2048, 00:26:27.740 "data_size": 63488 00:26:27.740 }, 00:26:27.740 { 00:26:27.740 "name": null, 00:26:27.740 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:27.740 "is_configured": false, 00:26:27.740 "data_offset": 2048, 00:26:27.740 "data_size": 63488 00:26:27.740 }, 00:26:27.740 { 00:26:27.740 "name": "BaseBdev3", 00:26:27.740 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:27.740 "is_configured": true, 00:26:27.740 "data_offset": 2048, 00:26:27.740 "data_size": 63488 00:26:27.740 }, 00:26:27.740 { 00:26:27.740 "name": "BaseBdev4", 00:26:27.740 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:27.740 "is_configured": true, 00:26:27.740 "data_offset": 2048, 00:26:27.740 "data_size": 63488 00:26:27.740 } 00:26:27.740 ] 00:26:27.740 }' 00:26:27.740 11:37:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:27.740 11:37:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:26:28.304 11:37:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # verify_raid_bdev_process raid_bdev1 none none 00:26:28.304 11:37:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:26:28.304 11:37:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:26:28.304 11:37:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@184 -- # local target=none 00:26:28.304 11:37:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:26:28.304 11:37:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:28.304 11:37:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:28.561 11:37:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:26:28.561 "name": "raid_bdev1", 00:26:28.561 "uuid": "7c90e45d-524b-4bc9-aeb4-cc1385867897", 00:26:28.561 "strip_size_kb": 0, 00:26:28.561 "state": "online", 00:26:28.561 "raid_level": "raid1", 00:26:28.561 "superblock": true, 00:26:28.561 "num_base_bdevs": 4, 00:26:28.561 "num_base_bdevs_discovered": 2, 00:26:28.561 "num_base_bdevs_operational": 2, 00:26:28.561 "base_bdevs_list": [ 00:26:28.561 { 00:26:28.561 "name": null, 00:26:28.561 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:28.561 "is_configured": false, 00:26:28.561 "data_offset": 2048, 00:26:28.561 "data_size": 63488 00:26:28.561 }, 00:26:28.561 { 00:26:28.561 "name": null, 00:26:28.561 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:28.561 "is_configured": false, 00:26:28.561 "data_offset": 2048, 00:26:28.561 "data_size": 63488 00:26:28.561 }, 00:26:28.561 { 00:26:28.561 "name": "BaseBdev3", 00:26:28.561 "uuid": "9d5cc731-fd33-5662-9881-23be31cf9255", 00:26:28.561 "is_configured": true, 00:26:28.561 "data_offset": 2048, 00:26:28.561 "data_size": 63488 00:26:28.561 }, 00:26:28.561 { 00:26:28.561 "name": "BaseBdev4", 00:26:28.561 "uuid": "2b53fa77-e730-5a29-9d47-0ffd908a409e", 00:26:28.561 "is_configured": true, 00:26:28.561 "data_offset": 2048, 00:26:28.561 "data_size": 63488 00:26:28.561 } 00:26:28.561 ] 00:26:28.561 }' 00:26:28.561 11:37:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:26:28.561 11:37:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:26:28.561 11:37:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:26:28.561 11:37:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:26:28.561 11:37:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@782 -- # killprocess 995796 00:26:28.561 11:37:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@948 -- # '[' -z 995796 ']' 00:26:28.561 11:37:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@952 -- # kill -0 995796 00:26:28.561 11:37:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@953 -- # uname 00:26:28.561 11:37:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:26:28.561 11:37:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 995796 00:26:28.818 11:37:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:26:28.818 11:37:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:26:28.818 11:37:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@966 -- # echo 'killing process with pid 995796' 00:26:28.818 killing process with pid 995796 00:26:28.818 11:37:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@967 -- # kill 995796 00:26:28.818 Received shutdown signal, test time was about 27.147438 seconds 00:26:28.818 00:26:28.818 Latency(us) 00:26:28.818 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:28.818 =================================================================================================================== 00:26:28.818 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:26:28.818 [2024-07-15 11:37:12.164966] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:26:28.818 [2024-07-15 11:37:12.165068] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:28.818 [2024-07-15 11:37:12.165127] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:28.818 [2024-07-15 11:37:12.165139] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1d1bf20 name raid_bdev1, state offline 00:26:28.818 11:37:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@972 -- # wait 995796 00:26:28.818 [2024-07-15 11:37:12.212364] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:26:29.075 11:37:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # return 0 00:26:29.075 00:26:29.075 real 0m33.646s 00:26:29.075 user 0m52.839s 00:26:29.075 sys 0m5.255s 00:26:29.075 11:37:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1124 -- # xtrace_disable 00:26:29.075 11:37:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:26:29.075 ************************************ 00:26:29.075 END TEST raid_rebuild_test_sb_io 00:26:29.075 ************************************ 00:26:29.075 11:37:12 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:26:29.075 11:37:12 bdev_raid -- bdev/bdev_raid.sh@884 -- # '[' n == y ']' 00:26:29.075 11:37:12 bdev_raid -- bdev/bdev_raid.sh@896 -- # base_blocklen=4096 00:26:29.075 11:37:12 bdev_raid -- bdev/bdev_raid.sh@898 -- # run_test raid_state_function_test_sb_4k raid_state_function_test raid1 2 true 00:26:29.075 11:37:12 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:26:29.075 11:37:12 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:26:29.075 11:37:12 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:26:29.075 ************************************ 00:26:29.075 START TEST raid_state_function_test_sb_4k 00:26:29.075 ************************************ 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1123 -- # raid_state_function_test raid1 2 true 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@220 -- # local raid_level=raid1 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=2 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@222 -- # local superblock=true 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@226 -- # local strip_size 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@230 -- # '[' raid1 '!=' raid1 ']' 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@234 -- # strip_size=0 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@237 -- # '[' true = true ']' 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@238 -- # superblock_create_arg=-s 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@244 -- # raid_pid=1000558 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 1000558' 00:26:29.075 Process raid pid: 1000558 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@246 -- # waitforlisten 1000558 /var/tmp/spdk-raid.sock 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@829 -- # '[' -z 1000558 ']' 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@834 -- # local max_retries=100 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:26:29.075 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@838 -- # xtrace_disable 00:26:29.075 11:37:12 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:26:29.075 [2024-07-15 11:37:12.601447] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:26:29.075 [2024-07-15 11:37:12.601515] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:29.333 [2024-07-15 11:37:12.732461] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:29.333 [2024-07-15 11:37:12.834384] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:26:29.333 [2024-07-15 11:37:12.889607] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:29.333 [2024-07-15 11:37:12.889633] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:30.260 11:37:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:26:30.260 11:37:13 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@862 -- # return 0 00:26:30.260 11:37:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:26:30.260 [2024-07-15 11:37:13.752778] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:26:30.261 [2024-07-15 11:37:13.752818] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:26:30.261 [2024-07-15 11:37:13.752829] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:26:30.261 [2024-07-15 11:37:13.752841] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:26:30.261 11:37:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:26:30.261 11:37:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:26:30.261 11:37:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:26:30.261 11:37:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:30.261 11:37:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:30.261 11:37:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:26:30.261 11:37:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:30.261 11:37:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:30.261 11:37:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:30.261 11:37:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:30.261 11:37:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:30.261 11:37:13 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:30.518 11:37:14 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:30.518 "name": "Existed_Raid", 00:26:30.518 "uuid": "5fe51a3e-ee8b-49ac-8742-242a673eeb69", 00:26:30.518 "strip_size_kb": 0, 00:26:30.518 "state": "configuring", 00:26:30.518 "raid_level": "raid1", 00:26:30.518 "superblock": true, 00:26:30.518 "num_base_bdevs": 2, 00:26:30.518 "num_base_bdevs_discovered": 0, 00:26:30.518 "num_base_bdevs_operational": 2, 00:26:30.518 "base_bdevs_list": [ 00:26:30.518 { 00:26:30.518 "name": "BaseBdev1", 00:26:30.518 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:30.518 "is_configured": false, 00:26:30.518 "data_offset": 0, 00:26:30.518 "data_size": 0 00:26:30.518 }, 00:26:30.518 { 00:26:30.518 "name": "BaseBdev2", 00:26:30.518 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:30.518 "is_configured": false, 00:26:30.518 "data_offset": 0, 00:26:30.518 "data_size": 0 00:26:30.518 } 00:26:30.518 ] 00:26:30.518 }' 00:26:30.518 11:37:14 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:30.518 11:37:14 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:26:31.081 11:37:14 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:26:31.337 [2024-07-15 11:37:14.839525] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:26:31.337 [2024-07-15 11:37:14.839558] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x19c6a80 name Existed_Raid, state configuring 00:26:31.337 11:37:14 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:26:31.594 [2024-07-15 11:37:15.084198] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:26:31.594 [2024-07-15 11:37:15.084229] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:26:31.594 [2024-07-15 11:37:15.084239] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:26:31.594 [2024-07-15 11:37:15.084250] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:26:31.594 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -b BaseBdev1 00:26:31.851 [2024-07-15 11:37:15.338735] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:26:31.851 BaseBdev1 00:26:31.851 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:26:31.851 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:26:31.851 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:26:31.851 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@899 -- # local i 00:26:31.851 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:26:31.851 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:26:31.851 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:26:32.108 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:26:32.365 [ 00:26:32.365 { 00:26:32.365 "name": "BaseBdev1", 00:26:32.365 "aliases": [ 00:26:32.365 "0e2fc37f-de6e-4a79-b91a-6525da35cb9a" 00:26:32.365 ], 00:26:32.365 "product_name": "Malloc disk", 00:26:32.365 "block_size": 4096, 00:26:32.365 "num_blocks": 8192, 00:26:32.365 "uuid": "0e2fc37f-de6e-4a79-b91a-6525da35cb9a", 00:26:32.365 "assigned_rate_limits": { 00:26:32.365 "rw_ios_per_sec": 0, 00:26:32.365 "rw_mbytes_per_sec": 0, 00:26:32.365 "r_mbytes_per_sec": 0, 00:26:32.365 "w_mbytes_per_sec": 0 00:26:32.365 }, 00:26:32.365 "claimed": true, 00:26:32.365 "claim_type": "exclusive_write", 00:26:32.365 "zoned": false, 00:26:32.365 "supported_io_types": { 00:26:32.365 "read": true, 00:26:32.365 "write": true, 00:26:32.365 "unmap": true, 00:26:32.365 "flush": true, 00:26:32.365 "reset": true, 00:26:32.365 "nvme_admin": false, 00:26:32.365 "nvme_io": false, 00:26:32.365 "nvme_io_md": false, 00:26:32.365 "write_zeroes": true, 00:26:32.365 "zcopy": true, 00:26:32.365 "get_zone_info": false, 00:26:32.365 "zone_management": false, 00:26:32.365 "zone_append": false, 00:26:32.365 "compare": false, 00:26:32.365 "compare_and_write": false, 00:26:32.365 "abort": true, 00:26:32.365 "seek_hole": false, 00:26:32.365 "seek_data": false, 00:26:32.365 "copy": true, 00:26:32.365 "nvme_iov_md": false 00:26:32.365 }, 00:26:32.365 "memory_domains": [ 00:26:32.365 { 00:26:32.365 "dma_device_id": "system", 00:26:32.365 "dma_device_type": 1 00:26:32.365 }, 00:26:32.365 { 00:26:32.365 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:32.365 "dma_device_type": 2 00:26:32.365 } 00:26:32.365 ], 00:26:32.365 "driver_specific": {} 00:26:32.365 } 00:26:32.365 ] 00:26:32.365 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@905 -- # return 0 00:26:32.365 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:26:32.365 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:26:32.365 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:26:32.365 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:32.365 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:32.365 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:26:32.365 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:32.365 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:32.365 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:32.365 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:32.365 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:32.365 11:37:15 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:32.622 11:37:16 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:32.622 "name": "Existed_Raid", 00:26:32.622 "uuid": "545606a7-b5a0-4e0e-9dde-622da69a8440", 00:26:32.622 "strip_size_kb": 0, 00:26:32.622 "state": "configuring", 00:26:32.622 "raid_level": "raid1", 00:26:32.622 "superblock": true, 00:26:32.622 "num_base_bdevs": 2, 00:26:32.622 "num_base_bdevs_discovered": 1, 00:26:32.622 "num_base_bdevs_operational": 2, 00:26:32.622 "base_bdevs_list": [ 00:26:32.622 { 00:26:32.622 "name": "BaseBdev1", 00:26:32.622 "uuid": "0e2fc37f-de6e-4a79-b91a-6525da35cb9a", 00:26:32.622 "is_configured": true, 00:26:32.622 "data_offset": 256, 00:26:32.622 "data_size": 7936 00:26:32.622 }, 00:26:32.622 { 00:26:32.622 "name": "BaseBdev2", 00:26:32.622 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:32.622 "is_configured": false, 00:26:32.622 "data_offset": 0, 00:26:32.622 "data_size": 0 00:26:32.622 } 00:26:32.622 ] 00:26:32.622 }' 00:26:32.622 11:37:16 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:32.622 11:37:16 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:26:33.186 11:37:16 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:26:33.443 [2024-07-15 11:37:16.934978] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:26:33.443 [2024-07-15 11:37:16.935019] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x19c6350 name Existed_Raid, state configuring 00:26:33.443 11:37:16 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:26:33.699 [2024-07-15 11:37:17.179655] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:26:33.699 [2024-07-15 11:37:17.181139] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:26:33.699 [2024-07-15 11:37:17.181171] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:26:33.699 11:37:17 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:26:33.699 11:37:17 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:26:33.699 11:37:17 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:26:33.699 11:37:17 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:26:33.699 11:37:17 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:26:33.699 11:37:17 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:33.700 11:37:17 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:33.700 11:37:17 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:26:33.700 11:37:17 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:33.700 11:37:17 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:33.700 11:37:17 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:33.700 11:37:17 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:33.700 11:37:17 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:33.700 11:37:17 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:33.956 11:37:17 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:33.956 "name": "Existed_Raid", 00:26:33.956 "uuid": "53334b15-7569-4888-826b-cc1339ec2186", 00:26:33.956 "strip_size_kb": 0, 00:26:33.956 "state": "configuring", 00:26:33.956 "raid_level": "raid1", 00:26:33.956 "superblock": true, 00:26:33.956 "num_base_bdevs": 2, 00:26:33.956 "num_base_bdevs_discovered": 1, 00:26:33.956 "num_base_bdevs_operational": 2, 00:26:33.956 "base_bdevs_list": [ 00:26:33.956 { 00:26:33.956 "name": "BaseBdev1", 00:26:33.956 "uuid": "0e2fc37f-de6e-4a79-b91a-6525da35cb9a", 00:26:33.956 "is_configured": true, 00:26:33.956 "data_offset": 256, 00:26:33.956 "data_size": 7936 00:26:33.956 }, 00:26:33.956 { 00:26:33.956 "name": "BaseBdev2", 00:26:33.956 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:33.956 "is_configured": false, 00:26:33.956 "data_offset": 0, 00:26:33.956 "data_size": 0 00:26:33.956 } 00:26:33.956 ] 00:26:33.956 }' 00:26:33.956 11:37:17 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:33.956 11:37:17 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:26:34.519 11:37:17 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -b BaseBdev2 00:26:34.776 [2024-07-15 11:37:18.121469] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:26:34.776 [2024-07-15 11:37:18.121622] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x19c7000 00:26:34.776 [2024-07-15 11:37:18.121635] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:26:34.776 [2024-07-15 11:37:18.121806] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x18e10c0 00:26:34.776 [2024-07-15 11:37:18.121938] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x19c7000 00:26:34.776 [2024-07-15 11:37:18.121949] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x19c7000 00:26:34.776 [2024-07-15 11:37:18.122044] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:34.776 BaseBdev2 00:26:34.776 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:26:34.776 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:26:34.776 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:26:34.776 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@899 -- # local i 00:26:34.776 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:26:34.776 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:26:34.776 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:26:34.776 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:26:35.033 [ 00:26:35.033 { 00:26:35.033 "name": "BaseBdev2", 00:26:35.033 "aliases": [ 00:26:35.033 "b21e0c64-cd47-4ccd-a404-9a5641e62cf0" 00:26:35.033 ], 00:26:35.033 "product_name": "Malloc disk", 00:26:35.033 "block_size": 4096, 00:26:35.033 "num_blocks": 8192, 00:26:35.033 "uuid": "b21e0c64-cd47-4ccd-a404-9a5641e62cf0", 00:26:35.033 "assigned_rate_limits": { 00:26:35.033 "rw_ios_per_sec": 0, 00:26:35.033 "rw_mbytes_per_sec": 0, 00:26:35.033 "r_mbytes_per_sec": 0, 00:26:35.033 "w_mbytes_per_sec": 0 00:26:35.033 }, 00:26:35.033 "claimed": true, 00:26:35.033 "claim_type": "exclusive_write", 00:26:35.033 "zoned": false, 00:26:35.033 "supported_io_types": { 00:26:35.033 "read": true, 00:26:35.033 "write": true, 00:26:35.033 "unmap": true, 00:26:35.033 "flush": true, 00:26:35.033 "reset": true, 00:26:35.033 "nvme_admin": false, 00:26:35.033 "nvme_io": false, 00:26:35.033 "nvme_io_md": false, 00:26:35.033 "write_zeroes": true, 00:26:35.033 "zcopy": true, 00:26:35.033 "get_zone_info": false, 00:26:35.033 "zone_management": false, 00:26:35.033 "zone_append": false, 00:26:35.033 "compare": false, 00:26:35.034 "compare_and_write": false, 00:26:35.034 "abort": true, 00:26:35.034 "seek_hole": false, 00:26:35.034 "seek_data": false, 00:26:35.034 "copy": true, 00:26:35.034 "nvme_iov_md": false 00:26:35.034 }, 00:26:35.034 "memory_domains": [ 00:26:35.034 { 00:26:35.034 "dma_device_id": "system", 00:26:35.034 "dma_device_type": 1 00:26:35.034 }, 00:26:35.034 { 00:26:35.034 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:35.034 "dma_device_type": 2 00:26:35.034 } 00:26:35.034 ], 00:26:35.034 "driver_specific": {} 00:26:35.034 } 00:26:35.034 ] 00:26:35.034 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@905 -- # return 0 00:26:35.034 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:26:35.034 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:26:35.034 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:26:35.034 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:26:35.034 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:26:35.034 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:35.034 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:35.034 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:26:35.034 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:35.034 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:35.034 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:35.034 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:35.034 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:35.034 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:35.290 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:35.290 "name": "Existed_Raid", 00:26:35.290 "uuid": "53334b15-7569-4888-826b-cc1339ec2186", 00:26:35.290 "strip_size_kb": 0, 00:26:35.290 "state": "online", 00:26:35.290 "raid_level": "raid1", 00:26:35.290 "superblock": true, 00:26:35.290 "num_base_bdevs": 2, 00:26:35.290 "num_base_bdevs_discovered": 2, 00:26:35.290 "num_base_bdevs_operational": 2, 00:26:35.290 "base_bdevs_list": [ 00:26:35.290 { 00:26:35.290 "name": "BaseBdev1", 00:26:35.290 "uuid": "0e2fc37f-de6e-4a79-b91a-6525da35cb9a", 00:26:35.290 "is_configured": true, 00:26:35.290 "data_offset": 256, 00:26:35.290 "data_size": 7936 00:26:35.290 }, 00:26:35.290 { 00:26:35.290 "name": "BaseBdev2", 00:26:35.290 "uuid": "b21e0c64-cd47-4ccd-a404-9a5641e62cf0", 00:26:35.290 "is_configured": true, 00:26:35.290 "data_offset": 256, 00:26:35.290 "data_size": 7936 00:26:35.290 } 00:26:35.290 ] 00:26:35.290 }' 00:26:35.290 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:35.290 11:37:18 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:26:35.853 11:37:19 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:26:35.853 11:37:19 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:26:35.853 11:37:19 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:26:35.853 11:37:19 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:26:35.853 11:37:19 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:26:35.853 11:37:19 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@198 -- # local name 00:26:35.853 11:37:19 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:26:35.853 11:37:19 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:26:36.110 [2024-07-15 11:37:19.553540] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:36.110 11:37:19 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:26:36.110 "name": "Existed_Raid", 00:26:36.110 "aliases": [ 00:26:36.110 "53334b15-7569-4888-826b-cc1339ec2186" 00:26:36.110 ], 00:26:36.110 "product_name": "Raid Volume", 00:26:36.110 "block_size": 4096, 00:26:36.110 "num_blocks": 7936, 00:26:36.110 "uuid": "53334b15-7569-4888-826b-cc1339ec2186", 00:26:36.110 "assigned_rate_limits": { 00:26:36.110 "rw_ios_per_sec": 0, 00:26:36.110 "rw_mbytes_per_sec": 0, 00:26:36.110 "r_mbytes_per_sec": 0, 00:26:36.110 "w_mbytes_per_sec": 0 00:26:36.110 }, 00:26:36.110 "claimed": false, 00:26:36.110 "zoned": false, 00:26:36.110 "supported_io_types": { 00:26:36.110 "read": true, 00:26:36.110 "write": true, 00:26:36.110 "unmap": false, 00:26:36.110 "flush": false, 00:26:36.110 "reset": true, 00:26:36.110 "nvme_admin": false, 00:26:36.110 "nvme_io": false, 00:26:36.110 "nvme_io_md": false, 00:26:36.110 "write_zeroes": true, 00:26:36.110 "zcopy": false, 00:26:36.110 "get_zone_info": false, 00:26:36.110 "zone_management": false, 00:26:36.110 "zone_append": false, 00:26:36.110 "compare": false, 00:26:36.110 "compare_and_write": false, 00:26:36.110 "abort": false, 00:26:36.110 "seek_hole": false, 00:26:36.110 "seek_data": false, 00:26:36.110 "copy": false, 00:26:36.110 "nvme_iov_md": false 00:26:36.110 }, 00:26:36.110 "memory_domains": [ 00:26:36.110 { 00:26:36.110 "dma_device_id": "system", 00:26:36.110 "dma_device_type": 1 00:26:36.110 }, 00:26:36.110 { 00:26:36.110 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:36.110 "dma_device_type": 2 00:26:36.110 }, 00:26:36.110 { 00:26:36.110 "dma_device_id": "system", 00:26:36.110 "dma_device_type": 1 00:26:36.110 }, 00:26:36.110 { 00:26:36.110 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:36.110 "dma_device_type": 2 00:26:36.110 } 00:26:36.110 ], 00:26:36.110 "driver_specific": { 00:26:36.110 "raid": { 00:26:36.110 "uuid": "53334b15-7569-4888-826b-cc1339ec2186", 00:26:36.110 "strip_size_kb": 0, 00:26:36.110 "state": "online", 00:26:36.110 "raid_level": "raid1", 00:26:36.110 "superblock": true, 00:26:36.110 "num_base_bdevs": 2, 00:26:36.110 "num_base_bdevs_discovered": 2, 00:26:36.110 "num_base_bdevs_operational": 2, 00:26:36.110 "base_bdevs_list": [ 00:26:36.110 { 00:26:36.110 "name": "BaseBdev1", 00:26:36.110 "uuid": "0e2fc37f-de6e-4a79-b91a-6525da35cb9a", 00:26:36.110 "is_configured": true, 00:26:36.110 "data_offset": 256, 00:26:36.110 "data_size": 7936 00:26:36.110 }, 00:26:36.110 { 00:26:36.110 "name": "BaseBdev2", 00:26:36.110 "uuid": "b21e0c64-cd47-4ccd-a404-9a5641e62cf0", 00:26:36.110 "is_configured": true, 00:26:36.110 "data_offset": 256, 00:26:36.110 "data_size": 7936 00:26:36.110 } 00:26:36.110 ] 00:26:36.110 } 00:26:36.110 } 00:26:36.110 }' 00:26:36.110 11:37:19 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:26:36.110 11:37:19 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:26:36.110 BaseBdev2' 00:26:36.110 11:37:19 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:26:36.110 11:37:19 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:26:36.110 11:37:19 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:26:36.369 11:37:19 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:26:36.369 "name": "BaseBdev1", 00:26:36.369 "aliases": [ 00:26:36.369 "0e2fc37f-de6e-4a79-b91a-6525da35cb9a" 00:26:36.369 ], 00:26:36.369 "product_name": "Malloc disk", 00:26:36.369 "block_size": 4096, 00:26:36.369 "num_blocks": 8192, 00:26:36.369 "uuid": "0e2fc37f-de6e-4a79-b91a-6525da35cb9a", 00:26:36.369 "assigned_rate_limits": { 00:26:36.369 "rw_ios_per_sec": 0, 00:26:36.369 "rw_mbytes_per_sec": 0, 00:26:36.369 "r_mbytes_per_sec": 0, 00:26:36.369 "w_mbytes_per_sec": 0 00:26:36.369 }, 00:26:36.369 "claimed": true, 00:26:36.369 "claim_type": "exclusive_write", 00:26:36.369 "zoned": false, 00:26:36.369 "supported_io_types": { 00:26:36.369 "read": true, 00:26:36.369 "write": true, 00:26:36.369 "unmap": true, 00:26:36.369 "flush": true, 00:26:36.369 "reset": true, 00:26:36.369 "nvme_admin": false, 00:26:36.369 "nvme_io": false, 00:26:36.369 "nvme_io_md": false, 00:26:36.369 "write_zeroes": true, 00:26:36.369 "zcopy": true, 00:26:36.369 "get_zone_info": false, 00:26:36.369 "zone_management": false, 00:26:36.369 "zone_append": false, 00:26:36.369 "compare": false, 00:26:36.369 "compare_and_write": false, 00:26:36.369 "abort": true, 00:26:36.369 "seek_hole": false, 00:26:36.369 "seek_data": false, 00:26:36.369 "copy": true, 00:26:36.369 "nvme_iov_md": false 00:26:36.369 }, 00:26:36.369 "memory_domains": [ 00:26:36.369 { 00:26:36.369 "dma_device_id": "system", 00:26:36.369 "dma_device_type": 1 00:26:36.369 }, 00:26:36.369 { 00:26:36.369 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:36.369 "dma_device_type": 2 00:26:36.369 } 00:26:36.369 ], 00:26:36.369 "driver_specific": {} 00:26:36.369 }' 00:26:36.369 11:37:19 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:26:36.369 11:37:19 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:26:36.369 11:37:19 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@205 -- # [[ 4096 == 4096 ]] 00:26:36.369 11:37:19 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:26:36.627 11:37:19 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:26:36.627 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:26:36.627 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:26:36.627 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:26:36.627 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:26:36.627 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:26:36.627 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:26:36.627 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:26:36.627 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:26:36.627 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:26:36.627 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:26:36.886 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:26:36.886 "name": "BaseBdev2", 00:26:36.886 "aliases": [ 00:26:36.886 "b21e0c64-cd47-4ccd-a404-9a5641e62cf0" 00:26:36.886 ], 00:26:36.886 "product_name": "Malloc disk", 00:26:36.886 "block_size": 4096, 00:26:36.886 "num_blocks": 8192, 00:26:36.886 "uuid": "b21e0c64-cd47-4ccd-a404-9a5641e62cf0", 00:26:36.886 "assigned_rate_limits": { 00:26:36.886 "rw_ios_per_sec": 0, 00:26:36.886 "rw_mbytes_per_sec": 0, 00:26:36.886 "r_mbytes_per_sec": 0, 00:26:36.886 "w_mbytes_per_sec": 0 00:26:36.886 }, 00:26:36.886 "claimed": true, 00:26:36.886 "claim_type": "exclusive_write", 00:26:36.886 "zoned": false, 00:26:36.886 "supported_io_types": { 00:26:36.886 "read": true, 00:26:36.886 "write": true, 00:26:36.886 "unmap": true, 00:26:36.886 "flush": true, 00:26:36.886 "reset": true, 00:26:36.886 "nvme_admin": false, 00:26:36.886 "nvme_io": false, 00:26:36.886 "nvme_io_md": false, 00:26:36.886 "write_zeroes": true, 00:26:36.886 "zcopy": true, 00:26:36.886 "get_zone_info": false, 00:26:36.886 "zone_management": false, 00:26:36.886 "zone_append": false, 00:26:36.886 "compare": false, 00:26:36.886 "compare_and_write": false, 00:26:36.886 "abort": true, 00:26:36.886 "seek_hole": false, 00:26:36.886 "seek_data": false, 00:26:36.886 "copy": true, 00:26:36.886 "nvme_iov_md": false 00:26:36.886 }, 00:26:36.886 "memory_domains": [ 00:26:36.886 { 00:26:36.886 "dma_device_id": "system", 00:26:36.886 "dma_device_type": 1 00:26:36.886 }, 00:26:36.886 { 00:26:36.886 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:36.886 "dma_device_type": 2 00:26:36.886 } 00:26:36.886 ], 00:26:36.886 "driver_specific": {} 00:26:36.886 }' 00:26:36.886 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:26:37.144 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:26:37.144 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@205 -- # [[ 4096 == 4096 ]] 00:26:37.144 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:26:37.144 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:26:37.144 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:26:37.144 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:26:37.144 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:26:37.144 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:26:37.144 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:26:37.402 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:26:37.402 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:26:37.402 11:37:20 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:26:37.660 [2024-07-15 11:37:21.013218] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:26:37.660 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@275 -- # local expected_state 00:26:37.660 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@276 -- # has_redundancy raid1 00:26:37.660 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@213 -- # case $1 in 00:26:37.660 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@214 -- # return 0 00:26:37.660 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@279 -- # expected_state=online 00:26:37.660 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:26:37.660 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:26:37.660 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:26:37.660 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:37.660 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:37.660 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:26:37.660 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:37.660 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:37.660 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:37.660 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:37.660 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:37.660 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:37.918 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:37.918 "name": "Existed_Raid", 00:26:37.918 "uuid": "53334b15-7569-4888-826b-cc1339ec2186", 00:26:37.918 "strip_size_kb": 0, 00:26:37.918 "state": "online", 00:26:37.918 "raid_level": "raid1", 00:26:37.918 "superblock": true, 00:26:37.918 "num_base_bdevs": 2, 00:26:37.918 "num_base_bdevs_discovered": 1, 00:26:37.918 "num_base_bdevs_operational": 1, 00:26:37.918 "base_bdevs_list": [ 00:26:37.918 { 00:26:37.918 "name": null, 00:26:37.918 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:37.918 "is_configured": false, 00:26:37.918 "data_offset": 256, 00:26:37.918 "data_size": 7936 00:26:37.918 }, 00:26:37.918 { 00:26:37.918 "name": "BaseBdev2", 00:26:37.918 "uuid": "b21e0c64-cd47-4ccd-a404-9a5641e62cf0", 00:26:37.918 "is_configured": true, 00:26:37.918 "data_offset": 256, 00:26:37.918 "data_size": 7936 00:26:37.918 } 00:26:37.918 ] 00:26:37.918 }' 00:26:37.919 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:37.919 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:26:38.483 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:26:38.483 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:26:38.483 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:38.483 11:37:21 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:26:38.741 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:26:38.741 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:26:38.741 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:26:38.998 [2024-07-15 11:37:22.350295] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:26:38.998 [2024-07-15 11:37:22.350378] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:26:38.998 [2024-07-15 11:37:22.363142] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:38.998 [2024-07-15 11:37:22.363179] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:38.998 [2024-07-15 11:37:22.363191] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x19c7000 name Existed_Raid, state offline 00:26:38.998 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:26:38.998 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:26:38.998 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:38.998 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:26:39.257 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:26:39.257 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:26:39.257 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@299 -- # '[' 2 -gt 2 ']' 00:26:39.257 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@341 -- # killprocess 1000558 00:26:39.257 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@948 -- # '[' -z 1000558 ']' 00:26:39.257 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@952 -- # kill -0 1000558 00:26:39.257 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@953 -- # uname 00:26:39.257 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:26:39.257 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1000558 00:26:39.257 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:26:39.257 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:26:39.257 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1000558' 00:26:39.257 killing process with pid 1000558 00:26:39.257 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@967 -- # kill 1000558 00:26:39.257 [2024-07-15 11:37:22.681006] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:26:39.257 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@972 -- # wait 1000558 00:26:39.257 [2024-07-15 11:37:22.681986] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:26:39.514 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@343 -- # return 0 00:26:39.514 00:26:39.514 real 0m10.370s 00:26:39.514 user 0m18.482s 00:26:39.514 sys 0m1.891s 00:26:39.514 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1124 -- # xtrace_disable 00:26:39.514 11:37:22 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:26:39.514 ************************************ 00:26:39.514 END TEST raid_state_function_test_sb_4k 00:26:39.514 ************************************ 00:26:39.514 11:37:22 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:26:39.514 11:37:22 bdev_raid -- bdev/bdev_raid.sh@899 -- # run_test raid_superblock_test_4k raid_superblock_test raid1 2 00:26:39.514 11:37:22 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:26:39.514 11:37:22 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:26:39.514 11:37:22 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:26:39.514 ************************************ 00:26:39.514 START TEST raid_superblock_test_4k 00:26:39.514 ************************************ 00:26:39.514 11:37:22 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1123 -- # raid_superblock_test raid1 2 00:26:39.514 11:37:22 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@392 -- # local raid_level=raid1 00:26:39.514 11:37:22 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@393 -- # local num_base_bdevs=2 00:26:39.514 11:37:22 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@394 -- # base_bdevs_malloc=() 00:26:39.514 11:37:22 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@394 -- # local base_bdevs_malloc 00:26:39.514 11:37:22 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # base_bdevs_pt=() 00:26:39.514 11:37:22 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # local base_bdevs_pt 00:26:39.514 11:37:22 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt_uuid=() 00:26:39.514 11:37:22 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt_uuid 00:26:39.514 11:37:22 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # local raid_bdev_name=raid_bdev1 00:26:39.514 11:37:22 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@398 -- # local strip_size 00:26:39.514 11:37:22 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@399 -- # local strip_size_create_arg 00:26:39.514 11:37:22 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@400 -- # local raid_bdev_uuid 00:26:39.514 11:37:22 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@401 -- # local raid_bdev 00:26:39.514 11:37:22 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@403 -- # '[' raid1 '!=' raid1 ']' 00:26:39.514 11:37:22 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@407 -- # strip_size=0 00:26:39.514 11:37:22 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@411 -- # raid_pid=1002115 00:26:39.515 11:37:22 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@412 -- # waitforlisten 1002115 /var/tmp/spdk-raid.sock 00:26:39.515 11:37:22 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@410 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -L bdev_raid 00:26:39.515 11:37:22 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@829 -- # '[' -z 1002115 ']' 00:26:39.515 11:37:22 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:26:39.515 11:37:22 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@834 -- # local max_retries=100 00:26:39.515 11:37:22 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:26:39.515 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:26:39.515 11:37:22 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@838 -- # xtrace_disable 00:26:39.515 11:37:22 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:26:39.515 [2024-07-15 11:37:23.047701] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:26:39.515 [2024-07-15 11:37:23.047765] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1002115 ] 00:26:39.775 [2024-07-15 11:37:23.163064] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:39.775 [2024-07-15 11:37:23.266195] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:26:39.775 [2024-07-15 11:37:23.326066] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:39.775 [2024-07-15 11:37:23.326099] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:40.340 11:37:23 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:26:40.340 11:37:23 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@862 -- # return 0 00:26:40.340 11:37:23 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@415 -- # (( i = 1 )) 00:26:40.340 11:37:23 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:26:40.340 11:37:23 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc1 00:26:40.340 11:37:23 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt1 00:26:40.340 11:37:23 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:26:40.340 11:37:23 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:26:40.340 11:37:23 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:26:40.340 11:37:23 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:26:40.340 11:37:23 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -b malloc1 00:26:40.641 malloc1 00:26:40.641 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:26:40.906 [2024-07-15 11:37:24.327253] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:26:40.906 [2024-07-15 11:37:24.327305] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:40.906 [2024-07-15 11:37:24.327327] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x976570 00:26:40.906 [2024-07-15 11:37:24.327339] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:40.906 [2024-07-15 11:37:24.329002] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:40.906 [2024-07-15 11:37:24.329032] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:26:40.906 pt1 00:26:40.906 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:26:40.906 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:26:40.906 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc2 00:26:40.906 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt2 00:26:40.906 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:26:40.906 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:26:40.906 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:26:40.906 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:26:40.906 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -b malloc2 00:26:41.164 malloc2 00:26:41.164 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:26:41.164 [2024-07-15 11:37:24.680992] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:26:41.164 [2024-07-15 11:37:24.681039] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:41.164 [2024-07-15 11:37:24.681055] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x977970 00:26:41.164 [2024-07-15 11:37:24.681068] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:41.164 [2024-07-15 11:37:24.682529] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:41.164 [2024-07-15 11:37:24.682557] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:26:41.164 pt2 00:26:41.164 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:26:41.164 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:26:41.164 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@429 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'pt1 pt2' -n raid_bdev1 -s 00:26:41.422 [2024-07-15 11:37:24.853452] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:26:41.422 [2024-07-15 11:37:24.854618] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:26:41.422 [2024-07-15 11:37:24.854760] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xb1a270 00:26:41.422 [2024-07-15 11:37:24.854773] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:26:41.422 [2024-07-15 11:37:24.854970] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x96e0e0 00:26:41.422 [2024-07-15 11:37:24.855109] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xb1a270 00:26:41.422 [2024-07-15 11:37:24.855119] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0xb1a270 00:26:41.422 [2024-07-15 11:37:24.855209] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:41.422 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@430 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:26:41.422 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:26:41.422 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:26:41.422 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:41.422 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:41.422 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:26:41.422 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:41.422 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:41.422 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:41.422 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:41.422 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:41.422 11:37:24 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:41.680 11:37:25 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:41.680 "name": "raid_bdev1", 00:26:41.680 "uuid": "888b7ceb-8e13-4d3e-bfd8-b6bc60ed836f", 00:26:41.680 "strip_size_kb": 0, 00:26:41.680 "state": "online", 00:26:41.680 "raid_level": "raid1", 00:26:41.680 "superblock": true, 00:26:41.680 "num_base_bdevs": 2, 00:26:41.680 "num_base_bdevs_discovered": 2, 00:26:41.680 "num_base_bdevs_operational": 2, 00:26:41.680 "base_bdevs_list": [ 00:26:41.680 { 00:26:41.680 "name": "pt1", 00:26:41.680 "uuid": "00000000-0000-0000-0000-000000000001", 00:26:41.680 "is_configured": true, 00:26:41.680 "data_offset": 256, 00:26:41.680 "data_size": 7936 00:26:41.680 }, 00:26:41.680 { 00:26:41.680 "name": "pt2", 00:26:41.680 "uuid": "00000000-0000-0000-0000-000000000002", 00:26:41.680 "is_configured": true, 00:26:41.680 "data_offset": 256, 00:26:41.680 "data_size": 7936 00:26:41.680 } 00:26:41.680 ] 00:26:41.680 }' 00:26:41.680 11:37:25 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:41.680 11:37:25 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:26:42.246 11:37:25 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_properties raid_bdev1 00:26:42.246 11:37:25 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:26:42.246 11:37:25 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:26:42.246 11:37:25 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:26:42.246 11:37:25 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:26:42.246 11:37:25 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@198 -- # local name 00:26:42.246 11:37:25 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:26:42.246 11:37:25 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:26:42.503 [2024-07-15 11:37:25.964638] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:42.503 11:37:25 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:26:42.503 "name": "raid_bdev1", 00:26:42.503 "aliases": [ 00:26:42.503 "888b7ceb-8e13-4d3e-bfd8-b6bc60ed836f" 00:26:42.503 ], 00:26:42.503 "product_name": "Raid Volume", 00:26:42.503 "block_size": 4096, 00:26:42.503 "num_blocks": 7936, 00:26:42.503 "uuid": "888b7ceb-8e13-4d3e-bfd8-b6bc60ed836f", 00:26:42.503 "assigned_rate_limits": { 00:26:42.503 "rw_ios_per_sec": 0, 00:26:42.503 "rw_mbytes_per_sec": 0, 00:26:42.503 "r_mbytes_per_sec": 0, 00:26:42.503 "w_mbytes_per_sec": 0 00:26:42.503 }, 00:26:42.503 "claimed": false, 00:26:42.503 "zoned": false, 00:26:42.503 "supported_io_types": { 00:26:42.503 "read": true, 00:26:42.503 "write": true, 00:26:42.503 "unmap": false, 00:26:42.503 "flush": false, 00:26:42.503 "reset": true, 00:26:42.503 "nvme_admin": false, 00:26:42.503 "nvme_io": false, 00:26:42.503 "nvme_io_md": false, 00:26:42.503 "write_zeroes": true, 00:26:42.503 "zcopy": false, 00:26:42.503 "get_zone_info": false, 00:26:42.503 "zone_management": false, 00:26:42.503 "zone_append": false, 00:26:42.503 "compare": false, 00:26:42.503 "compare_and_write": false, 00:26:42.503 "abort": false, 00:26:42.503 "seek_hole": false, 00:26:42.503 "seek_data": false, 00:26:42.503 "copy": false, 00:26:42.503 "nvme_iov_md": false 00:26:42.503 }, 00:26:42.503 "memory_domains": [ 00:26:42.503 { 00:26:42.503 "dma_device_id": "system", 00:26:42.503 "dma_device_type": 1 00:26:42.503 }, 00:26:42.503 { 00:26:42.503 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:42.503 "dma_device_type": 2 00:26:42.503 }, 00:26:42.503 { 00:26:42.503 "dma_device_id": "system", 00:26:42.503 "dma_device_type": 1 00:26:42.503 }, 00:26:42.503 { 00:26:42.503 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:42.503 "dma_device_type": 2 00:26:42.503 } 00:26:42.503 ], 00:26:42.503 "driver_specific": { 00:26:42.503 "raid": { 00:26:42.503 "uuid": "888b7ceb-8e13-4d3e-bfd8-b6bc60ed836f", 00:26:42.503 "strip_size_kb": 0, 00:26:42.503 "state": "online", 00:26:42.503 "raid_level": "raid1", 00:26:42.503 "superblock": true, 00:26:42.503 "num_base_bdevs": 2, 00:26:42.503 "num_base_bdevs_discovered": 2, 00:26:42.503 "num_base_bdevs_operational": 2, 00:26:42.503 "base_bdevs_list": [ 00:26:42.503 { 00:26:42.503 "name": "pt1", 00:26:42.503 "uuid": "00000000-0000-0000-0000-000000000001", 00:26:42.503 "is_configured": true, 00:26:42.503 "data_offset": 256, 00:26:42.503 "data_size": 7936 00:26:42.503 }, 00:26:42.503 { 00:26:42.503 "name": "pt2", 00:26:42.503 "uuid": "00000000-0000-0000-0000-000000000002", 00:26:42.503 "is_configured": true, 00:26:42.503 "data_offset": 256, 00:26:42.503 "data_size": 7936 00:26:42.503 } 00:26:42.503 ] 00:26:42.503 } 00:26:42.503 } 00:26:42.503 }' 00:26:42.503 11:37:25 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:26:42.503 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:26:42.503 pt2' 00:26:42.503 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:26:42.503 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:26:42.503 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:26:42.761 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:26:42.761 "name": "pt1", 00:26:42.761 "aliases": [ 00:26:42.761 "00000000-0000-0000-0000-000000000001" 00:26:42.761 ], 00:26:42.761 "product_name": "passthru", 00:26:42.761 "block_size": 4096, 00:26:42.761 "num_blocks": 8192, 00:26:42.761 "uuid": "00000000-0000-0000-0000-000000000001", 00:26:42.761 "assigned_rate_limits": { 00:26:42.761 "rw_ios_per_sec": 0, 00:26:42.761 "rw_mbytes_per_sec": 0, 00:26:42.761 "r_mbytes_per_sec": 0, 00:26:42.761 "w_mbytes_per_sec": 0 00:26:42.761 }, 00:26:42.761 "claimed": true, 00:26:42.761 "claim_type": "exclusive_write", 00:26:42.761 "zoned": false, 00:26:42.761 "supported_io_types": { 00:26:42.761 "read": true, 00:26:42.761 "write": true, 00:26:42.761 "unmap": true, 00:26:42.761 "flush": true, 00:26:42.761 "reset": true, 00:26:42.761 "nvme_admin": false, 00:26:42.761 "nvme_io": false, 00:26:42.761 "nvme_io_md": false, 00:26:42.761 "write_zeroes": true, 00:26:42.761 "zcopy": true, 00:26:42.761 "get_zone_info": false, 00:26:42.761 "zone_management": false, 00:26:42.761 "zone_append": false, 00:26:42.761 "compare": false, 00:26:42.761 "compare_and_write": false, 00:26:42.761 "abort": true, 00:26:42.761 "seek_hole": false, 00:26:42.761 "seek_data": false, 00:26:42.761 "copy": true, 00:26:42.761 "nvme_iov_md": false 00:26:42.761 }, 00:26:42.761 "memory_domains": [ 00:26:42.761 { 00:26:42.761 "dma_device_id": "system", 00:26:42.761 "dma_device_type": 1 00:26:42.761 }, 00:26:42.761 { 00:26:42.761 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:42.761 "dma_device_type": 2 00:26:42.761 } 00:26:42.761 ], 00:26:42.761 "driver_specific": { 00:26:42.761 "passthru": { 00:26:42.761 "name": "pt1", 00:26:42.761 "base_bdev_name": "malloc1" 00:26:42.761 } 00:26:42.761 } 00:26:42.761 }' 00:26:42.761 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:26:42.761 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:26:42.761 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@205 -- # [[ 4096 == 4096 ]] 00:26:43.018 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:26:43.018 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:26:43.018 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:26:43.018 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:26:43.018 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:26:43.018 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:26:43.018 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:26:43.018 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:26:43.275 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:26:43.275 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:26:43.275 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:26:43.275 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:26:43.533 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:26:43.533 "name": "pt2", 00:26:43.533 "aliases": [ 00:26:43.533 "00000000-0000-0000-0000-000000000002" 00:26:43.533 ], 00:26:43.533 "product_name": "passthru", 00:26:43.533 "block_size": 4096, 00:26:43.533 "num_blocks": 8192, 00:26:43.533 "uuid": "00000000-0000-0000-0000-000000000002", 00:26:43.533 "assigned_rate_limits": { 00:26:43.533 "rw_ios_per_sec": 0, 00:26:43.533 "rw_mbytes_per_sec": 0, 00:26:43.533 "r_mbytes_per_sec": 0, 00:26:43.533 "w_mbytes_per_sec": 0 00:26:43.533 }, 00:26:43.533 "claimed": true, 00:26:43.533 "claim_type": "exclusive_write", 00:26:43.533 "zoned": false, 00:26:43.533 "supported_io_types": { 00:26:43.533 "read": true, 00:26:43.533 "write": true, 00:26:43.533 "unmap": true, 00:26:43.533 "flush": true, 00:26:43.533 "reset": true, 00:26:43.533 "nvme_admin": false, 00:26:43.533 "nvme_io": false, 00:26:43.533 "nvme_io_md": false, 00:26:43.533 "write_zeroes": true, 00:26:43.533 "zcopy": true, 00:26:43.533 "get_zone_info": false, 00:26:43.533 "zone_management": false, 00:26:43.533 "zone_append": false, 00:26:43.533 "compare": false, 00:26:43.533 "compare_and_write": false, 00:26:43.533 "abort": true, 00:26:43.533 "seek_hole": false, 00:26:43.533 "seek_data": false, 00:26:43.533 "copy": true, 00:26:43.533 "nvme_iov_md": false 00:26:43.533 }, 00:26:43.533 "memory_domains": [ 00:26:43.533 { 00:26:43.533 "dma_device_id": "system", 00:26:43.533 "dma_device_type": 1 00:26:43.533 }, 00:26:43.533 { 00:26:43.533 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:43.533 "dma_device_type": 2 00:26:43.533 } 00:26:43.533 ], 00:26:43.533 "driver_specific": { 00:26:43.533 "passthru": { 00:26:43.533 "name": "pt2", 00:26:43.533 "base_bdev_name": "malloc2" 00:26:43.533 } 00:26:43.533 } 00:26:43.533 }' 00:26:43.533 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:26:43.533 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:26:43.533 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@205 -- # [[ 4096 == 4096 ]] 00:26:43.533 11:37:26 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:26:43.533 11:37:27 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:26:43.533 11:37:27 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:26:43.533 11:37:27 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:26:43.533 11:37:27 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:26:43.791 11:37:27 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:26:43.791 11:37:27 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:26:43.791 11:37:27 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:26:43.791 11:37:27 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:26:43.791 11:37:27 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@434 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:26:43.791 11:37:27 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@434 -- # jq -r '.[] | .uuid' 00:26:44.049 [2024-07-15 11:37:27.456601] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:44.049 11:37:27 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@434 -- # raid_bdev_uuid=888b7ceb-8e13-4d3e-bfd8-b6bc60ed836f 00:26:44.049 11:37:27 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # '[' -z 888b7ceb-8e13-4d3e-bfd8-b6bc60ed836f ']' 00:26:44.049 11:37:27 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@440 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:26:44.306 [2024-07-15 11:37:27.700991] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:26:44.306 [2024-07-15 11:37:27.701011] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:26:44.306 [2024-07-15 11:37:27.701067] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:44.306 [2024-07-15 11:37:27.701123] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:44.306 [2024-07-15 11:37:27.701135] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xb1a270 name raid_bdev1, state offline 00:26:44.306 11:37:27 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@441 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:44.306 11:37:27 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@441 -- # jq -r '.[]' 00:26:44.562 11:37:27 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@441 -- # raid_bdev= 00:26:44.562 11:37:27 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # '[' -n '' ']' 00:26:44.562 11:37:27 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:26:44.562 11:37:27 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt1 00:26:44.818 11:37:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:26:44.818 11:37:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:26:45.074 11:37:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@450 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs 00:26:45.074 11:37:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@450 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:26:45.330 11:37:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@450 -- # '[' false == true ']' 00:26:45.330 11:37:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@456 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'malloc1 malloc2' -n raid_bdev1 00:26:45.330 11:37:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@648 -- # local es=0 00:26:45.330 11:37:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'malloc1 malloc2' -n raid_bdev1 00:26:45.330 11:37:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:26:45.330 11:37:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:26:45.330 11:37:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:26:45.330 11:37:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:26:45.330 11:37:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:26:45.330 11:37:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:26:45.331 11:37:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:26:45.331 11:37:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py ]] 00:26:45.331 11:37:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'malloc1 malloc2' -n raid_bdev1 00:26:45.331 [2024-07-15 11:37:28.916156] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:26:45.331 [2024-07-15 11:37:28.917556] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:26:45.331 [2024-07-15 11:37:28.917614] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:26:45.331 [2024-07-15 11:37:28.917655] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:26:45.331 [2024-07-15 11:37:28.917674] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:26:45.331 [2024-07-15 11:37:28.917684] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xb19ff0 name raid_bdev1, state configuring 00:26:45.331 request: 00:26:45.331 { 00:26:45.331 "name": "raid_bdev1", 00:26:45.331 "raid_level": "raid1", 00:26:45.331 "base_bdevs": [ 00:26:45.331 "malloc1", 00:26:45.331 "malloc2" 00:26:45.331 ], 00:26:45.331 "superblock": false, 00:26:45.331 "method": "bdev_raid_create", 00:26:45.331 "req_id": 1 00:26:45.331 } 00:26:45.331 Got JSON-RPC error response 00:26:45.331 response: 00:26:45.331 { 00:26:45.331 "code": -17, 00:26:45.331 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:26:45.331 } 00:26:45.586 11:37:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@651 -- # es=1 00:26:45.586 11:37:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:26:45.586 11:37:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:26:45.586 11:37:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:26:45.586 11:37:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@458 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:45.587 11:37:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@458 -- # jq -r '.[]' 00:26:45.587 11:37:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@458 -- # raid_bdev= 00:26:45.587 11:37:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # '[' -n '' ']' 00:26:45.587 11:37:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@464 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:26:46.148 [2024-07-15 11:37:29.654042] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:26:46.148 [2024-07-15 11:37:29.654092] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:46.148 [2024-07-15 11:37:29.654115] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x9767a0 00:26:46.148 [2024-07-15 11:37:29.654128] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:46.148 [2024-07-15 11:37:29.655814] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:46.148 [2024-07-15 11:37:29.655845] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:26:46.148 [2024-07-15 11:37:29.655918] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:26:46.148 [2024-07-15 11:37:29.655958] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:26:46.148 pt1 00:26:46.148 11:37:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@467 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:26:46.148 11:37:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:26:46.148 11:37:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:26:46.148 11:37:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:46.148 11:37:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:46.148 11:37:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:26:46.148 11:37:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:46.148 11:37:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:46.148 11:37:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:46.148 11:37:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:46.148 11:37:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:46.148 11:37:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:46.404 11:37:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:46.404 "name": "raid_bdev1", 00:26:46.404 "uuid": "888b7ceb-8e13-4d3e-bfd8-b6bc60ed836f", 00:26:46.404 "strip_size_kb": 0, 00:26:46.404 "state": "configuring", 00:26:46.404 "raid_level": "raid1", 00:26:46.404 "superblock": true, 00:26:46.404 "num_base_bdevs": 2, 00:26:46.404 "num_base_bdevs_discovered": 1, 00:26:46.404 "num_base_bdevs_operational": 2, 00:26:46.404 "base_bdevs_list": [ 00:26:46.404 { 00:26:46.404 "name": "pt1", 00:26:46.404 "uuid": "00000000-0000-0000-0000-000000000001", 00:26:46.404 "is_configured": true, 00:26:46.404 "data_offset": 256, 00:26:46.404 "data_size": 7936 00:26:46.404 }, 00:26:46.404 { 00:26:46.404 "name": null, 00:26:46.404 "uuid": "00000000-0000-0000-0000-000000000002", 00:26:46.404 "is_configured": false, 00:26:46.404 "data_offset": 256, 00:26:46.404 "data_size": 7936 00:26:46.404 } 00:26:46.404 ] 00:26:46.404 }' 00:26:46.404 11:37:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:46.404 11:37:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:26:46.965 11:37:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@469 -- # '[' 2 -gt 2 ']' 00:26:46.965 11:37:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@477 -- # (( i = 1 )) 00:26:46.965 11:37:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:26:46.965 11:37:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:26:47.220 [2024-07-15 11:37:30.756996] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:26:47.220 [2024-07-15 11:37:30.757050] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:47.220 [2024-07-15 11:37:30.757069] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xb0e6f0 00:26:47.220 [2024-07-15 11:37:30.757082] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:47.220 [2024-07-15 11:37:30.757457] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:47.220 [2024-07-15 11:37:30.757476] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:26:47.220 [2024-07-15 11:37:30.757545] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:26:47.221 [2024-07-15 11:37:30.757566] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:26:47.221 [2024-07-15 11:37:30.757667] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xb0f590 00:26:47.221 [2024-07-15 11:37:30.757678] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:26:47.221 [2024-07-15 11:37:30.757849] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x970540 00:26:47.221 [2024-07-15 11:37:30.757984] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xb0f590 00:26:47.221 [2024-07-15 11:37:30.757995] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0xb0f590 00:26:47.221 [2024-07-15 11:37:30.758096] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:47.221 pt2 00:26:47.221 11:37:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:26:47.221 11:37:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:26:47.221 11:37:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@482 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:26:47.221 11:37:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:26:47.221 11:37:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:26:47.221 11:37:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:47.221 11:37:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:47.221 11:37:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:26:47.221 11:37:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:47.221 11:37:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:47.221 11:37:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:47.221 11:37:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:47.221 11:37:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:47.221 11:37:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:47.487 11:37:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:47.487 "name": "raid_bdev1", 00:26:47.487 "uuid": "888b7ceb-8e13-4d3e-bfd8-b6bc60ed836f", 00:26:47.487 "strip_size_kb": 0, 00:26:47.487 "state": "online", 00:26:47.487 "raid_level": "raid1", 00:26:47.487 "superblock": true, 00:26:47.487 "num_base_bdevs": 2, 00:26:47.487 "num_base_bdevs_discovered": 2, 00:26:47.487 "num_base_bdevs_operational": 2, 00:26:47.487 "base_bdevs_list": [ 00:26:47.487 { 00:26:47.487 "name": "pt1", 00:26:47.487 "uuid": "00000000-0000-0000-0000-000000000001", 00:26:47.487 "is_configured": true, 00:26:47.487 "data_offset": 256, 00:26:47.487 "data_size": 7936 00:26:47.487 }, 00:26:47.487 { 00:26:47.487 "name": "pt2", 00:26:47.487 "uuid": "00000000-0000-0000-0000-000000000002", 00:26:47.487 "is_configured": true, 00:26:47.487 "data_offset": 256, 00:26:47.487 "data_size": 7936 00:26:47.487 } 00:26:47.487 ] 00:26:47.487 }' 00:26:47.487 11:37:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:47.487 11:37:31 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:26:48.054 11:37:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_properties raid_bdev1 00:26:48.054 11:37:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:26:48.054 11:37:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:26:48.054 11:37:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:26:48.054 11:37:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:26:48.054 11:37:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@198 -- # local name 00:26:48.054 11:37:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:26:48.054 11:37:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:26:48.310 [2024-07-15 11:37:31.844130] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:48.310 11:37:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:26:48.310 "name": "raid_bdev1", 00:26:48.310 "aliases": [ 00:26:48.310 "888b7ceb-8e13-4d3e-bfd8-b6bc60ed836f" 00:26:48.310 ], 00:26:48.310 "product_name": "Raid Volume", 00:26:48.310 "block_size": 4096, 00:26:48.310 "num_blocks": 7936, 00:26:48.310 "uuid": "888b7ceb-8e13-4d3e-bfd8-b6bc60ed836f", 00:26:48.310 "assigned_rate_limits": { 00:26:48.310 "rw_ios_per_sec": 0, 00:26:48.310 "rw_mbytes_per_sec": 0, 00:26:48.310 "r_mbytes_per_sec": 0, 00:26:48.310 "w_mbytes_per_sec": 0 00:26:48.310 }, 00:26:48.310 "claimed": false, 00:26:48.310 "zoned": false, 00:26:48.310 "supported_io_types": { 00:26:48.310 "read": true, 00:26:48.310 "write": true, 00:26:48.310 "unmap": false, 00:26:48.310 "flush": false, 00:26:48.310 "reset": true, 00:26:48.310 "nvme_admin": false, 00:26:48.310 "nvme_io": false, 00:26:48.310 "nvme_io_md": false, 00:26:48.310 "write_zeroes": true, 00:26:48.310 "zcopy": false, 00:26:48.310 "get_zone_info": false, 00:26:48.310 "zone_management": false, 00:26:48.310 "zone_append": false, 00:26:48.310 "compare": false, 00:26:48.310 "compare_and_write": false, 00:26:48.310 "abort": false, 00:26:48.310 "seek_hole": false, 00:26:48.310 "seek_data": false, 00:26:48.310 "copy": false, 00:26:48.310 "nvme_iov_md": false 00:26:48.310 }, 00:26:48.310 "memory_domains": [ 00:26:48.310 { 00:26:48.310 "dma_device_id": "system", 00:26:48.310 "dma_device_type": 1 00:26:48.310 }, 00:26:48.310 { 00:26:48.310 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:48.310 "dma_device_type": 2 00:26:48.310 }, 00:26:48.310 { 00:26:48.310 "dma_device_id": "system", 00:26:48.310 "dma_device_type": 1 00:26:48.310 }, 00:26:48.310 { 00:26:48.310 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:48.310 "dma_device_type": 2 00:26:48.310 } 00:26:48.310 ], 00:26:48.310 "driver_specific": { 00:26:48.310 "raid": { 00:26:48.311 "uuid": "888b7ceb-8e13-4d3e-bfd8-b6bc60ed836f", 00:26:48.311 "strip_size_kb": 0, 00:26:48.311 "state": "online", 00:26:48.311 "raid_level": "raid1", 00:26:48.311 "superblock": true, 00:26:48.311 "num_base_bdevs": 2, 00:26:48.311 "num_base_bdevs_discovered": 2, 00:26:48.311 "num_base_bdevs_operational": 2, 00:26:48.311 "base_bdevs_list": [ 00:26:48.311 { 00:26:48.311 "name": "pt1", 00:26:48.311 "uuid": "00000000-0000-0000-0000-000000000001", 00:26:48.311 "is_configured": true, 00:26:48.311 "data_offset": 256, 00:26:48.311 "data_size": 7936 00:26:48.311 }, 00:26:48.311 { 00:26:48.311 "name": "pt2", 00:26:48.311 "uuid": "00000000-0000-0000-0000-000000000002", 00:26:48.311 "is_configured": true, 00:26:48.311 "data_offset": 256, 00:26:48.311 "data_size": 7936 00:26:48.311 } 00:26:48.311 ] 00:26:48.311 } 00:26:48.311 } 00:26:48.311 }' 00:26:48.311 11:37:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:26:48.567 11:37:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:26:48.567 pt2' 00:26:48.567 11:37:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:26:48.567 11:37:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:26:48.567 11:37:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:26:48.567 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:26:48.567 "name": "pt1", 00:26:48.567 "aliases": [ 00:26:48.567 "00000000-0000-0000-0000-000000000001" 00:26:48.567 ], 00:26:48.567 "product_name": "passthru", 00:26:48.567 "block_size": 4096, 00:26:48.567 "num_blocks": 8192, 00:26:48.567 "uuid": "00000000-0000-0000-0000-000000000001", 00:26:48.567 "assigned_rate_limits": { 00:26:48.567 "rw_ios_per_sec": 0, 00:26:48.567 "rw_mbytes_per_sec": 0, 00:26:48.567 "r_mbytes_per_sec": 0, 00:26:48.567 "w_mbytes_per_sec": 0 00:26:48.567 }, 00:26:48.567 "claimed": true, 00:26:48.567 "claim_type": "exclusive_write", 00:26:48.567 "zoned": false, 00:26:48.567 "supported_io_types": { 00:26:48.567 "read": true, 00:26:48.567 "write": true, 00:26:48.567 "unmap": true, 00:26:48.567 "flush": true, 00:26:48.567 "reset": true, 00:26:48.567 "nvme_admin": false, 00:26:48.567 "nvme_io": false, 00:26:48.567 "nvme_io_md": false, 00:26:48.567 "write_zeroes": true, 00:26:48.567 "zcopy": true, 00:26:48.567 "get_zone_info": false, 00:26:48.567 "zone_management": false, 00:26:48.567 "zone_append": false, 00:26:48.567 "compare": false, 00:26:48.567 "compare_and_write": false, 00:26:48.567 "abort": true, 00:26:48.567 "seek_hole": false, 00:26:48.567 "seek_data": false, 00:26:48.567 "copy": true, 00:26:48.567 "nvme_iov_md": false 00:26:48.567 }, 00:26:48.567 "memory_domains": [ 00:26:48.567 { 00:26:48.567 "dma_device_id": "system", 00:26:48.567 "dma_device_type": 1 00:26:48.567 }, 00:26:48.567 { 00:26:48.567 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:48.567 "dma_device_type": 2 00:26:48.567 } 00:26:48.567 ], 00:26:48.567 "driver_specific": { 00:26:48.567 "passthru": { 00:26:48.567 "name": "pt1", 00:26:48.567 "base_bdev_name": "malloc1" 00:26:48.567 } 00:26:48.567 } 00:26:48.567 }' 00:26:48.823 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:26:48.823 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:26:48.824 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@205 -- # [[ 4096 == 4096 ]] 00:26:48.824 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:26:48.824 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:26:48.824 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:26:48.824 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:26:48.824 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:26:49.080 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:26:49.080 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:26:49.080 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:26:49.080 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:26:49.080 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:26:49.080 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:26:49.080 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:26:49.337 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:26:49.337 "name": "pt2", 00:26:49.337 "aliases": [ 00:26:49.337 "00000000-0000-0000-0000-000000000002" 00:26:49.337 ], 00:26:49.337 "product_name": "passthru", 00:26:49.337 "block_size": 4096, 00:26:49.337 "num_blocks": 8192, 00:26:49.337 "uuid": "00000000-0000-0000-0000-000000000002", 00:26:49.337 "assigned_rate_limits": { 00:26:49.337 "rw_ios_per_sec": 0, 00:26:49.337 "rw_mbytes_per_sec": 0, 00:26:49.337 "r_mbytes_per_sec": 0, 00:26:49.337 "w_mbytes_per_sec": 0 00:26:49.337 }, 00:26:49.337 "claimed": true, 00:26:49.337 "claim_type": "exclusive_write", 00:26:49.337 "zoned": false, 00:26:49.337 "supported_io_types": { 00:26:49.337 "read": true, 00:26:49.337 "write": true, 00:26:49.337 "unmap": true, 00:26:49.337 "flush": true, 00:26:49.337 "reset": true, 00:26:49.337 "nvme_admin": false, 00:26:49.337 "nvme_io": false, 00:26:49.337 "nvme_io_md": false, 00:26:49.337 "write_zeroes": true, 00:26:49.337 "zcopy": true, 00:26:49.337 "get_zone_info": false, 00:26:49.337 "zone_management": false, 00:26:49.337 "zone_append": false, 00:26:49.337 "compare": false, 00:26:49.337 "compare_and_write": false, 00:26:49.337 "abort": true, 00:26:49.337 "seek_hole": false, 00:26:49.337 "seek_data": false, 00:26:49.337 "copy": true, 00:26:49.337 "nvme_iov_md": false 00:26:49.337 }, 00:26:49.337 "memory_domains": [ 00:26:49.337 { 00:26:49.337 "dma_device_id": "system", 00:26:49.337 "dma_device_type": 1 00:26:49.337 }, 00:26:49.338 { 00:26:49.338 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:49.338 "dma_device_type": 2 00:26:49.338 } 00:26:49.338 ], 00:26:49.338 "driver_specific": { 00:26:49.338 "passthru": { 00:26:49.338 "name": "pt2", 00:26:49.338 "base_bdev_name": "malloc2" 00:26:49.338 } 00:26:49.338 } 00:26:49.338 }' 00:26:49.338 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:26:49.338 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:26:49.338 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@205 -- # [[ 4096 == 4096 ]] 00:26:49.338 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:26:49.338 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:26:49.595 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@206 -- # [[ null == null ]] 00:26:49.595 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:26:49.595 11:37:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:26:49.595 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@207 -- # [[ null == null ]] 00:26:49.595 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:26:49.595 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:26:49.595 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@208 -- # [[ null == null ]] 00:26:49.595 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@486 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:26:49.595 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@486 -- # jq -r '.[] | .uuid' 00:26:49.935 [2024-07-15 11:37:33.340102] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:49.935 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@486 -- # '[' 888b7ceb-8e13-4d3e-bfd8-b6bc60ed836f '!=' 888b7ceb-8e13-4d3e-bfd8-b6bc60ed836f ']' 00:26:49.935 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@490 -- # has_redundancy raid1 00:26:49.935 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@213 -- # case $1 in 00:26:49.935 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@214 -- # return 0 00:26:49.935 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@492 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt1 00:26:50.192 [2024-07-15 11:37:33.588544] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:26:50.192 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@495 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:26:50.192 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:26:50.192 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:26:50.192 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:50.192 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:50.192 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:26:50.192 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:50.192 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:50.192 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:50.192 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:50.192 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:50.192 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:50.450 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:50.450 "name": "raid_bdev1", 00:26:50.450 "uuid": "888b7ceb-8e13-4d3e-bfd8-b6bc60ed836f", 00:26:50.450 "strip_size_kb": 0, 00:26:50.450 "state": "online", 00:26:50.450 "raid_level": "raid1", 00:26:50.450 "superblock": true, 00:26:50.450 "num_base_bdevs": 2, 00:26:50.450 "num_base_bdevs_discovered": 1, 00:26:50.450 "num_base_bdevs_operational": 1, 00:26:50.450 "base_bdevs_list": [ 00:26:50.450 { 00:26:50.450 "name": null, 00:26:50.450 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:50.450 "is_configured": false, 00:26:50.450 "data_offset": 256, 00:26:50.450 "data_size": 7936 00:26:50.450 }, 00:26:50.450 { 00:26:50.450 "name": "pt2", 00:26:50.450 "uuid": "00000000-0000-0000-0000-000000000002", 00:26:50.450 "is_configured": true, 00:26:50.450 "data_offset": 256, 00:26:50.450 "data_size": 7936 00:26:50.450 } 00:26:50.450 ] 00:26:50.450 }' 00:26:50.450 11:37:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:50.450 11:37:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:26:51.015 11:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@498 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:26:51.271 [2024-07-15 11:37:34.663360] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:26:51.271 [2024-07-15 11:37:34.663386] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:26:51.272 [2024-07-15 11:37:34.663440] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:51.272 [2024-07-15 11:37:34.663483] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:51.272 [2024-07-15 11:37:34.663494] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xb0f590 name raid_bdev1, state offline 00:26:51.272 11:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@499 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:51.272 11:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@499 -- # jq -r '.[]' 00:26:51.528 11:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@499 -- # raid_bdev= 00:26:51.528 11:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # '[' -n '' ']' 00:26:51.528 11:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@505 -- # (( i = 1 )) 00:26:51.528 11:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@505 -- # (( i < num_base_bdevs )) 00:26:51.528 11:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:26:51.784 11:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@505 -- # (( i++ )) 00:26:51.784 11:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@505 -- # (( i < num_base_bdevs )) 00:26:51.784 11:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@510 -- # (( i = 1 )) 00:26:51.784 11:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@510 -- # (( i < num_base_bdevs - 1 )) 00:26:51.784 11:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@518 -- # i=1 00:26:51.784 11:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@519 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:26:52.041 [2024-07-15 11:37:35.397266] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:26:52.041 [2024-07-15 11:37:35.397315] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:52.041 [2024-07-15 11:37:35.397333] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x977160 00:26:52.041 [2024-07-15 11:37:35.397352] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:52.041 [2024-07-15 11:37:35.399058] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:52.041 [2024-07-15 11:37:35.399089] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:26:52.041 [2024-07-15 11:37:35.399159] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:26:52.041 [2024-07-15 11:37:35.399188] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:26:52.041 [2024-07-15 11:37:35.399278] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x96d380 00:26:52.041 [2024-07-15 11:37:35.399289] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:26:52.041 [2024-07-15 11:37:35.399469] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x96ea80 00:26:52.041 [2024-07-15 11:37:35.399592] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x96d380 00:26:52.041 [2024-07-15 11:37:35.399602] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x96d380 00:26:52.041 [2024-07-15 11:37:35.399701] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:52.041 pt2 00:26:52.041 11:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@522 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:26:52.041 11:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:26:52.041 11:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:26:52.041 11:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:52.041 11:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:52.041 11:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:26:52.041 11:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:52.041 11:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:52.041 11:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:52.041 11:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:52.041 11:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:52.041 11:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:52.298 11:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:52.298 "name": "raid_bdev1", 00:26:52.298 "uuid": "888b7ceb-8e13-4d3e-bfd8-b6bc60ed836f", 00:26:52.298 "strip_size_kb": 0, 00:26:52.298 "state": "online", 00:26:52.298 "raid_level": "raid1", 00:26:52.298 "superblock": true, 00:26:52.298 "num_base_bdevs": 2, 00:26:52.298 "num_base_bdevs_discovered": 1, 00:26:52.298 "num_base_bdevs_operational": 1, 00:26:52.298 "base_bdevs_list": [ 00:26:52.298 { 00:26:52.298 "name": null, 00:26:52.298 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:52.298 "is_configured": false, 00:26:52.298 "data_offset": 256, 00:26:52.298 "data_size": 7936 00:26:52.298 }, 00:26:52.298 { 00:26:52.298 "name": "pt2", 00:26:52.298 "uuid": "00000000-0000-0000-0000-000000000002", 00:26:52.298 "is_configured": true, 00:26:52.298 "data_offset": 256, 00:26:52.298 "data_size": 7936 00:26:52.298 } 00:26:52.298 ] 00:26:52.298 }' 00:26:52.298 11:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:52.298 11:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:26:52.859 11:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@525 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:26:52.859 [2024-07-15 11:37:36.415961] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:26:52.859 [2024-07-15 11:37:36.415990] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:26:52.859 [2024-07-15 11:37:36.416046] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:52.859 [2024-07-15 11:37:36.416090] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:52.859 [2024-07-15 11:37:36.416101] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x96d380 name raid_bdev1, state offline 00:26:52.860 11:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@526 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:52.860 11:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@526 -- # jq -r '.[]' 00:26:53.116 11:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@526 -- # raid_bdev= 00:26:53.116 11:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # '[' -n '' ']' 00:26:53.116 11:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@531 -- # '[' 2 -gt 2 ']' 00:26:53.116 11:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@539 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:26:53.373 [2024-07-15 11:37:36.909238] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:26:53.373 [2024-07-15 11:37:36.909290] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:53.373 [2024-07-15 11:37:36.909310] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xb19520 00:26:53.373 [2024-07-15 11:37:36.909323] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:53.373 [2024-07-15 11:37:36.910974] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:53.373 [2024-07-15 11:37:36.911002] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:26:53.373 [2024-07-15 11:37:36.911072] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:26:53.373 [2024-07-15 11:37:36.911100] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:26:53.373 [2024-07-15 11:37:36.911200] bdev_raid.c:3547:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:26:53.373 [2024-07-15 11:37:36.911213] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:26:53.373 [2024-07-15 11:37:36.911227] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x96e3f0 name raid_bdev1, state configuring 00:26:53.373 [2024-07-15 11:37:36.911251] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:26:53.373 [2024-07-15 11:37:36.911309] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x9702b0 00:26:53.373 [2024-07-15 11:37:36.911319] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:26:53.373 [2024-07-15 11:37:36.911483] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x96d350 00:26:53.373 [2024-07-15 11:37:36.911604] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x9702b0 00:26:53.373 [2024-07-15 11:37:36.911614] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x9702b0 00:26:53.373 [2024-07-15 11:37:36.911712] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:53.373 pt1 00:26:53.373 11:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@541 -- # '[' 2 -gt 2 ']' 00:26:53.373 11:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@553 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:26:53.373 11:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:26:53.373 11:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:26:53.373 11:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:53.373 11:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:53.373 11:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:26:53.373 11:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:53.373 11:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:53.373 11:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:53.373 11:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:53.373 11:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:53.373 11:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:53.630 11:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:53.630 "name": "raid_bdev1", 00:26:53.630 "uuid": "888b7ceb-8e13-4d3e-bfd8-b6bc60ed836f", 00:26:53.630 "strip_size_kb": 0, 00:26:53.630 "state": "online", 00:26:53.630 "raid_level": "raid1", 00:26:53.630 "superblock": true, 00:26:53.630 "num_base_bdevs": 2, 00:26:53.630 "num_base_bdevs_discovered": 1, 00:26:53.630 "num_base_bdevs_operational": 1, 00:26:53.630 "base_bdevs_list": [ 00:26:53.630 { 00:26:53.630 "name": null, 00:26:53.630 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:53.630 "is_configured": false, 00:26:53.630 "data_offset": 256, 00:26:53.630 "data_size": 7936 00:26:53.630 }, 00:26:53.630 { 00:26:53.630 "name": "pt2", 00:26:53.630 "uuid": "00000000-0000-0000-0000-000000000002", 00:26:53.630 "is_configured": true, 00:26:53.630 "data_offset": 256, 00:26:53.630 "data_size": 7936 00:26:53.630 } 00:26:53.630 ] 00:26:53.630 }' 00:26:53.630 11:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:53.630 11:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:26:54.192 11:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@554 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs online 00:26:54.192 11:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@554 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:26:54.448 11:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@554 -- # [[ false == \f\a\l\s\e ]] 00:26:54.448 11:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@557 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:26:54.448 11:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@557 -- # jq -r '.[] | .uuid' 00:26:54.706 [2024-07-15 11:37:38.200904] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:54.706 11:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@557 -- # '[' 888b7ceb-8e13-4d3e-bfd8-b6bc60ed836f '!=' 888b7ceb-8e13-4d3e-bfd8-b6bc60ed836f ']' 00:26:54.706 11:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@562 -- # killprocess 1002115 00:26:54.706 11:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@948 -- # '[' -z 1002115 ']' 00:26:54.706 11:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@952 -- # kill -0 1002115 00:26:54.706 11:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@953 -- # uname 00:26:54.706 11:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:26:54.706 11:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1002115 00:26:54.706 11:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:26:54.706 11:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:26:54.706 11:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1002115' 00:26:54.706 killing process with pid 1002115 00:26:54.706 11:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@967 -- # kill 1002115 00:26:54.706 [2024-07-15 11:37:38.271593] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:26:54.706 [2024-07-15 11:37:38.271652] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:54.706 [2024-07-15 11:37:38.271696] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:54.706 [2024-07-15 11:37:38.271708] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x9702b0 name raid_bdev1, state offline 00:26:54.706 11:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@972 -- # wait 1002115 00:26:54.706 [2024-07-15 11:37:38.291093] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:26:54.966 11:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@564 -- # return 0 00:26:54.966 00:26:54.966 real 0m15.526s 00:26:54.966 user 0m28.139s 00:26:54.966 sys 0m2.894s 00:26:54.966 11:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1124 -- # xtrace_disable 00:26:54.966 11:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:26:54.966 ************************************ 00:26:54.966 END TEST raid_superblock_test_4k 00:26:54.966 ************************************ 00:26:54.966 11:37:38 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:26:54.966 11:37:38 bdev_raid -- bdev/bdev_raid.sh@900 -- # '[' true = true ']' 00:26:55.226 11:37:38 bdev_raid -- bdev/bdev_raid.sh@901 -- # run_test raid_rebuild_test_sb_4k raid_rebuild_test raid1 2 true false true 00:26:55.226 11:37:38 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:26:55.226 11:37:38 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:26:55.226 11:37:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:26:55.226 ************************************ 00:26:55.226 START TEST raid_rebuild_test_sb_4k 00:26:55.226 ************************************ 00:26:55.226 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1123 -- # raid_rebuild_test raid1 2 true false true 00:26:55.226 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@568 -- # local raid_level=raid1 00:26:55.226 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@569 -- # local num_base_bdevs=2 00:26:55.226 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@570 -- # local superblock=true 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@571 -- # local background_io=false 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@572 -- # local verify=true 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@573 -- # (( i = 1 )) 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@575 -- # echo BaseBdev1 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@575 -- # echo BaseBdev2 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@573 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@573 -- # local base_bdevs 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # local raid_bdev_name=raid_bdev1 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@575 -- # local strip_size 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # local create_arg 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@577 -- # local raid_bdev_size 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@578 -- # local data_offset 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@580 -- # '[' raid1 '!=' raid1 ']' 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@588 -- # strip_size=0 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@591 -- # '[' true = true ']' 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@592 -- # create_arg+=' -s' 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@596 -- # raid_pid=1004531 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@597 -- # waitforlisten 1004531 /var/tmp/spdk-raid.sock 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@595 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@829 -- # '[' -z 1004531 ']' 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@834 -- # local max_retries=100 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:26:55.227 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@838 -- # xtrace_disable 00:26:55.227 11:37:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:26:55.227 [2024-07-15 11:37:38.664876] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:26:55.227 [2024-07-15 11:37:38.664946] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1004531 ] 00:26:55.227 I/O size of 3145728 is greater than zero copy threshold (65536). 00:26:55.227 Zero copy mechanism will not be used. 00:26:55.227 [2024-07-15 11:37:38.786699] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:55.483 [2024-07-15 11:37:38.889917] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:26:55.483 [2024-07-15 11:37:38.949229] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:55.483 [2024-07-15 11:37:38.949266] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:56.045 11:37:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:26:56.045 11:37:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@862 -- # return 0 00:26:56.045 11:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:26:56.045 11:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -b BaseBdev1_malloc 00:26:56.301 BaseBdev1_malloc 00:26:56.301 11:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:26:56.558 [2024-07-15 11:37:40.081980] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:26:56.558 [2024-07-15 11:37:40.082033] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:56.558 [2024-07-15 11:37:40.082058] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x145dd40 00:26:56.558 [2024-07-15 11:37:40.082071] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:56.558 [2024-07-15 11:37:40.083839] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:56.558 [2024-07-15 11:37:40.083868] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:26:56.558 BaseBdev1 00:26:56.558 11:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:26:56.558 11:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -b BaseBdev2_malloc 00:26:56.814 BaseBdev2_malloc 00:26:56.814 11:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:26:57.071 [2024-07-15 11:37:40.588185] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:26:57.071 [2024-07-15 11:37:40.588232] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:57.071 [2024-07-15 11:37:40.588257] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x145e860 00:26:57.071 [2024-07-15 11:37:40.588269] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:57.071 [2024-07-15 11:37:40.589808] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:57.071 [2024-07-15 11:37:40.589836] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:26:57.071 BaseBdev2 00:26:57.071 11:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@606 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -b spare_malloc 00:26:57.327 spare_malloc 00:26:57.327 11:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@607 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:26:57.584 spare_delay 00:26:57.584 11:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@608 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:26:57.841 [2024-07-15 11:37:41.326724] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:26:57.841 [2024-07-15 11:37:41.326770] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:57.841 [2024-07-15 11:37:41.326792] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x160cec0 00:26:57.841 [2024-07-15 11:37:41.326804] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:57.841 [2024-07-15 11:37:41.328423] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:57.841 [2024-07-15 11:37:41.328451] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:26:57.841 spare 00:26:57.841 11:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@611 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2' -n raid_bdev1 00:26:58.097 [2024-07-15 11:37:41.559360] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:26:58.097 [2024-07-15 11:37:41.560686] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:26:58.097 [2024-07-15 11:37:41.560855] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x160e070 00:26:58.097 [2024-07-15 11:37:41.560868] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:26:58.097 [2024-07-15 11:37:41.561076] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1607490 00:26:58.097 [2024-07-15 11:37:41.561220] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x160e070 00:26:58.097 [2024-07-15 11:37:41.561230] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x160e070 00:26:58.097 [2024-07-15 11:37:41.561331] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:58.097 11:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@612 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:26:58.097 11:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:26:58.097 11:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:26:58.097 11:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:26:58.097 11:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:26:58.097 11:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:26:58.097 11:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:26:58.097 11:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:26:58.097 11:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:26:58.097 11:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:26:58.097 11:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:58.097 11:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:58.354 11:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:26:58.354 "name": "raid_bdev1", 00:26:58.354 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:26:58.354 "strip_size_kb": 0, 00:26:58.354 "state": "online", 00:26:58.354 "raid_level": "raid1", 00:26:58.354 "superblock": true, 00:26:58.354 "num_base_bdevs": 2, 00:26:58.354 "num_base_bdevs_discovered": 2, 00:26:58.354 "num_base_bdevs_operational": 2, 00:26:58.354 "base_bdevs_list": [ 00:26:58.354 { 00:26:58.354 "name": "BaseBdev1", 00:26:58.354 "uuid": "70c04a77-d834-57c6-8e08-125f482f214f", 00:26:58.354 "is_configured": true, 00:26:58.354 "data_offset": 256, 00:26:58.354 "data_size": 7936 00:26:58.354 }, 00:26:58.354 { 00:26:58.354 "name": "BaseBdev2", 00:26:58.354 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:26:58.354 "is_configured": true, 00:26:58.354 "data_offset": 256, 00:26:58.354 "data_size": 7936 00:26:58.354 } 00:26:58.354 ] 00:26:58.354 }' 00:26:58.354 11:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:26:58.354 11:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:26:58.916 11:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@615 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:26:58.916 11:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@615 -- # jq -r '.[].num_blocks' 00:26:59.173 [2024-07-15 11:37:42.630411] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:59.173 11:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@615 -- # raid_bdev_size=7936 00:26:59.173 11:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@618 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:26:59.173 11:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@618 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:26:59.429 11:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@618 -- # data_offset=256 00:26:59.429 11:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@620 -- # '[' false = true ']' 00:26:59.430 11:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@623 -- # '[' true = true ']' 00:26:59.430 11:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@624 -- # local write_unit_size 00:26:59.430 11:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@627 -- # nbd_start_disks /var/tmp/spdk-raid.sock raid_bdev1 /dev/nbd0 00:26:59.430 11:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:26:59.430 11:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:26:59.430 11:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:26:59.430 11:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:26:59.430 11:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:26:59.430 11:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:26:59.430 11:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:26:59.430 11:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:26:59.430 11:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:26:59.686 [2024-07-15 11:37:43.123518] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1607490 00:26:59.686 /dev/nbd0 00:26:59.686 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:26:59.686 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:26:59.686 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:26:59.686 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@867 -- # local i 00:26:59.686 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:26:59.686 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:26:59.686 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:26:59.687 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # break 00:26:59.687 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:26:59.687 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:26:59.687 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:26:59.687 1+0 records in 00:26:59.687 1+0 records out 00:26:59.687 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000269031 s, 15.2 MB/s 00:26:59.687 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:26:59.687 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # size=4096 00:26:59.687 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:26:59.687 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:26:59.687 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@887 -- # return 0 00:26:59.687 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:26:59.687 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:26:59.687 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@628 -- # '[' raid1 = raid5f ']' 00:26:59.687 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@632 -- # write_unit_size=1 00:26:59.687 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@634 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:27:00.616 7936+0 records in 00:27:00.616 7936+0 records out 00:27:00.616 32505856 bytes (33 MB, 31 MiB) copied, 0.744937 s, 43.6 MB/s 00:27:00.616 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@635 -- # nbd_stop_disks /var/tmp/spdk-raid.sock /dev/nbd0 00:27:00.616 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:27:00.616 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:27:00.616 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:27:00.616 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:27:00.616 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:27:00.616 11:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd0 00:27:00.871 [2024-07-15 11:37:44.213123] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:00.871 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:27:00.871 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:27:00.871 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:27:00.871 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:27:00.871 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:27:00.871 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:27:00.871 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:27:00.871 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:27:00.871 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@639 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev1 00:27:00.871 [2024-07-15 11:37:44.453802] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:27:01.127 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@642 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:01.127 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:27:01.127 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:27:01.127 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:01.127 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:01.127 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:27:01.127 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:01.127 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:01.127 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:01.127 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:01.127 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:01.127 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:01.382 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:01.382 "name": "raid_bdev1", 00:27:01.382 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:01.382 "strip_size_kb": 0, 00:27:01.382 "state": "online", 00:27:01.382 "raid_level": "raid1", 00:27:01.382 "superblock": true, 00:27:01.382 "num_base_bdevs": 2, 00:27:01.382 "num_base_bdevs_discovered": 1, 00:27:01.382 "num_base_bdevs_operational": 1, 00:27:01.382 "base_bdevs_list": [ 00:27:01.382 { 00:27:01.382 "name": null, 00:27:01.382 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:01.382 "is_configured": false, 00:27:01.382 "data_offset": 256, 00:27:01.382 "data_size": 7936 00:27:01.382 }, 00:27:01.382 { 00:27:01.382 "name": "BaseBdev2", 00:27:01.382 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:01.382 "is_configured": true, 00:27:01.382 "data_offset": 256, 00:27:01.382 "data_size": 7936 00:27:01.382 } 00:27:01.382 ] 00:27:01.382 }' 00:27:01.382 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:01.382 11:37:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:27:01.941 11:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@645 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:27:02.197 [2024-07-15 11:37:45.536689] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:27:02.197 [2024-07-15 11:37:45.541658] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x160dce0 00:27:02.197 [2024-07-15 11:37:45.543859] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:27:02.197 11:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@646 -- # sleep 1 00:27:03.124 11:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@649 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:03.124 11:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:27:03.124 11:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:27:03.124 11:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local target=spare 00:27:03.124 11:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:27:03.124 11:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:03.124 11:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:03.380 11:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:03.380 "name": "raid_bdev1", 00:27:03.380 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:03.380 "strip_size_kb": 0, 00:27:03.380 "state": "online", 00:27:03.380 "raid_level": "raid1", 00:27:03.380 "superblock": true, 00:27:03.380 "num_base_bdevs": 2, 00:27:03.380 "num_base_bdevs_discovered": 2, 00:27:03.380 "num_base_bdevs_operational": 2, 00:27:03.380 "process": { 00:27:03.380 "type": "rebuild", 00:27:03.380 "target": "spare", 00:27:03.380 "progress": { 00:27:03.380 "blocks": 3072, 00:27:03.380 "percent": 38 00:27:03.380 } 00:27:03.380 }, 00:27:03.380 "base_bdevs_list": [ 00:27:03.380 { 00:27:03.380 "name": "spare", 00:27:03.380 "uuid": "c5be393d-e1ed-558a-b2fa-2ace2cb72166", 00:27:03.380 "is_configured": true, 00:27:03.380 "data_offset": 256, 00:27:03.380 "data_size": 7936 00:27:03.380 }, 00:27:03.380 { 00:27:03.380 "name": "BaseBdev2", 00:27:03.380 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:03.380 "is_configured": true, 00:27:03.380 "data_offset": 256, 00:27:03.380 "data_size": 7936 00:27:03.380 } 00:27:03.380 ] 00:27:03.380 }' 00:27:03.380 11:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:27:03.380 11:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:03.380 11:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:27:03.380 11:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:27:03.380 11:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@652 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev spare 00:27:03.636 [2024-07-15 11:37:47.130709] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:27:03.636 [2024-07-15 11:37:47.156752] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:27:03.636 [2024-07-15 11:37:47.156804] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:03.636 [2024-07-15 11:37:47.156820] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:27:03.636 [2024-07-15 11:37:47.156828] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:27:03.636 11:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@655 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:03.636 11:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:27:03.636 11:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:27:03.636 11:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:03.636 11:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:03.636 11:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:27:03.636 11:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:03.636 11:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:03.636 11:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:03.636 11:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:03.636 11:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:03.636 11:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:03.892 11:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:03.892 "name": "raid_bdev1", 00:27:03.892 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:03.892 "strip_size_kb": 0, 00:27:03.892 "state": "online", 00:27:03.892 "raid_level": "raid1", 00:27:03.892 "superblock": true, 00:27:03.892 "num_base_bdevs": 2, 00:27:03.892 "num_base_bdevs_discovered": 1, 00:27:03.892 "num_base_bdevs_operational": 1, 00:27:03.892 "base_bdevs_list": [ 00:27:03.892 { 00:27:03.892 "name": null, 00:27:03.892 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:03.892 "is_configured": false, 00:27:03.892 "data_offset": 256, 00:27:03.892 "data_size": 7936 00:27:03.892 }, 00:27:03.892 { 00:27:03.892 "name": "BaseBdev2", 00:27:03.892 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:03.892 "is_configured": true, 00:27:03.892 "data_offset": 256, 00:27:03.892 "data_size": 7936 00:27:03.892 } 00:27:03.892 ] 00:27:03.892 }' 00:27:03.892 11:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:03.892 11:37:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:27:04.455 11:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@658 -- # verify_raid_bdev_process raid_bdev1 none none 00:27:04.455 11:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:27:04.455 11:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:27:04.455 11:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local target=none 00:27:04.455 11:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:27:04.455 11:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:04.455 11:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:04.711 11:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:04.711 "name": "raid_bdev1", 00:27:04.711 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:04.711 "strip_size_kb": 0, 00:27:04.711 "state": "online", 00:27:04.711 "raid_level": "raid1", 00:27:04.711 "superblock": true, 00:27:04.711 "num_base_bdevs": 2, 00:27:04.711 "num_base_bdevs_discovered": 1, 00:27:04.711 "num_base_bdevs_operational": 1, 00:27:04.711 "base_bdevs_list": [ 00:27:04.711 { 00:27:04.711 "name": null, 00:27:04.711 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:04.711 "is_configured": false, 00:27:04.711 "data_offset": 256, 00:27:04.712 "data_size": 7936 00:27:04.712 }, 00:27:04.712 { 00:27:04.712 "name": "BaseBdev2", 00:27:04.712 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:04.712 "is_configured": true, 00:27:04.712 "data_offset": 256, 00:27:04.712 "data_size": 7936 00:27:04.712 } 00:27:04.712 ] 00:27:04.712 }' 00:27:04.712 11:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:27:04.969 11:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:27:04.969 11:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:27:04.969 11:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:27:04.969 11:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@661 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:27:05.257 [2024-07-15 11:37:48.589068] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:27:05.257 [2024-07-15 11:37:48.594012] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x160dce0 00:27:05.257 [2024-07-15 11:37:48.595466] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:27:05.257 11:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@662 -- # sleep 1 00:27:06.185 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@663 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:06.185 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:27:06.185 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:27:06.185 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local target=spare 00:27:06.185 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:27:06.185 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:06.185 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:06.442 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:06.442 "name": "raid_bdev1", 00:27:06.442 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:06.442 "strip_size_kb": 0, 00:27:06.442 "state": "online", 00:27:06.442 "raid_level": "raid1", 00:27:06.442 "superblock": true, 00:27:06.442 "num_base_bdevs": 2, 00:27:06.442 "num_base_bdevs_discovered": 2, 00:27:06.442 "num_base_bdevs_operational": 2, 00:27:06.442 "process": { 00:27:06.442 "type": "rebuild", 00:27:06.442 "target": "spare", 00:27:06.442 "progress": { 00:27:06.442 "blocks": 3072, 00:27:06.442 "percent": 38 00:27:06.442 } 00:27:06.442 }, 00:27:06.442 "base_bdevs_list": [ 00:27:06.442 { 00:27:06.442 "name": "spare", 00:27:06.442 "uuid": "c5be393d-e1ed-558a-b2fa-2ace2cb72166", 00:27:06.442 "is_configured": true, 00:27:06.442 "data_offset": 256, 00:27:06.442 "data_size": 7936 00:27:06.442 }, 00:27:06.442 { 00:27:06.442 "name": "BaseBdev2", 00:27:06.442 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:06.442 "is_configured": true, 00:27:06.442 "data_offset": 256, 00:27:06.442 "data_size": 7936 00:27:06.442 } 00:27:06.442 ] 00:27:06.442 }' 00:27:06.442 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:27:06.442 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:06.442 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:27:06.442 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:27:06.442 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@665 -- # '[' true = true ']' 00:27:06.442 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@665 -- # '[' = false ']' 00:27:06.442 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev_raid.sh: line 665: [: =: unary operator expected 00:27:06.442 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@690 -- # local num_base_bdevs_operational=2 00:27:06.442 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@692 -- # '[' raid1 = raid1 ']' 00:27:06.442 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@692 -- # '[' 2 -gt 2 ']' 00:27:06.442 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@705 -- # local timeout=997 00:27:06.442 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:27:06.442 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:06.442 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:27:06.442 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:27:06.442 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local target=spare 00:27:06.442 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:27:06.442 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:06.442 11:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:06.699 11:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:06.699 "name": "raid_bdev1", 00:27:06.699 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:06.699 "strip_size_kb": 0, 00:27:06.699 "state": "online", 00:27:06.699 "raid_level": "raid1", 00:27:06.699 "superblock": true, 00:27:06.699 "num_base_bdevs": 2, 00:27:06.699 "num_base_bdevs_discovered": 2, 00:27:06.699 "num_base_bdevs_operational": 2, 00:27:06.699 "process": { 00:27:06.699 "type": "rebuild", 00:27:06.699 "target": "spare", 00:27:06.699 "progress": { 00:27:06.699 "blocks": 3840, 00:27:06.699 "percent": 48 00:27:06.699 } 00:27:06.699 }, 00:27:06.699 "base_bdevs_list": [ 00:27:06.699 { 00:27:06.699 "name": "spare", 00:27:06.699 "uuid": "c5be393d-e1ed-558a-b2fa-2ace2cb72166", 00:27:06.699 "is_configured": true, 00:27:06.699 "data_offset": 256, 00:27:06.699 "data_size": 7936 00:27:06.699 }, 00:27:06.699 { 00:27:06.699 "name": "BaseBdev2", 00:27:06.699 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:06.699 "is_configured": true, 00:27:06.699 "data_offset": 256, 00:27:06.699 "data_size": 7936 00:27:06.699 } 00:27:06.699 ] 00:27:06.699 }' 00:27:06.699 11:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:27:06.699 11:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:06.699 11:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:27:06.955 11:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:27:06.955 11:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@710 -- # sleep 1 00:27:07.887 11:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:27:07.887 11:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:07.887 11:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:27:07.887 11:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:27:07.887 11:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local target=spare 00:27:07.887 11:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:27:07.887 11:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:07.887 11:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:08.146 11:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:08.146 "name": "raid_bdev1", 00:27:08.146 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:08.146 "strip_size_kb": 0, 00:27:08.146 "state": "online", 00:27:08.146 "raid_level": "raid1", 00:27:08.146 "superblock": true, 00:27:08.146 "num_base_bdevs": 2, 00:27:08.146 "num_base_bdevs_discovered": 2, 00:27:08.146 "num_base_bdevs_operational": 2, 00:27:08.146 "process": { 00:27:08.146 "type": "rebuild", 00:27:08.146 "target": "spare", 00:27:08.146 "progress": { 00:27:08.146 "blocks": 7424, 00:27:08.146 "percent": 93 00:27:08.146 } 00:27:08.146 }, 00:27:08.146 "base_bdevs_list": [ 00:27:08.146 { 00:27:08.146 "name": "spare", 00:27:08.146 "uuid": "c5be393d-e1ed-558a-b2fa-2ace2cb72166", 00:27:08.146 "is_configured": true, 00:27:08.146 "data_offset": 256, 00:27:08.146 "data_size": 7936 00:27:08.146 }, 00:27:08.146 { 00:27:08.146 "name": "BaseBdev2", 00:27:08.146 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:08.146 "is_configured": true, 00:27:08.146 "data_offset": 256, 00:27:08.146 "data_size": 7936 00:27:08.146 } 00:27:08.146 ] 00:27:08.146 }' 00:27:08.146 11:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:27:08.146 11:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:08.146 11:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:27:08.146 11:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:27:08.146 11:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@710 -- # sleep 1 00:27:08.146 [2024-07-15 11:37:51.719610] bdev_raid.c:2789:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:27:08.146 [2024-07-15 11:37:51.719675] bdev_raid.c:2504:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:27:08.146 [2024-07-15 11:37:51.719756] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:09.079 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:27:09.079 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:09.079 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:27:09.079 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:27:09.079 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local target=spare 00:27:09.079 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:27:09.079 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:09.079 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:09.338 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:09.338 "name": "raid_bdev1", 00:27:09.338 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:09.338 "strip_size_kb": 0, 00:27:09.338 "state": "online", 00:27:09.338 "raid_level": "raid1", 00:27:09.338 "superblock": true, 00:27:09.338 "num_base_bdevs": 2, 00:27:09.338 "num_base_bdevs_discovered": 2, 00:27:09.338 "num_base_bdevs_operational": 2, 00:27:09.338 "base_bdevs_list": [ 00:27:09.338 { 00:27:09.338 "name": "spare", 00:27:09.338 "uuid": "c5be393d-e1ed-558a-b2fa-2ace2cb72166", 00:27:09.338 "is_configured": true, 00:27:09.338 "data_offset": 256, 00:27:09.338 "data_size": 7936 00:27:09.338 }, 00:27:09.338 { 00:27:09.338 "name": "BaseBdev2", 00:27:09.338 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:09.338 "is_configured": true, 00:27:09.338 "data_offset": 256, 00:27:09.338 "data_size": 7936 00:27:09.338 } 00:27:09.338 ] 00:27:09.338 }' 00:27:09.338 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:27:09.652 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # [[ none == \r\e\b\u\i\l\d ]] 00:27:09.652 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:27:09.652 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # [[ none == \s\p\a\r\e ]] 00:27:09.652 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # break 00:27:09.652 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@714 -- # verify_raid_bdev_process raid_bdev1 none none 00:27:09.652 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:27:09.652 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:27:09.652 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local target=none 00:27:09.652 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:27:09.652 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:09.652 11:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:09.910 11:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:09.910 "name": "raid_bdev1", 00:27:09.910 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:09.910 "strip_size_kb": 0, 00:27:09.910 "state": "online", 00:27:09.910 "raid_level": "raid1", 00:27:09.910 "superblock": true, 00:27:09.910 "num_base_bdevs": 2, 00:27:09.910 "num_base_bdevs_discovered": 2, 00:27:09.910 "num_base_bdevs_operational": 2, 00:27:09.910 "base_bdevs_list": [ 00:27:09.910 { 00:27:09.910 "name": "spare", 00:27:09.910 "uuid": "c5be393d-e1ed-558a-b2fa-2ace2cb72166", 00:27:09.910 "is_configured": true, 00:27:09.910 "data_offset": 256, 00:27:09.910 "data_size": 7936 00:27:09.910 }, 00:27:09.910 { 00:27:09.910 "name": "BaseBdev2", 00:27:09.910 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:09.910 "is_configured": true, 00:27:09.910 "data_offset": 256, 00:27:09.910 "data_size": 7936 00:27:09.910 } 00:27:09.910 ] 00:27:09.910 }' 00:27:09.910 11:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:27:09.910 11:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:27:09.910 11:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:27:09.910 11:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:27:09.910 11:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:27:09.910 11:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:27:09.910 11:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:27:09.910 11:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:09.910 11:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:09.910 11:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:27:09.910 11:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:09.910 11:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:09.910 11:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:09.910 11:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:09.910 11:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:09.910 11:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:10.167 11:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:10.167 "name": "raid_bdev1", 00:27:10.167 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:10.167 "strip_size_kb": 0, 00:27:10.167 "state": "online", 00:27:10.167 "raid_level": "raid1", 00:27:10.167 "superblock": true, 00:27:10.167 "num_base_bdevs": 2, 00:27:10.167 "num_base_bdevs_discovered": 2, 00:27:10.167 "num_base_bdevs_operational": 2, 00:27:10.167 "base_bdevs_list": [ 00:27:10.167 { 00:27:10.167 "name": "spare", 00:27:10.167 "uuid": "c5be393d-e1ed-558a-b2fa-2ace2cb72166", 00:27:10.167 "is_configured": true, 00:27:10.167 "data_offset": 256, 00:27:10.167 "data_size": 7936 00:27:10.167 }, 00:27:10.167 { 00:27:10.167 "name": "BaseBdev2", 00:27:10.167 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:10.167 "is_configured": true, 00:27:10.167 "data_offset": 256, 00:27:10.167 "data_size": 7936 00:27:10.167 } 00:27:10.167 ] 00:27:10.167 }' 00:27:10.167 11:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:10.168 11:37:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:27:10.733 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@718 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:27:10.990 [2024-07-15 11:37:54.408025] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:10.990 [2024-07-15 11:37:54.408050] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:10.990 [2024-07-15 11:37:54.408108] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:10.990 [2024-07-15 11:37:54.408164] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:10.990 [2024-07-15 11:37:54.408176] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x160e070 name raid_bdev1, state offline 00:27:10.990 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@719 -- # jq length 00:27:10.990 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@719 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:11.247 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@719 -- # [[ 0 == 0 ]] 00:27:11.247 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@721 -- # '[' true = true ']' 00:27:11.247 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@722 -- # '[' false = true ']' 00:27:11.247 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@736 -- # nbd_start_disks /var/tmp/spdk-raid.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:27:11.247 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:27:11.247 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:27:11.247 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:27:11.247 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:27:11.247 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:27:11.247 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:27:11.247 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:27:11.247 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:27:11.247 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:27:11.505 /dev/nbd0 00:27:11.505 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:27:11.505 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:27:11.505 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:27:11.505 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@867 -- # local i 00:27:11.505 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:27:11.505 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:27:11.505 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:27:11.505 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # break 00:27:11.505 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:27:11.505 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:27:11.505 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:27:11.505 1+0 records in 00:27:11.505 1+0 records out 00:27:11.505 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000226832 s, 18.1 MB/s 00:27:11.505 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:27:11.505 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # size=4096 00:27:11.505 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:27:11.505 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:27:11.505 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@887 -- # return 0 00:27:11.505 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:27:11.505 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:27:11.505 11:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk spare /dev/nbd1 00:27:11.763 /dev/nbd1 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@867 -- # local i 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # break 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:27:11.763 1+0 records in 00:27:11.763 1+0 records out 00:27:11.763 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000341059 s, 12.0 MB/s 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # size=4096 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@887 -- # return 0 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@737 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@738 -- # nbd_stop_disks /var/tmp/spdk-raid.sock '/dev/nbd0 /dev/nbd1' 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:27:11.763 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd0 00:27:12.021 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:27:12.021 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:27:12.021 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:27:12.021 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:27:12.021 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:27:12.021 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:27:12.279 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:27:12.279 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:27:12.279 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:27:12.279 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd1 00:27:12.536 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:27:12.536 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:27:12.536 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:27:12.536 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:27:12.536 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:27:12.536 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:27:12.536 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:27:12.536 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:27:12.536 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@742 -- # '[' true = true ']' 00:27:12.536 11:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@744 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:27:12.793 11:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@745 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:27:12.793 [2024-07-15 11:37:56.377415] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:27:12.794 [2024-07-15 11:37:56.377463] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:12.794 [2024-07-15 11:37:56.377483] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x160d500 00:27:12.794 [2024-07-15 11:37:56.377496] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:12.794 [2024-07-15 11:37:56.379112] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:12.794 [2024-07-15 11:37:56.379140] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:27:12.794 [2024-07-15 11:37:56.379220] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:27:12.794 [2024-07-15 11:37:56.379246] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:27:12.794 [2024-07-15 11:37:56.379345] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:12.794 spare 00:27:13.051 11:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@747 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:27:13.051 11:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:27:13.051 11:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:27:13.051 11:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:13.051 11:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:13.051 11:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:27:13.051 11:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:13.051 11:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:13.051 11:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:13.051 11:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:13.051 11:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:13.051 11:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:13.051 [2024-07-15 11:37:56.479661] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x160c260 00:27:13.051 [2024-07-15 11:37:56.479680] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:27:13.051 [2024-07-15 11:37:56.479880] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1607490 00:27:13.051 [2024-07-15 11:37:56.480035] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x160c260 00:27:13.051 [2024-07-15 11:37:56.480046] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x160c260 00:27:13.051 [2024-07-15 11:37:56.480151] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:13.309 11:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:13.309 "name": "raid_bdev1", 00:27:13.309 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:13.309 "strip_size_kb": 0, 00:27:13.309 "state": "online", 00:27:13.309 "raid_level": "raid1", 00:27:13.309 "superblock": true, 00:27:13.309 "num_base_bdevs": 2, 00:27:13.309 "num_base_bdevs_discovered": 2, 00:27:13.309 "num_base_bdevs_operational": 2, 00:27:13.309 "base_bdevs_list": [ 00:27:13.309 { 00:27:13.309 "name": "spare", 00:27:13.309 "uuid": "c5be393d-e1ed-558a-b2fa-2ace2cb72166", 00:27:13.309 "is_configured": true, 00:27:13.309 "data_offset": 256, 00:27:13.309 "data_size": 7936 00:27:13.309 }, 00:27:13.309 { 00:27:13.309 "name": "BaseBdev2", 00:27:13.309 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:13.309 "is_configured": true, 00:27:13.309 "data_offset": 256, 00:27:13.309 "data_size": 7936 00:27:13.309 } 00:27:13.309 ] 00:27:13.309 }' 00:27:13.309 11:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:13.309 11:37:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:27:13.874 11:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@748 -- # verify_raid_bdev_process raid_bdev1 none none 00:27:13.874 11:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:27:13.874 11:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:27:13.874 11:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local target=none 00:27:13.874 11:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:27:13.874 11:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:13.875 11:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:13.875 11:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:13.875 "name": "raid_bdev1", 00:27:13.875 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:13.875 "strip_size_kb": 0, 00:27:13.875 "state": "online", 00:27:13.875 "raid_level": "raid1", 00:27:13.875 "superblock": true, 00:27:13.875 "num_base_bdevs": 2, 00:27:13.875 "num_base_bdevs_discovered": 2, 00:27:13.875 "num_base_bdevs_operational": 2, 00:27:13.875 "base_bdevs_list": [ 00:27:13.875 { 00:27:13.875 "name": "spare", 00:27:13.875 "uuid": "c5be393d-e1ed-558a-b2fa-2ace2cb72166", 00:27:13.875 "is_configured": true, 00:27:13.875 "data_offset": 256, 00:27:13.875 "data_size": 7936 00:27:13.875 }, 00:27:13.875 { 00:27:13.875 "name": "BaseBdev2", 00:27:13.875 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:13.875 "is_configured": true, 00:27:13.875 "data_offset": 256, 00:27:13.875 "data_size": 7936 00:27:13.875 } 00:27:13.875 ] 00:27:13.875 }' 00:27:13.875 11:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:27:14.132 11:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:27:14.132 11:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:27:14.132 11:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:27:14.132 11:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@749 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:14.132 11:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@749 -- # jq -r '.[].base_bdevs_list[0].name' 00:27:14.390 11:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@749 -- # [[ spare == \s\p\a\r\e ]] 00:27:14.390 11:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@752 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev spare 00:27:14.648 [2024-07-15 11:37:57.993826] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:27:14.648 11:37:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@753 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:14.648 11:37:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:27:14.648 11:37:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:27:14.648 11:37:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:14.648 11:37:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:14.648 11:37:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:27:14.648 11:37:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:14.648 11:37:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:14.648 11:37:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:14.648 11:37:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:14.648 11:37:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:14.648 11:37:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:14.906 11:37:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:14.906 "name": "raid_bdev1", 00:27:14.906 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:14.906 "strip_size_kb": 0, 00:27:14.906 "state": "online", 00:27:14.906 "raid_level": "raid1", 00:27:14.906 "superblock": true, 00:27:14.906 "num_base_bdevs": 2, 00:27:14.906 "num_base_bdevs_discovered": 1, 00:27:14.906 "num_base_bdevs_operational": 1, 00:27:14.906 "base_bdevs_list": [ 00:27:14.906 { 00:27:14.906 "name": null, 00:27:14.907 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:14.907 "is_configured": false, 00:27:14.907 "data_offset": 256, 00:27:14.907 "data_size": 7936 00:27:14.907 }, 00:27:14.907 { 00:27:14.907 "name": "BaseBdev2", 00:27:14.907 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:14.907 "is_configured": true, 00:27:14.907 "data_offset": 256, 00:27:14.907 "data_size": 7936 00:27:14.907 } 00:27:14.907 ] 00:27:14.907 }' 00:27:14.907 11:37:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:14.907 11:37:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:27:15.472 11:37:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@754 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:27:15.472 [2024-07-15 11:37:59.060679] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:27:15.472 [2024-07-15 11:37:59.060823] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:27:15.472 [2024-07-15 11:37:59.060840] bdev_raid.c:3620:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:27:15.472 [2024-07-15 11:37:59.060867] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:27:15.472 [2024-07-15 11:37:59.065658] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1607490 00:27:15.730 [2024-07-15 11:37:59.068007] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:27:15.730 11:37:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@755 -- # sleep 1 00:27:16.663 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@756 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:16.663 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:27:16.663 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:27:16.663 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local target=spare 00:27:16.663 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:27:16.663 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:16.663 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:16.921 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:16.921 "name": "raid_bdev1", 00:27:16.921 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:16.921 "strip_size_kb": 0, 00:27:16.921 "state": "online", 00:27:16.921 "raid_level": "raid1", 00:27:16.921 "superblock": true, 00:27:16.921 "num_base_bdevs": 2, 00:27:16.921 "num_base_bdevs_discovered": 2, 00:27:16.921 "num_base_bdevs_operational": 2, 00:27:16.921 "process": { 00:27:16.921 "type": "rebuild", 00:27:16.921 "target": "spare", 00:27:16.921 "progress": { 00:27:16.921 "blocks": 3072, 00:27:16.921 "percent": 38 00:27:16.921 } 00:27:16.921 }, 00:27:16.921 "base_bdevs_list": [ 00:27:16.921 { 00:27:16.921 "name": "spare", 00:27:16.921 "uuid": "c5be393d-e1ed-558a-b2fa-2ace2cb72166", 00:27:16.921 "is_configured": true, 00:27:16.921 "data_offset": 256, 00:27:16.921 "data_size": 7936 00:27:16.921 }, 00:27:16.921 { 00:27:16.921 "name": "BaseBdev2", 00:27:16.921 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:16.921 "is_configured": true, 00:27:16.921 "data_offset": 256, 00:27:16.921 "data_size": 7936 00:27:16.921 } 00:27:16.921 ] 00:27:16.921 }' 00:27:16.921 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:27:16.921 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:16.921 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:27:16.921 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:27:16.921 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@759 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:27:17.179 [2024-07-15 11:38:00.650271] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:27:17.180 [2024-07-15 11:38:00.680736] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:27:17.180 [2024-07-15 11:38:00.680776] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:17.180 [2024-07-15 11:38:00.680791] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:27:17.180 [2024-07-15 11:38:00.680800] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:27:17.180 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@760 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:17.180 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:27:17.180 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:27:17.180 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:17.180 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:17.180 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:27:17.180 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:17.180 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:17.180 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:17.180 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:17.180 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:17.180 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:17.437 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:17.437 "name": "raid_bdev1", 00:27:17.437 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:17.437 "strip_size_kb": 0, 00:27:17.437 "state": "online", 00:27:17.437 "raid_level": "raid1", 00:27:17.437 "superblock": true, 00:27:17.437 "num_base_bdevs": 2, 00:27:17.437 "num_base_bdevs_discovered": 1, 00:27:17.437 "num_base_bdevs_operational": 1, 00:27:17.437 "base_bdevs_list": [ 00:27:17.437 { 00:27:17.437 "name": null, 00:27:17.437 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:17.437 "is_configured": false, 00:27:17.437 "data_offset": 256, 00:27:17.437 "data_size": 7936 00:27:17.437 }, 00:27:17.437 { 00:27:17.437 "name": "BaseBdev2", 00:27:17.437 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:17.437 "is_configured": true, 00:27:17.437 "data_offset": 256, 00:27:17.437 "data_size": 7936 00:27:17.437 } 00:27:17.437 ] 00:27:17.437 }' 00:27:17.437 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:17.437 11:38:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:27:18.000 11:38:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@761 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:27:18.257 [2024-07-15 11:38:01.755996] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:27:18.257 [2024-07-15 11:38:01.756044] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:18.257 [2024-07-15 11:38:01.756066] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x160d730 00:27:18.257 [2024-07-15 11:38:01.756079] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:18.257 [2024-07-15 11:38:01.756445] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:18.257 [2024-07-15 11:38:01.756463] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:27:18.257 [2024-07-15 11:38:01.756539] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:27:18.257 [2024-07-15 11:38:01.756552] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:27:18.257 [2024-07-15 11:38:01.756569] bdev_raid.c:3620:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:27:18.257 [2024-07-15 11:38:01.756589] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:27:18.257 [2024-07-15 11:38:01.761469] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x160eaa0 00:27:18.257 spare 00:27:18.257 [2024-07-15 11:38:01.762932] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:27:18.257 11:38:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@762 -- # sleep 1 00:27:19.621 11:38:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@763 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:19.621 11:38:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:27:19.621 11:38:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:27:19.621 11:38:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local target=spare 00:27:19.621 11:38:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:27:19.621 11:38:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:19.621 11:38:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:19.621 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:19.621 "name": "raid_bdev1", 00:27:19.621 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:19.621 "strip_size_kb": 0, 00:27:19.621 "state": "online", 00:27:19.621 "raid_level": "raid1", 00:27:19.621 "superblock": true, 00:27:19.621 "num_base_bdevs": 2, 00:27:19.621 "num_base_bdevs_discovered": 2, 00:27:19.621 "num_base_bdevs_operational": 2, 00:27:19.621 "process": { 00:27:19.621 "type": "rebuild", 00:27:19.621 "target": "spare", 00:27:19.621 "progress": { 00:27:19.621 "blocks": 3072, 00:27:19.621 "percent": 38 00:27:19.621 } 00:27:19.621 }, 00:27:19.621 "base_bdevs_list": [ 00:27:19.621 { 00:27:19.621 "name": "spare", 00:27:19.621 "uuid": "c5be393d-e1ed-558a-b2fa-2ace2cb72166", 00:27:19.621 "is_configured": true, 00:27:19.621 "data_offset": 256, 00:27:19.621 "data_size": 7936 00:27:19.621 }, 00:27:19.621 { 00:27:19.621 "name": "BaseBdev2", 00:27:19.621 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:19.621 "is_configured": true, 00:27:19.621 "data_offset": 256, 00:27:19.621 "data_size": 7936 00:27:19.621 } 00:27:19.621 ] 00:27:19.621 }' 00:27:19.621 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:27:19.621 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:19.621 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:27:19.621 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:27:19.621 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@766 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:27:19.877 [2024-07-15 11:38:03.354227] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:27:19.877 [2024-07-15 11:38:03.375204] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:27:19.877 [2024-07-15 11:38:03.375247] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:19.877 [2024-07-15 11:38:03.375262] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:27:19.877 [2024-07-15 11:38:03.375270] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:27:19.877 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@767 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:19.877 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:27:19.877 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:27:19.877 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:19.877 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:19.877 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:27:19.877 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:19.877 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:19.877 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:19.877 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:19.877 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:19.877 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:20.134 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:20.134 "name": "raid_bdev1", 00:27:20.134 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:20.134 "strip_size_kb": 0, 00:27:20.134 "state": "online", 00:27:20.134 "raid_level": "raid1", 00:27:20.134 "superblock": true, 00:27:20.134 "num_base_bdevs": 2, 00:27:20.134 "num_base_bdevs_discovered": 1, 00:27:20.134 "num_base_bdevs_operational": 1, 00:27:20.134 "base_bdevs_list": [ 00:27:20.134 { 00:27:20.134 "name": null, 00:27:20.134 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:20.134 "is_configured": false, 00:27:20.134 "data_offset": 256, 00:27:20.134 "data_size": 7936 00:27:20.134 }, 00:27:20.134 { 00:27:20.134 "name": "BaseBdev2", 00:27:20.134 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:20.134 "is_configured": true, 00:27:20.134 "data_offset": 256, 00:27:20.134 "data_size": 7936 00:27:20.134 } 00:27:20.134 ] 00:27:20.134 }' 00:27:20.134 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:20.134 11:38:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:27:20.700 11:38:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@768 -- # verify_raid_bdev_process raid_bdev1 none none 00:27:20.700 11:38:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:27:20.700 11:38:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:27:20.700 11:38:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local target=none 00:27:20.700 11:38:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:27:20.700 11:38:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:20.700 11:38:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:20.956 11:38:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:20.956 "name": "raid_bdev1", 00:27:20.956 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:20.956 "strip_size_kb": 0, 00:27:20.956 "state": "online", 00:27:20.956 "raid_level": "raid1", 00:27:20.956 "superblock": true, 00:27:20.956 "num_base_bdevs": 2, 00:27:20.956 "num_base_bdevs_discovered": 1, 00:27:20.956 "num_base_bdevs_operational": 1, 00:27:20.956 "base_bdevs_list": [ 00:27:20.956 { 00:27:20.956 "name": null, 00:27:20.957 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:20.957 "is_configured": false, 00:27:20.957 "data_offset": 256, 00:27:20.957 "data_size": 7936 00:27:20.957 }, 00:27:20.957 { 00:27:20.957 "name": "BaseBdev2", 00:27:20.957 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:20.957 "is_configured": true, 00:27:20.957 "data_offset": 256, 00:27:20.957 "data_size": 7936 00:27:20.957 } 00:27:20.957 ] 00:27:20.957 }' 00:27:20.957 11:38:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:27:20.957 11:38:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:27:20.957 11:38:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:27:21.213 11:38:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:27:21.213 11:38:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@771 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete BaseBdev1 00:27:21.470 11:38:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@772 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:27:21.727 [2024-07-15 11:38:05.309434] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:27:21.727 [2024-07-15 11:38:05.309483] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:21.727 [2024-07-15 11:38:05.309513] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1608650 00:27:21.727 [2024-07-15 11:38:05.309526] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:21.727 [2024-07-15 11:38:05.309868] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:21.727 [2024-07-15 11:38:05.309885] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:27:21.727 [2024-07-15 11:38:05.309957] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:27:21.727 [2024-07-15 11:38:05.309970] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:27:21.728 [2024-07-15 11:38:05.309980] bdev_raid.c:3581:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:27:21.728 BaseBdev1 00:27:21.985 11:38:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@773 -- # sleep 1 00:27:22.917 11:38:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@774 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:22.917 11:38:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:27:22.917 11:38:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:27:22.917 11:38:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:22.917 11:38:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:22.917 11:38:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:27:22.917 11:38:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:22.917 11:38:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:22.917 11:38:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:22.917 11:38:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:22.917 11:38:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:22.917 11:38:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:23.174 11:38:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:23.174 "name": "raid_bdev1", 00:27:23.174 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:23.174 "strip_size_kb": 0, 00:27:23.174 "state": "online", 00:27:23.174 "raid_level": "raid1", 00:27:23.174 "superblock": true, 00:27:23.174 "num_base_bdevs": 2, 00:27:23.174 "num_base_bdevs_discovered": 1, 00:27:23.174 "num_base_bdevs_operational": 1, 00:27:23.174 "base_bdevs_list": [ 00:27:23.174 { 00:27:23.174 "name": null, 00:27:23.174 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:23.174 "is_configured": false, 00:27:23.174 "data_offset": 256, 00:27:23.174 "data_size": 7936 00:27:23.174 }, 00:27:23.174 { 00:27:23.174 "name": "BaseBdev2", 00:27:23.174 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:23.174 "is_configured": true, 00:27:23.174 "data_offset": 256, 00:27:23.174 "data_size": 7936 00:27:23.174 } 00:27:23.174 ] 00:27:23.174 }' 00:27:23.174 11:38:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:23.174 11:38:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:27:23.752 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@775 -- # verify_raid_bdev_process raid_bdev1 none none 00:27:23.752 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:27:23.752 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:27:23.752 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local target=none 00:27:23.752 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:27:23.752 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:23.752 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:24.010 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:24.010 "name": "raid_bdev1", 00:27:24.010 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:24.010 "strip_size_kb": 0, 00:27:24.010 "state": "online", 00:27:24.010 "raid_level": "raid1", 00:27:24.010 "superblock": true, 00:27:24.010 "num_base_bdevs": 2, 00:27:24.010 "num_base_bdevs_discovered": 1, 00:27:24.010 "num_base_bdevs_operational": 1, 00:27:24.010 "base_bdevs_list": [ 00:27:24.010 { 00:27:24.010 "name": null, 00:27:24.010 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:24.010 "is_configured": false, 00:27:24.010 "data_offset": 256, 00:27:24.010 "data_size": 7936 00:27:24.010 }, 00:27:24.010 { 00:27:24.010 "name": "BaseBdev2", 00:27:24.010 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:24.010 "is_configured": true, 00:27:24.010 "data_offset": 256, 00:27:24.010 "data_size": 7936 00:27:24.010 } 00:27:24.010 ] 00:27:24.010 }' 00:27:24.010 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:27:24.010 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:27:24.010 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:27:24.010 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:27:24.010 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@776 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:27:24.010 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@648 -- # local es=0 00:27:24.010 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:27:24.010 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:27:24.010 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:27:24.010 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:27:24.010 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:27:24.010 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:27:24.010 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:27:24.010 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:27:24.010 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py ]] 00:27:24.010 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:27:24.330 [2024-07-15 11:38:07.731881] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:24.330 [2024-07-15 11:38:07.732008] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:27:24.330 [2024-07-15 11:38:07.732025] bdev_raid.c:3581:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:27:24.330 request: 00:27:24.330 { 00:27:24.330 "base_bdev": "BaseBdev1", 00:27:24.330 "raid_bdev": "raid_bdev1", 00:27:24.330 "method": "bdev_raid_add_base_bdev", 00:27:24.330 "req_id": 1 00:27:24.330 } 00:27:24.330 Got JSON-RPC error response 00:27:24.330 response: 00:27:24.330 { 00:27:24.330 "code": -22, 00:27:24.330 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:27:24.330 } 00:27:24.330 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@651 -- # es=1 00:27:24.330 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:27:24.330 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:27:24.330 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:27:24.330 11:38:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@777 -- # sleep 1 00:27:25.284 11:38:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@778 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:25.284 11:38:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:27:25.284 11:38:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:27:25.284 11:38:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:25.284 11:38:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:25.284 11:38:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:27:25.284 11:38:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:25.284 11:38:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:25.284 11:38:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:25.284 11:38:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:25.284 11:38:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:25.284 11:38:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:25.541 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:25.541 "name": "raid_bdev1", 00:27:25.541 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:25.541 "strip_size_kb": 0, 00:27:25.541 "state": "online", 00:27:25.541 "raid_level": "raid1", 00:27:25.541 "superblock": true, 00:27:25.541 "num_base_bdevs": 2, 00:27:25.541 "num_base_bdevs_discovered": 1, 00:27:25.541 "num_base_bdevs_operational": 1, 00:27:25.541 "base_bdevs_list": [ 00:27:25.541 { 00:27:25.541 "name": null, 00:27:25.541 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:25.541 "is_configured": false, 00:27:25.541 "data_offset": 256, 00:27:25.541 "data_size": 7936 00:27:25.541 }, 00:27:25.541 { 00:27:25.541 "name": "BaseBdev2", 00:27:25.541 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:25.541 "is_configured": true, 00:27:25.541 "data_offset": 256, 00:27:25.541 "data_size": 7936 00:27:25.541 } 00:27:25.541 ] 00:27:25.541 }' 00:27:25.541 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:25.541 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:27:26.103 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@779 -- # verify_raid_bdev_process raid_bdev1 none none 00:27:26.103 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:27:26.103 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:27:26.103 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local target=none 00:27:26.103 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:27:26.103 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:26.103 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:26.359 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:26.359 "name": "raid_bdev1", 00:27:26.359 "uuid": "ff8f66ba-310d-4d45-a181-60b51508321c", 00:27:26.359 "strip_size_kb": 0, 00:27:26.359 "state": "online", 00:27:26.359 "raid_level": "raid1", 00:27:26.359 "superblock": true, 00:27:26.359 "num_base_bdevs": 2, 00:27:26.359 "num_base_bdevs_discovered": 1, 00:27:26.359 "num_base_bdevs_operational": 1, 00:27:26.359 "base_bdevs_list": [ 00:27:26.359 { 00:27:26.359 "name": null, 00:27:26.359 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:26.359 "is_configured": false, 00:27:26.359 "data_offset": 256, 00:27:26.359 "data_size": 7936 00:27:26.359 }, 00:27:26.359 { 00:27:26.359 "name": "BaseBdev2", 00:27:26.359 "uuid": "c909dbe8-1284-5977-8e81-cf74527961c6", 00:27:26.359 "is_configured": true, 00:27:26.359 "data_offset": 256, 00:27:26.359 "data_size": 7936 00:27:26.359 } 00:27:26.359 ] 00:27:26.359 }' 00:27:26.359 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:27:26.359 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:27:26.359 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:27:26.359 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:27:26.359 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@782 -- # killprocess 1004531 00:27:26.359 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@948 -- # '[' -z 1004531 ']' 00:27:26.359 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@952 -- # kill -0 1004531 00:27:26.359 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@953 -- # uname 00:27:26.359 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:27:26.359 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1004531 00:27:26.359 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:27:26.359 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:27:26.359 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1004531' 00:27:26.359 killing process with pid 1004531 00:27:26.616 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@967 -- # kill 1004531 00:27:26.616 Received shutdown signal, test time was about 60.000000 seconds 00:27:26.616 00:27:26.616 Latency(us) 00:27:26.616 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:26.616 =================================================================================================================== 00:27:26.616 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:27:26.616 [2024-07-15 11:38:09.954102] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:27:26.616 [2024-07-15 11:38:09.954194] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:26.616 [2024-07-15 11:38:09.954235] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:26.616 [2024-07-15 11:38:09.954248] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x160c260 name raid_bdev1, state offline 00:27:26.616 11:38:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@972 -- # wait 1004531 00:27:26.616 [2024-07-15 11:38:09.985607] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:27:26.616 11:38:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@784 -- # return 0 00:27:26.616 00:27:26.616 real 0m31.610s 00:27:26.616 user 0m49.289s 00:27:26.616 sys 0m5.158s 00:27:26.616 11:38:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1124 -- # xtrace_disable 00:27:26.616 11:38:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:27:26.616 ************************************ 00:27:26.616 END TEST raid_rebuild_test_sb_4k 00:27:26.616 ************************************ 00:27:26.872 11:38:10 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:27:26.872 11:38:10 bdev_raid -- bdev/bdev_raid.sh@904 -- # base_malloc_params='-m 32' 00:27:26.872 11:38:10 bdev_raid -- bdev/bdev_raid.sh@905 -- # run_test raid_state_function_test_sb_md_separate raid_state_function_test raid1 2 true 00:27:26.873 11:38:10 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:27:26.873 11:38:10 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:27:26.873 11:38:10 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:27:26.873 ************************************ 00:27:26.873 START TEST raid_state_function_test_sb_md_separate 00:27:26.873 ************************************ 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1123 -- # raid_state_function_test raid1 2 true 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@220 -- # local raid_level=raid1 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=2 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@222 -- # local superblock=true 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@226 -- # local strip_size 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@230 -- # '[' raid1 '!=' raid1 ']' 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@234 -- # strip_size=0 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@237 -- # '[' true = true ']' 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@238 -- # superblock_create_arg=-s 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@244 -- # raid_pid=1009036 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 1009036' 00:27:26.873 Process raid pid: 1009036 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@246 -- # waitforlisten 1009036 /var/tmp/spdk-raid.sock 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@829 -- # '[' -z 1009036 ']' 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@834 -- # local max_retries=100 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:27:26.873 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@838 -- # xtrace_disable 00:27:26.873 11:38:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:27:26.873 [2024-07-15 11:38:10.362303] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:27:26.873 [2024-07-15 11:38:10.362367] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:27.129 [2024-07-15 11:38:10.491041] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:27.129 [2024-07-15 11:38:10.593212] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:27:27.129 [2024-07-15 11:38:10.650462] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:27.129 [2024-07-15 11:38:10.650496] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:28.058 11:38:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:27:28.058 11:38:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@862 -- # return 0 00:27:28.059 11:38:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:27:28.059 [2024-07-15 11:38:11.510855] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:27:28.059 [2024-07-15 11:38:11.510898] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:27:28.059 [2024-07-15 11:38:11.510910] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:27:28.059 [2024-07-15 11:38:11.510922] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:27:28.059 11:38:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:27:28.059 11:38:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:27:28.059 11:38:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:27:28.059 11:38:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:28.059 11:38:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:28.059 11:38:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:27:28.059 11:38:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:28.059 11:38:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:28.059 11:38:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:28.059 11:38:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:28.059 11:38:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:28.059 11:38:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:28.315 11:38:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:28.315 "name": "Existed_Raid", 00:27:28.315 "uuid": "7302e542-0558-44a9-af47-ec083f19757b", 00:27:28.315 "strip_size_kb": 0, 00:27:28.315 "state": "configuring", 00:27:28.315 "raid_level": "raid1", 00:27:28.315 "superblock": true, 00:27:28.315 "num_base_bdevs": 2, 00:27:28.315 "num_base_bdevs_discovered": 0, 00:27:28.315 "num_base_bdevs_operational": 2, 00:27:28.315 "base_bdevs_list": [ 00:27:28.315 { 00:27:28.315 "name": "BaseBdev1", 00:27:28.315 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:28.315 "is_configured": false, 00:27:28.315 "data_offset": 0, 00:27:28.315 "data_size": 0 00:27:28.315 }, 00:27:28.315 { 00:27:28.315 "name": "BaseBdev2", 00:27:28.315 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:28.315 "is_configured": false, 00:27:28.315 "data_offset": 0, 00:27:28.315 "data_size": 0 00:27:28.315 } 00:27:28.315 ] 00:27:28.315 }' 00:27:28.315 11:38:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:28.315 11:38:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:27:28.878 11:38:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:27:29.134 [2024-07-15 11:38:12.593579] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:27:29.134 [2024-07-15 11:38:12.593613] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x23eda80 name Existed_Raid, state configuring 00:27:29.134 11:38:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:27:29.391 [2024-07-15 11:38:12.842263] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:27:29.391 [2024-07-15 11:38:12.842302] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:27:29.391 [2024-07-15 11:38:12.842313] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:27:29.391 [2024-07-15 11:38:12.842324] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:27:29.391 11:38:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -m 32 -b BaseBdev1 00:27:29.649 [2024-07-15 11:38:13.097422] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:29.649 BaseBdev1 00:27:29.649 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:27:29.649 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:27:29.649 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:27:29.649 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@899 -- # local i 00:27:29.649 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:27:29.649 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:27:29.649 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:27:29.905 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:27:30.162 [ 00:27:30.162 { 00:27:30.162 "name": "BaseBdev1", 00:27:30.162 "aliases": [ 00:27:30.162 "1c4a0d38-ab55-4740-bf3b-4eff53e118df" 00:27:30.162 ], 00:27:30.162 "product_name": "Malloc disk", 00:27:30.162 "block_size": 4096, 00:27:30.162 "num_blocks": 8192, 00:27:30.162 "uuid": "1c4a0d38-ab55-4740-bf3b-4eff53e118df", 00:27:30.162 "md_size": 32, 00:27:30.162 "md_interleave": false, 00:27:30.162 "dif_type": 0, 00:27:30.162 "assigned_rate_limits": { 00:27:30.162 "rw_ios_per_sec": 0, 00:27:30.162 "rw_mbytes_per_sec": 0, 00:27:30.162 "r_mbytes_per_sec": 0, 00:27:30.162 "w_mbytes_per_sec": 0 00:27:30.162 }, 00:27:30.162 "claimed": true, 00:27:30.162 "claim_type": "exclusive_write", 00:27:30.162 "zoned": false, 00:27:30.162 "supported_io_types": { 00:27:30.162 "read": true, 00:27:30.162 "write": true, 00:27:30.162 "unmap": true, 00:27:30.162 "flush": true, 00:27:30.162 "reset": true, 00:27:30.162 "nvme_admin": false, 00:27:30.162 "nvme_io": false, 00:27:30.162 "nvme_io_md": false, 00:27:30.162 "write_zeroes": true, 00:27:30.162 "zcopy": true, 00:27:30.162 "get_zone_info": false, 00:27:30.162 "zone_management": false, 00:27:30.162 "zone_append": false, 00:27:30.162 "compare": false, 00:27:30.162 "compare_and_write": false, 00:27:30.162 "abort": true, 00:27:30.162 "seek_hole": false, 00:27:30.162 "seek_data": false, 00:27:30.162 "copy": true, 00:27:30.162 "nvme_iov_md": false 00:27:30.162 }, 00:27:30.162 "memory_domains": [ 00:27:30.162 { 00:27:30.162 "dma_device_id": "system", 00:27:30.162 "dma_device_type": 1 00:27:30.162 }, 00:27:30.162 { 00:27:30.162 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:30.162 "dma_device_type": 2 00:27:30.162 } 00:27:30.162 ], 00:27:30.162 "driver_specific": {} 00:27:30.162 } 00:27:30.162 ] 00:27:30.162 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@905 -- # return 0 00:27:30.162 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:27:30.162 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:27:30.162 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:27:30.162 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:30.162 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:30.162 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:27:30.162 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:30.162 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:30.162 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:30.162 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:30.162 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:30.162 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:30.418 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:30.418 "name": "Existed_Raid", 00:27:30.418 "uuid": "8b29d420-f9b1-475c-9381-6d27bd380f8a", 00:27:30.418 "strip_size_kb": 0, 00:27:30.418 "state": "configuring", 00:27:30.418 "raid_level": "raid1", 00:27:30.418 "superblock": true, 00:27:30.418 "num_base_bdevs": 2, 00:27:30.418 "num_base_bdevs_discovered": 1, 00:27:30.418 "num_base_bdevs_operational": 2, 00:27:30.418 "base_bdevs_list": [ 00:27:30.419 { 00:27:30.419 "name": "BaseBdev1", 00:27:30.419 "uuid": "1c4a0d38-ab55-4740-bf3b-4eff53e118df", 00:27:30.419 "is_configured": true, 00:27:30.419 "data_offset": 256, 00:27:30.419 "data_size": 7936 00:27:30.419 }, 00:27:30.419 { 00:27:30.419 "name": "BaseBdev2", 00:27:30.419 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:30.419 "is_configured": false, 00:27:30.419 "data_offset": 0, 00:27:30.419 "data_size": 0 00:27:30.419 } 00:27:30.419 ] 00:27:30.419 }' 00:27:30.419 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:30.419 11:38:13 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:27:30.982 11:38:14 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:27:31.240 [2024-07-15 11:38:14.669611] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:27:31.240 [2024-07-15 11:38:14.669652] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x23ed350 name Existed_Raid, state configuring 00:27:31.240 11:38:14 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:27:31.497 [2024-07-15 11:38:14.930328] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:31.497 [2024-07-15 11:38:14.931746] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:27:31.497 [2024-07-15 11:38:14.931778] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:27:31.497 11:38:14 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:27:31.497 11:38:14 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:27:31.497 11:38:14 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:27:31.497 11:38:14 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:27:31.497 11:38:14 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:27:31.497 11:38:14 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:31.497 11:38:14 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:31.497 11:38:14 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:27:31.497 11:38:14 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:31.497 11:38:14 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:31.497 11:38:14 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:31.497 11:38:14 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:31.497 11:38:14 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:31.497 11:38:14 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:31.754 11:38:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:31.754 "name": "Existed_Raid", 00:27:31.754 "uuid": "c95cd731-166b-4a1d-91b5-426c0e0214a2", 00:27:31.754 "strip_size_kb": 0, 00:27:31.754 "state": "configuring", 00:27:31.754 "raid_level": "raid1", 00:27:31.754 "superblock": true, 00:27:31.754 "num_base_bdevs": 2, 00:27:31.754 "num_base_bdevs_discovered": 1, 00:27:31.754 "num_base_bdevs_operational": 2, 00:27:31.754 "base_bdevs_list": [ 00:27:31.754 { 00:27:31.754 "name": "BaseBdev1", 00:27:31.754 "uuid": "1c4a0d38-ab55-4740-bf3b-4eff53e118df", 00:27:31.754 "is_configured": true, 00:27:31.754 "data_offset": 256, 00:27:31.754 "data_size": 7936 00:27:31.754 }, 00:27:31.754 { 00:27:31.754 "name": "BaseBdev2", 00:27:31.754 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:31.754 "is_configured": false, 00:27:31.754 "data_offset": 0, 00:27:31.754 "data_size": 0 00:27:31.754 } 00:27:31.754 ] 00:27:31.754 }' 00:27:31.754 11:38:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:31.754 11:38:15 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:27:32.315 11:38:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -m 32 -b BaseBdev2 00:27:32.572 [2024-07-15 11:38:16.049316] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:32.572 [2024-07-15 11:38:16.049456] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x23ef210 00:27:32.572 [2024-07-15 11:38:16.049469] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:27:32.572 [2024-07-15 11:38:16.049532] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x23eec50 00:27:32.572 [2024-07-15 11:38:16.049629] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x23ef210 00:27:32.572 [2024-07-15 11:38:16.049639] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x23ef210 00:27:32.572 [2024-07-15 11:38:16.049701] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:32.572 BaseBdev2 00:27:32.572 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:27:32.572 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:27:32.572 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:27:32.572 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@899 -- # local i 00:27:32.572 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:27:32.572 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:27:32.572 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:27:32.828 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:27:33.085 [ 00:27:33.085 { 00:27:33.085 "name": "BaseBdev2", 00:27:33.085 "aliases": [ 00:27:33.085 "f09f1835-8164-4c00-95dc-296fb16f8945" 00:27:33.085 ], 00:27:33.085 "product_name": "Malloc disk", 00:27:33.085 "block_size": 4096, 00:27:33.085 "num_blocks": 8192, 00:27:33.085 "uuid": "f09f1835-8164-4c00-95dc-296fb16f8945", 00:27:33.085 "md_size": 32, 00:27:33.085 "md_interleave": false, 00:27:33.085 "dif_type": 0, 00:27:33.085 "assigned_rate_limits": { 00:27:33.085 "rw_ios_per_sec": 0, 00:27:33.085 "rw_mbytes_per_sec": 0, 00:27:33.085 "r_mbytes_per_sec": 0, 00:27:33.085 "w_mbytes_per_sec": 0 00:27:33.085 }, 00:27:33.085 "claimed": true, 00:27:33.085 "claim_type": "exclusive_write", 00:27:33.085 "zoned": false, 00:27:33.085 "supported_io_types": { 00:27:33.085 "read": true, 00:27:33.085 "write": true, 00:27:33.085 "unmap": true, 00:27:33.085 "flush": true, 00:27:33.085 "reset": true, 00:27:33.085 "nvme_admin": false, 00:27:33.085 "nvme_io": false, 00:27:33.085 "nvme_io_md": false, 00:27:33.085 "write_zeroes": true, 00:27:33.085 "zcopy": true, 00:27:33.085 "get_zone_info": false, 00:27:33.085 "zone_management": false, 00:27:33.085 "zone_append": false, 00:27:33.085 "compare": false, 00:27:33.085 "compare_and_write": false, 00:27:33.085 "abort": true, 00:27:33.085 "seek_hole": false, 00:27:33.085 "seek_data": false, 00:27:33.085 "copy": true, 00:27:33.085 "nvme_iov_md": false 00:27:33.085 }, 00:27:33.085 "memory_domains": [ 00:27:33.085 { 00:27:33.085 "dma_device_id": "system", 00:27:33.085 "dma_device_type": 1 00:27:33.085 }, 00:27:33.085 { 00:27:33.085 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:33.085 "dma_device_type": 2 00:27:33.085 } 00:27:33.085 ], 00:27:33.085 "driver_specific": {} 00:27:33.085 } 00:27:33.085 ] 00:27:33.085 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@905 -- # return 0 00:27:33.085 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:27:33.085 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:27:33.085 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:27:33.085 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:27:33.086 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:27:33.086 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:33.086 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:33.086 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:27:33.086 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:33.086 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:33.086 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:33.086 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:33.086 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:33.086 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:33.342 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:33.342 "name": "Existed_Raid", 00:27:33.342 "uuid": "c95cd731-166b-4a1d-91b5-426c0e0214a2", 00:27:33.342 "strip_size_kb": 0, 00:27:33.342 "state": "online", 00:27:33.342 "raid_level": "raid1", 00:27:33.342 "superblock": true, 00:27:33.342 "num_base_bdevs": 2, 00:27:33.342 "num_base_bdevs_discovered": 2, 00:27:33.342 "num_base_bdevs_operational": 2, 00:27:33.342 "base_bdevs_list": [ 00:27:33.342 { 00:27:33.342 "name": "BaseBdev1", 00:27:33.342 "uuid": "1c4a0d38-ab55-4740-bf3b-4eff53e118df", 00:27:33.342 "is_configured": true, 00:27:33.342 "data_offset": 256, 00:27:33.342 "data_size": 7936 00:27:33.342 }, 00:27:33.342 { 00:27:33.342 "name": "BaseBdev2", 00:27:33.342 "uuid": "f09f1835-8164-4c00-95dc-296fb16f8945", 00:27:33.342 "is_configured": true, 00:27:33.342 "data_offset": 256, 00:27:33.342 "data_size": 7936 00:27:33.342 } 00:27:33.342 ] 00:27:33.342 }' 00:27:33.342 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:33.342 11:38:16 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:27:33.905 11:38:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:27:33.905 11:38:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:27:33.905 11:38:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:27:33.905 11:38:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:27:33.905 11:38:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:27:33.905 11:38:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@198 -- # local name 00:27:33.905 11:38:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:27:33.905 11:38:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:27:34.162 [2024-07-15 11:38:17.561641] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:34.162 11:38:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:27:34.162 "name": "Existed_Raid", 00:27:34.162 "aliases": [ 00:27:34.162 "c95cd731-166b-4a1d-91b5-426c0e0214a2" 00:27:34.162 ], 00:27:34.162 "product_name": "Raid Volume", 00:27:34.162 "block_size": 4096, 00:27:34.162 "num_blocks": 7936, 00:27:34.162 "uuid": "c95cd731-166b-4a1d-91b5-426c0e0214a2", 00:27:34.162 "md_size": 32, 00:27:34.162 "md_interleave": false, 00:27:34.162 "dif_type": 0, 00:27:34.162 "assigned_rate_limits": { 00:27:34.162 "rw_ios_per_sec": 0, 00:27:34.162 "rw_mbytes_per_sec": 0, 00:27:34.162 "r_mbytes_per_sec": 0, 00:27:34.162 "w_mbytes_per_sec": 0 00:27:34.162 }, 00:27:34.162 "claimed": false, 00:27:34.162 "zoned": false, 00:27:34.162 "supported_io_types": { 00:27:34.162 "read": true, 00:27:34.162 "write": true, 00:27:34.162 "unmap": false, 00:27:34.162 "flush": false, 00:27:34.162 "reset": true, 00:27:34.162 "nvme_admin": false, 00:27:34.162 "nvme_io": false, 00:27:34.162 "nvme_io_md": false, 00:27:34.162 "write_zeroes": true, 00:27:34.162 "zcopy": false, 00:27:34.162 "get_zone_info": false, 00:27:34.162 "zone_management": false, 00:27:34.162 "zone_append": false, 00:27:34.162 "compare": false, 00:27:34.162 "compare_and_write": false, 00:27:34.162 "abort": false, 00:27:34.162 "seek_hole": false, 00:27:34.162 "seek_data": false, 00:27:34.162 "copy": false, 00:27:34.162 "nvme_iov_md": false 00:27:34.162 }, 00:27:34.162 "memory_domains": [ 00:27:34.162 { 00:27:34.162 "dma_device_id": "system", 00:27:34.162 "dma_device_type": 1 00:27:34.162 }, 00:27:34.162 { 00:27:34.162 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:34.162 "dma_device_type": 2 00:27:34.162 }, 00:27:34.162 { 00:27:34.162 "dma_device_id": "system", 00:27:34.162 "dma_device_type": 1 00:27:34.162 }, 00:27:34.162 { 00:27:34.162 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:34.162 "dma_device_type": 2 00:27:34.162 } 00:27:34.162 ], 00:27:34.162 "driver_specific": { 00:27:34.162 "raid": { 00:27:34.162 "uuid": "c95cd731-166b-4a1d-91b5-426c0e0214a2", 00:27:34.162 "strip_size_kb": 0, 00:27:34.162 "state": "online", 00:27:34.162 "raid_level": "raid1", 00:27:34.162 "superblock": true, 00:27:34.162 "num_base_bdevs": 2, 00:27:34.162 "num_base_bdevs_discovered": 2, 00:27:34.162 "num_base_bdevs_operational": 2, 00:27:34.162 "base_bdevs_list": [ 00:27:34.162 { 00:27:34.162 "name": "BaseBdev1", 00:27:34.162 "uuid": "1c4a0d38-ab55-4740-bf3b-4eff53e118df", 00:27:34.162 "is_configured": true, 00:27:34.162 "data_offset": 256, 00:27:34.162 "data_size": 7936 00:27:34.162 }, 00:27:34.162 { 00:27:34.162 "name": "BaseBdev2", 00:27:34.162 "uuid": "f09f1835-8164-4c00-95dc-296fb16f8945", 00:27:34.162 "is_configured": true, 00:27:34.162 "data_offset": 256, 00:27:34.162 "data_size": 7936 00:27:34.162 } 00:27:34.162 ] 00:27:34.162 } 00:27:34.162 } 00:27:34.162 }' 00:27:34.162 11:38:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:27:34.162 11:38:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:27:34.162 BaseBdev2' 00:27:34.162 11:38:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:27:34.162 11:38:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:27:34.162 11:38:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:27:34.419 11:38:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:27:34.419 "name": "BaseBdev1", 00:27:34.419 "aliases": [ 00:27:34.419 "1c4a0d38-ab55-4740-bf3b-4eff53e118df" 00:27:34.419 ], 00:27:34.419 "product_name": "Malloc disk", 00:27:34.419 "block_size": 4096, 00:27:34.419 "num_blocks": 8192, 00:27:34.419 "uuid": "1c4a0d38-ab55-4740-bf3b-4eff53e118df", 00:27:34.419 "md_size": 32, 00:27:34.419 "md_interleave": false, 00:27:34.419 "dif_type": 0, 00:27:34.419 "assigned_rate_limits": { 00:27:34.419 "rw_ios_per_sec": 0, 00:27:34.419 "rw_mbytes_per_sec": 0, 00:27:34.419 "r_mbytes_per_sec": 0, 00:27:34.419 "w_mbytes_per_sec": 0 00:27:34.419 }, 00:27:34.419 "claimed": true, 00:27:34.419 "claim_type": "exclusive_write", 00:27:34.419 "zoned": false, 00:27:34.419 "supported_io_types": { 00:27:34.419 "read": true, 00:27:34.419 "write": true, 00:27:34.419 "unmap": true, 00:27:34.419 "flush": true, 00:27:34.419 "reset": true, 00:27:34.419 "nvme_admin": false, 00:27:34.419 "nvme_io": false, 00:27:34.419 "nvme_io_md": false, 00:27:34.419 "write_zeroes": true, 00:27:34.419 "zcopy": true, 00:27:34.419 "get_zone_info": false, 00:27:34.419 "zone_management": false, 00:27:34.419 "zone_append": false, 00:27:34.419 "compare": false, 00:27:34.419 "compare_and_write": false, 00:27:34.419 "abort": true, 00:27:34.419 "seek_hole": false, 00:27:34.419 "seek_data": false, 00:27:34.419 "copy": true, 00:27:34.419 "nvme_iov_md": false 00:27:34.419 }, 00:27:34.419 "memory_domains": [ 00:27:34.419 { 00:27:34.419 "dma_device_id": "system", 00:27:34.419 "dma_device_type": 1 00:27:34.419 }, 00:27:34.419 { 00:27:34.419 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:34.419 "dma_device_type": 2 00:27:34.419 } 00:27:34.419 ], 00:27:34.419 "driver_specific": {} 00:27:34.419 }' 00:27:34.419 11:38:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:27:34.419 11:38:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:27:34.419 11:38:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@205 -- # [[ 4096 == 4096 ]] 00:27:34.419 11:38:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:27:34.675 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:27:34.675 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@206 -- # [[ 32 == 32 ]] 00:27:34.675 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:27:34.675 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:27:34.675 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@207 -- # [[ false == false ]] 00:27:34.675 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:27:34.675 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:27:34.675 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@208 -- # [[ 0 == 0 ]] 00:27:34.675 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:27:34.675 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:27:34.675 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:27:34.931 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:27:34.931 "name": "BaseBdev2", 00:27:34.931 "aliases": [ 00:27:34.931 "f09f1835-8164-4c00-95dc-296fb16f8945" 00:27:34.931 ], 00:27:34.931 "product_name": "Malloc disk", 00:27:34.931 "block_size": 4096, 00:27:34.931 "num_blocks": 8192, 00:27:34.931 "uuid": "f09f1835-8164-4c00-95dc-296fb16f8945", 00:27:34.931 "md_size": 32, 00:27:34.931 "md_interleave": false, 00:27:34.931 "dif_type": 0, 00:27:34.931 "assigned_rate_limits": { 00:27:34.931 "rw_ios_per_sec": 0, 00:27:34.931 "rw_mbytes_per_sec": 0, 00:27:34.931 "r_mbytes_per_sec": 0, 00:27:34.931 "w_mbytes_per_sec": 0 00:27:34.931 }, 00:27:34.931 "claimed": true, 00:27:34.931 "claim_type": "exclusive_write", 00:27:34.931 "zoned": false, 00:27:34.931 "supported_io_types": { 00:27:34.931 "read": true, 00:27:34.931 "write": true, 00:27:34.931 "unmap": true, 00:27:34.931 "flush": true, 00:27:34.931 "reset": true, 00:27:34.931 "nvme_admin": false, 00:27:34.931 "nvme_io": false, 00:27:34.931 "nvme_io_md": false, 00:27:34.931 "write_zeroes": true, 00:27:34.931 "zcopy": true, 00:27:34.931 "get_zone_info": false, 00:27:34.931 "zone_management": false, 00:27:34.931 "zone_append": false, 00:27:34.931 "compare": false, 00:27:34.931 "compare_and_write": false, 00:27:34.931 "abort": true, 00:27:34.931 "seek_hole": false, 00:27:34.931 "seek_data": false, 00:27:34.931 "copy": true, 00:27:34.931 "nvme_iov_md": false 00:27:34.931 }, 00:27:34.931 "memory_domains": [ 00:27:34.931 { 00:27:34.931 "dma_device_id": "system", 00:27:34.931 "dma_device_type": 1 00:27:34.931 }, 00:27:34.931 { 00:27:34.931 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:34.931 "dma_device_type": 2 00:27:34.931 } 00:27:34.931 ], 00:27:34.931 "driver_specific": {} 00:27:34.931 }' 00:27:34.931 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:27:35.186 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:27:35.186 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@205 -- # [[ 4096 == 4096 ]] 00:27:35.187 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:27:35.187 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:27:35.187 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@206 -- # [[ 32 == 32 ]] 00:27:35.187 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:27:35.187 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:27:35.187 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@207 -- # [[ false == false ]] 00:27:35.187 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:27:35.443 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:27:35.443 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@208 -- # [[ 0 == 0 ]] 00:27:35.443 11:38:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:27:35.701 [2024-07-15 11:38:19.053360] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:27:35.701 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@275 -- # local expected_state 00:27:35.701 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@276 -- # has_redundancy raid1 00:27:35.701 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@213 -- # case $1 in 00:27:35.701 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@214 -- # return 0 00:27:35.701 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@279 -- # expected_state=online 00:27:35.701 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:27:35.701 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:27:35.702 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:27:35.702 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:35.702 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:35.702 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:27:35.702 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:35.702 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:35.702 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:35.702 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:35.702 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:35.702 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:35.960 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:35.960 "name": "Existed_Raid", 00:27:35.960 "uuid": "c95cd731-166b-4a1d-91b5-426c0e0214a2", 00:27:35.960 "strip_size_kb": 0, 00:27:35.960 "state": "online", 00:27:35.960 "raid_level": "raid1", 00:27:35.960 "superblock": true, 00:27:35.960 "num_base_bdevs": 2, 00:27:35.960 "num_base_bdevs_discovered": 1, 00:27:35.960 "num_base_bdevs_operational": 1, 00:27:35.960 "base_bdevs_list": [ 00:27:35.960 { 00:27:35.960 "name": null, 00:27:35.960 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:35.960 "is_configured": false, 00:27:35.960 "data_offset": 256, 00:27:35.960 "data_size": 7936 00:27:35.960 }, 00:27:35.960 { 00:27:35.960 "name": "BaseBdev2", 00:27:35.960 "uuid": "f09f1835-8164-4c00-95dc-296fb16f8945", 00:27:35.960 "is_configured": true, 00:27:35.960 "data_offset": 256, 00:27:35.960 "data_size": 7936 00:27:35.960 } 00:27:35.960 ] 00:27:35.960 }' 00:27:35.960 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:35.960 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:27:36.524 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:27:36.524 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:27:36.524 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:36.524 11:38:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:27:36.783 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:27:36.783 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:27:36.783 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:27:37.042 [2024-07-15 11:38:20.415644] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:27:37.042 [2024-07-15 11:38:20.415728] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:37.042 [2024-07-15 11:38:20.428153] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:37.042 [2024-07-15 11:38:20.428189] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:37.042 [2024-07-15 11:38:20.428202] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x23ef210 name Existed_Raid, state offline 00:27:37.042 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:27:37.042 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:27:37.042 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:37.042 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:27:37.300 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:27:37.300 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:27:37.300 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@299 -- # '[' 2 -gt 2 ']' 00:27:37.300 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@341 -- # killprocess 1009036 00:27:37.300 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@948 -- # '[' -z 1009036 ']' 00:27:37.300 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@952 -- # kill -0 1009036 00:27:37.300 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@953 -- # uname 00:27:37.300 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:27:37.300 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1009036 00:27:37.300 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:27:37.300 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:27:37.300 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1009036' 00:27:37.300 killing process with pid 1009036 00:27:37.300 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@967 -- # kill 1009036 00:27:37.300 [2024-07-15 11:38:20.749259] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:27:37.300 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@972 -- # wait 1009036 00:27:37.300 [2024-07-15 11:38:20.750145] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:27:37.558 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@343 -- # return 0 00:27:37.558 00:27:37.558 real 0m10.663s 00:27:37.558 user 0m18.971s 00:27:37.558 sys 0m2.008s 00:27:37.558 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1124 -- # xtrace_disable 00:27:37.558 11:38:20 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:27:37.558 ************************************ 00:27:37.558 END TEST raid_state_function_test_sb_md_separate 00:27:37.558 ************************************ 00:27:37.558 11:38:20 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:27:37.558 11:38:20 bdev_raid -- bdev/bdev_raid.sh@906 -- # run_test raid_superblock_test_md_separate raid_superblock_test raid1 2 00:27:37.558 11:38:20 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:27:37.558 11:38:20 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:27:37.558 11:38:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:27:37.558 ************************************ 00:27:37.558 START TEST raid_superblock_test_md_separate 00:27:37.558 ************************************ 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1123 -- # raid_superblock_test raid1 2 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@392 -- # local raid_level=raid1 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@393 -- # local num_base_bdevs=2 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@394 -- # base_bdevs_malloc=() 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@394 -- # local base_bdevs_malloc 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # base_bdevs_pt=() 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # local base_bdevs_pt 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt_uuid=() 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt_uuid 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # local raid_bdev_name=raid_bdev1 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@398 -- # local strip_size 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@399 -- # local strip_size_create_arg 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@400 -- # local raid_bdev_uuid 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@401 -- # local raid_bdev 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@403 -- # '[' raid1 '!=' raid1 ']' 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@407 -- # strip_size=0 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@411 -- # raid_pid=1010655 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@412 -- # waitforlisten 1010655 /var/tmp/spdk-raid.sock 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@410 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -L bdev_raid 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@829 -- # '[' -z 1010655 ']' 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@834 -- # local max_retries=100 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:27:37.558 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@838 -- # xtrace_disable 00:27:37.558 11:38:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:27:37.558 [2024-07-15 11:38:21.100960] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:27:37.558 [2024-07-15 11:38:21.101028] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1010655 ] 00:27:37.816 [2024-07-15 11:38:21.228630] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:37.816 [2024-07-15 11:38:21.330605] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:27:37.816 [2024-07-15 11:38:21.393756] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:37.816 [2024-07-15 11:38:21.393792] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:38.452 11:38:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:27:38.452 11:38:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@862 -- # return 0 00:27:38.452 11:38:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@415 -- # (( i = 1 )) 00:27:38.452 11:38:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:27:38.452 11:38:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc1 00:27:38.452 11:38:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt1 00:27:38.452 11:38:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:27:38.452 11:38:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:27:38.452 11:38:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:27:38.452 11:38:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:27:38.452 11:38:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -m 32 -b malloc1 00:27:38.710 malloc1 00:27:38.710 11:38:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:27:38.968 [2024-07-15 11:38:22.515484] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:27:38.968 [2024-07-15 11:38:22.515532] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:38.968 [2024-07-15 11:38:22.515554] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1868830 00:27:38.968 [2024-07-15 11:38:22.515566] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:38.968 [2024-07-15 11:38:22.517148] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:38.968 [2024-07-15 11:38:22.517176] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:27:38.968 pt1 00:27:38.968 11:38:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:27:38.968 11:38:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:27:38.968 11:38:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc2 00:27:38.968 11:38:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt2 00:27:38.968 11:38:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:27:38.968 11:38:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:27:38.968 11:38:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:27:38.968 11:38:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:27:38.968 11:38:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -m 32 -b malloc2 00:27:39.225 malloc2 00:27:39.225 11:38:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:27:39.484 [2024-07-15 11:38:23.002251] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:27:39.484 [2024-07-15 11:38:23.002300] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:39.484 [2024-07-15 11:38:23.002319] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x185a250 00:27:39.484 [2024-07-15 11:38:23.002332] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:39.484 [2024-07-15 11:38:23.003707] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:39.484 [2024-07-15 11:38:23.003734] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:27:39.484 pt2 00:27:39.484 11:38:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:27:39.484 11:38:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:27:39.484 11:38:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@429 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'pt1 pt2' -n raid_bdev1 -s 00:27:39.742 [2024-07-15 11:38:23.246931] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:27:39.742 [2024-07-15 11:38:23.248309] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:27:39.742 [2024-07-15 11:38:23.248459] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x185ad20 00:27:39.742 [2024-07-15 11:38:23.248480] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:27:39.742 [2024-07-15 11:38:23.248558] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x184ea60 00:27:39.742 [2024-07-15 11:38:23.248679] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x185ad20 00:27:39.742 [2024-07-15 11:38:23.248690] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x185ad20 00:27:39.742 [2024-07-15 11:38:23.248762] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:39.742 11:38:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@430 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:27:39.742 11:38:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:27:39.742 11:38:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:27:39.742 11:38:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:39.742 11:38:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:39.742 11:38:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:27:39.742 11:38:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:39.742 11:38:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:39.742 11:38:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:39.742 11:38:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:39.742 11:38:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:39.742 11:38:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:40.000 11:38:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:40.000 "name": "raid_bdev1", 00:27:40.000 "uuid": "08e31d09-ef91-4656-bebd-14a9474273f4", 00:27:40.000 "strip_size_kb": 0, 00:27:40.000 "state": "online", 00:27:40.000 "raid_level": "raid1", 00:27:40.000 "superblock": true, 00:27:40.000 "num_base_bdevs": 2, 00:27:40.000 "num_base_bdevs_discovered": 2, 00:27:40.000 "num_base_bdevs_operational": 2, 00:27:40.000 "base_bdevs_list": [ 00:27:40.000 { 00:27:40.000 "name": "pt1", 00:27:40.000 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:40.000 "is_configured": true, 00:27:40.000 "data_offset": 256, 00:27:40.000 "data_size": 7936 00:27:40.000 }, 00:27:40.000 { 00:27:40.000 "name": "pt2", 00:27:40.000 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:40.000 "is_configured": true, 00:27:40.000 "data_offset": 256, 00:27:40.000 "data_size": 7936 00:27:40.000 } 00:27:40.000 ] 00:27:40.000 }' 00:27:40.000 11:38:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:40.000 11:38:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:27:40.567 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_properties raid_bdev1 00:27:40.567 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:27:40.567 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:27:40.567 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:27:40.567 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:27:40.567 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@198 -- # local name 00:27:40.567 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:27:40.567 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:27:40.825 [2024-07-15 11:38:24.346189] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:40.825 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:27:40.825 "name": "raid_bdev1", 00:27:40.825 "aliases": [ 00:27:40.825 "08e31d09-ef91-4656-bebd-14a9474273f4" 00:27:40.825 ], 00:27:40.825 "product_name": "Raid Volume", 00:27:40.825 "block_size": 4096, 00:27:40.825 "num_blocks": 7936, 00:27:40.825 "uuid": "08e31d09-ef91-4656-bebd-14a9474273f4", 00:27:40.825 "md_size": 32, 00:27:40.825 "md_interleave": false, 00:27:40.825 "dif_type": 0, 00:27:40.825 "assigned_rate_limits": { 00:27:40.825 "rw_ios_per_sec": 0, 00:27:40.825 "rw_mbytes_per_sec": 0, 00:27:40.825 "r_mbytes_per_sec": 0, 00:27:40.825 "w_mbytes_per_sec": 0 00:27:40.825 }, 00:27:40.825 "claimed": false, 00:27:40.825 "zoned": false, 00:27:40.825 "supported_io_types": { 00:27:40.825 "read": true, 00:27:40.825 "write": true, 00:27:40.825 "unmap": false, 00:27:40.825 "flush": false, 00:27:40.825 "reset": true, 00:27:40.825 "nvme_admin": false, 00:27:40.825 "nvme_io": false, 00:27:40.825 "nvme_io_md": false, 00:27:40.825 "write_zeroes": true, 00:27:40.825 "zcopy": false, 00:27:40.825 "get_zone_info": false, 00:27:40.825 "zone_management": false, 00:27:40.825 "zone_append": false, 00:27:40.825 "compare": false, 00:27:40.825 "compare_and_write": false, 00:27:40.825 "abort": false, 00:27:40.825 "seek_hole": false, 00:27:40.825 "seek_data": false, 00:27:40.825 "copy": false, 00:27:40.825 "nvme_iov_md": false 00:27:40.825 }, 00:27:40.825 "memory_domains": [ 00:27:40.825 { 00:27:40.825 "dma_device_id": "system", 00:27:40.825 "dma_device_type": 1 00:27:40.825 }, 00:27:40.825 { 00:27:40.825 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:40.825 "dma_device_type": 2 00:27:40.825 }, 00:27:40.825 { 00:27:40.825 "dma_device_id": "system", 00:27:40.825 "dma_device_type": 1 00:27:40.825 }, 00:27:40.825 { 00:27:40.825 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:40.825 "dma_device_type": 2 00:27:40.825 } 00:27:40.825 ], 00:27:40.825 "driver_specific": { 00:27:40.825 "raid": { 00:27:40.825 "uuid": "08e31d09-ef91-4656-bebd-14a9474273f4", 00:27:40.825 "strip_size_kb": 0, 00:27:40.825 "state": "online", 00:27:40.825 "raid_level": "raid1", 00:27:40.825 "superblock": true, 00:27:40.825 "num_base_bdevs": 2, 00:27:40.825 "num_base_bdevs_discovered": 2, 00:27:40.826 "num_base_bdevs_operational": 2, 00:27:40.826 "base_bdevs_list": [ 00:27:40.826 { 00:27:40.826 "name": "pt1", 00:27:40.826 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:40.826 "is_configured": true, 00:27:40.826 "data_offset": 256, 00:27:40.826 "data_size": 7936 00:27:40.826 }, 00:27:40.826 { 00:27:40.826 "name": "pt2", 00:27:40.826 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:40.826 "is_configured": true, 00:27:40.826 "data_offset": 256, 00:27:40.826 "data_size": 7936 00:27:40.826 } 00:27:40.826 ] 00:27:40.826 } 00:27:40.826 } 00:27:40.826 }' 00:27:40.826 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:27:40.826 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:27:40.826 pt2' 00:27:40.826 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:27:41.085 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:27:41.085 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:27:41.085 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:27:41.085 "name": "pt1", 00:27:41.085 "aliases": [ 00:27:41.085 "00000000-0000-0000-0000-000000000001" 00:27:41.085 ], 00:27:41.085 "product_name": "passthru", 00:27:41.085 "block_size": 4096, 00:27:41.085 "num_blocks": 8192, 00:27:41.085 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:41.085 "md_size": 32, 00:27:41.085 "md_interleave": false, 00:27:41.085 "dif_type": 0, 00:27:41.085 "assigned_rate_limits": { 00:27:41.085 "rw_ios_per_sec": 0, 00:27:41.085 "rw_mbytes_per_sec": 0, 00:27:41.085 "r_mbytes_per_sec": 0, 00:27:41.085 "w_mbytes_per_sec": 0 00:27:41.085 }, 00:27:41.085 "claimed": true, 00:27:41.085 "claim_type": "exclusive_write", 00:27:41.085 "zoned": false, 00:27:41.085 "supported_io_types": { 00:27:41.085 "read": true, 00:27:41.085 "write": true, 00:27:41.085 "unmap": true, 00:27:41.085 "flush": true, 00:27:41.085 "reset": true, 00:27:41.085 "nvme_admin": false, 00:27:41.085 "nvme_io": false, 00:27:41.085 "nvme_io_md": false, 00:27:41.085 "write_zeroes": true, 00:27:41.085 "zcopy": true, 00:27:41.085 "get_zone_info": false, 00:27:41.085 "zone_management": false, 00:27:41.085 "zone_append": false, 00:27:41.085 "compare": false, 00:27:41.085 "compare_and_write": false, 00:27:41.085 "abort": true, 00:27:41.085 "seek_hole": false, 00:27:41.085 "seek_data": false, 00:27:41.085 "copy": true, 00:27:41.085 "nvme_iov_md": false 00:27:41.085 }, 00:27:41.085 "memory_domains": [ 00:27:41.085 { 00:27:41.085 "dma_device_id": "system", 00:27:41.085 "dma_device_type": 1 00:27:41.085 }, 00:27:41.085 { 00:27:41.085 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:41.085 "dma_device_type": 2 00:27:41.085 } 00:27:41.085 ], 00:27:41.085 "driver_specific": { 00:27:41.085 "passthru": { 00:27:41.085 "name": "pt1", 00:27:41.085 "base_bdev_name": "malloc1" 00:27:41.085 } 00:27:41.085 } 00:27:41.085 }' 00:27:41.085 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:27:41.343 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:27:41.343 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@205 -- # [[ 4096 == 4096 ]] 00:27:41.343 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:27:41.343 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:27:41.343 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@206 -- # [[ 32 == 32 ]] 00:27:41.343 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:27:41.343 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:27:41.343 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@207 -- # [[ false == false ]] 00:27:41.343 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:27:41.601 11:38:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:27:41.601 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@208 -- # [[ 0 == 0 ]] 00:27:41.601 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:27:41.601 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:27:41.601 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:27:41.860 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:27:41.860 "name": "pt2", 00:27:41.860 "aliases": [ 00:27:41.860 "00000000-0000-0000-0000-000000000002" 00:27:41.860 ], 00:27:41.860 "product_name": "passthru", 00:27:41.860 "block_size": 4096, 00:27:41.860 "num_blocks": 8192, 00:27:41.860 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:41.860 "md_size": 32, 00:27:41.860 "md_interleave": false, 00:27:41.860 "dif_type": 0, 00:27:41.860 "assigned_rate_limits": { 00:27:41.860 "rw_ios_per_sec": 0, 00:27:41.860 "rw_mbytes_per_sec": 0, 00:27:41.860 "r_mbytes_per_sec": 0, 00:27:41.860 "w_mbytes_per_sec": 0 00:27:41.860 }, 00:27:41.860 "claimed": true, 00:27:41.860 "claim_type": "exclusive_write", 00:27:41.860 "zoned": false, 00:27:41.860 "supported_io_types": { 00:27:41.860 "read": true, 00:27:41.860 "write": true, 00:27:41.860 "unmap": true, 00:27:41.860 "flush": true, 00:27:41.860 "reset": true, 00:27:41.860 "nvme_admin": false, 00:27:41.860 "nvme_io": false, 00:27:41.860 "nvme_io_md": false, 00:27:41.860 "write_zeroes": true, 00:27:41.860 "zcopy": true, 00:27:41.860 "get_zone_info": false, 00:27:41.860 "zone_management": false, 00:27:41.860 "zone_append": false, 00:27:41.860 "compare": false, 00:27:41.860 "compare_and_write": false, 00:27:41.860 "abort": true, 00:27:41.860 "seek_hole": false, 00:27:41.860 "seek_data": false, 00:27:41.860 "copy": true, 00:27:41.860 "nvme_iov_md": false 00:27:41.860 }, 00:27:41.860 "memory_domains": [ 00:27:41.860 { 00:27:41.860 "dma_device_id": "system", 00:27:41.860 "dma_device_type": 1 00:27:41.860 }, 00:27:41.860 { 00:27:41.860 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:41.860 "dma_device_type": 2 00:27:41.860 } 00:27:41.860 ], 00:27:41.860 "driver_specific": { 00:27:41.860 "passthru": { 00:27:41.860 "name": "pt2", 00:27:41.860 "base_bdev_name": "malloc2" 00:27:41.860 } 00:27:41.860 } 00:27:41.860 }' 00:27:41.860 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:27:41.860 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:27:41.860 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@205 -- # [[ 4096 == 4096 ]] 00:27:41.860 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:27:41.860 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:27:41.860 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@206 -- # [[ 32 == 32 ]] 00:27:41.860 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:27:42.118 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:27:42.118 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@207 -- # [[ false == false ]] 00:27:42.118 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:27:42.118 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:27:42.118 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@208 -- # [[ 0 == 0 ]] 00:27:42.118 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@434 -- # jq -r '.[] | .uuid' 00:27:42.118 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@434 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:27:42.377 [2024-07-15 11:38:25.838130] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:42.377 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@434 -- # raid_bdev_uuid=08e31d09-ef91-4656-bebd-14a9474273f4 00:27:42.377 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # '[' -z 08e31d09-ef91-4656-bebd-14a9474273f4 ']' 00:27:42.377 11:38:25 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@440 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:27:42.636 [2024-07-15 11:38:26.086540] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:42.636 [2024-07-15 11:38:26.086562] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:42.636 [2024-07-15 11:38:26.086615] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:42.636 [2024-07-15 11:38:26.086668] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:42.636 [2024-07-15 11:38:26.086680] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x185ad20 name raid_bdev1, state offline 00:27:42.636 11:38:26 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@441 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:42.636 11:38:26 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@441 -- # jq -r '.[]' 00:27:42.894 11:38:26 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@441 -- # raid_bdev= 00:27:42.894 11:38:26 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # '[' -n '' ']' 00:27:42.894 11:38:26 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:27:42.894 11:38:26 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt1 00:27:43.152 11:38:26 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:27:43.153 11:38:26 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:27:43.411 11:38:26 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@450 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs 00:27:43.411 11:38:26 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@450 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:27:43.670 11:38:27 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@450 -- # '[' false == true ']' 00:27:43.670 11:38:27 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@456 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'malloc1 malloc2' -n raid_bdev1 00:27:43.670 11:38:27 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@648 -- # local es=0 00:27:43.670 11:38:27 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'malloc1 malloc2' -n raid_bdev1 00:27:43.670 11:38:27 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:27:43.670 11:38:27 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:27:43.670 11:38:27 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:27:43.670 11:38:27 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:27:43.670 11:38:27 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:27:43.670 11:38:27 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:27:43.670 11:38:27 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:27:43.670 11:38:27 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py ]] 00:27:43.670 11:38:27 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'malloc1 malloc2' -n raid_bdev1 00:27:43.927 [2024-07-15 11:38:27.313733] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:27:43.927 [2024-07-15 11:38:27.315076] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:27:43.927 [2024-07-15 11:38:27.315130] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:27:43.927 [2024-07-15 11:38:27.315168] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:27:43.927 [2024-07-15 11:38:27.315187] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:43.927 [2024-07-15 11:38:27.315197] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x16caed0 name raid_bdev1, state configuring 00:27:43.927 request: 00:27:43.927 { 00:27:43.927 "name": "raid_bdev1", 00:27:43.927 "raid_level": "raid1", 00:27:43.927 "base_bdevs": [ 00:27:43.927 "malloc1", 00:27:43.927 "malloc2" 00:27:43.927 ], 00:27:43.927 "superblock": false, 00:27:43.927 "method": "bdev_raid_create", 00:27:43.927 "req_id": 1 00:27:43.927 } 00:27:43.927 Got JSON-RPC error response 00:27:43.927 response: 00:27:43.927 { 00:27:43.927 "code": -17, 00:27:43.927 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:27:43.927 } 00:27:43.927 11:38:27 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@651 -- # es=1 00:27:43.927 11:38:27 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:27:43.927 11:38:27 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:27:43.927 11:38:27 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:27:43.927 11:38:27 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@458 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:43.927 11:38:27 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@458 -- # jq -r '.[]' 00:27:44.189 11:38:27 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@458 -- # raid_bdev= 00:27:44.189 11:38:27 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # '[' -n '' ']' 00:27:44.190 11:38:27 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@464 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:27:44.458 [2024-07-15 11:38:27.802977] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:27:44.458 [2024-07-15 11:38:27.803015] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:44.458 [2024-07-15 11:38:27.803032] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1868ee0 00:27:44.458 [2024-07-15 11:38:27.803044] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:44.458 [2024-07-15 11:38:27.804488] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:44.458 [2024-07-15 11:38:27.804514] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:27:44.458 [2024-07-15 11:38:27.804557] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:27:44.458 [2024-07-15 11:38:27.804583] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:27:44.458 pt1 00:27:44.458 11:38:27 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@467 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:27:44.458 11:38:27 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:27:44.458 11:38:27 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:27:44.458 11:38:27 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:44.458 11:38:27 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:44.458 11:38:27 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:27:44.458 11:38:27 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:44.458 11:38:27 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:44.458 11:38:27 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:44.458 11:38:27 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:44.458 11:38:27 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:44.458 11:38:27 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:44.718 11:38:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:44.718 "name": "raid_bdev1", 00:27:44.718 "uuid": "08e31d09-ef91-4656-bebd-14a9474273f4", 00:27:44.718 "strip_size_kb": 0, 00:27:44.718 "state": "configuring", 00:27:44.718 "raid_level": "raid1", 00:27:44.718 "superblock": true, 00:27:44.718 "num_base_bdevs": 2, 00:27:44.718 "num_base_bdevs_discovered": 1, 00:27:44.718 "num_base_bdevs_operational": 2, 00:27:44.718 "base_bdevs_list": [ 00:27:44.718 { 00:27:44.718 "name": "pt1", 00:27:44.718 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:44.718 "is_configured": true, 00:27:44.718 "data_offset": 256, 00:27:44.718 "data_size": 7936 00:27:44.718 }, 00:27:44.718 { 00:27:44.718 "name": null, 00:27:44.718 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:44.718 "is_configured": false, 00:27:44.718 "data_offset": 256, 00:27:44.718 "data_size": 7936 00:27:44.718 } 00:27:44.718 ] 00:27:44.718 }' 00:27:44.718 11:38:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:44.718 11:38:28 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:27:45.284 11:38:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@469 -- # '[' 2 -gt 2 ']' 00:27:45.284 11:38:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@477 -- # (( i = 1 )) 00:27:45.284 11:38:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:27:45.284 11:38:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:27:45.544 [2024-07-15 11:38:28.905910] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:27:45.544 [2024-07-15 11:38:28.905963] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:45.544 [2024-07-15 11:38:28.905981] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x16cb490 00:27:45.544 [2024-07-15 11:38:28.905994] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:45.544 [2024-07-15 11:38:28.906181] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:45.544 [2024-07-15 11:38:28.906198] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:27:45.544 [2024-07-15 11:38:28.906239] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:27:45.544 [2024-07-15 11:38:28.906258] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:27:45.544 [2024-07-15 11:38:28.906346] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x184f5d0 00:27:45.544 [2024-07-15 11:38:28.906356] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:27:45.544 [2024-07-15 11:38:28.906409] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1850800 00:27:45.544 [2024-07-15 11:38:28.906511] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x184f5d0 00:27:45.544 [2024-07-15 11:38:28.906521] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x184f5d0 00:27:45.544 [2024-07-15 11:38:28.906588] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:45.544 pt2 00:27:45.544 11:38:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:27:45.544 11:38:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:27:45.544 11:38:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@482 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:27:45.544 11:38:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:27:45.544 11:38:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:27:45.544 11:38:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:45.544 11:38:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:45.544 11:38:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:27:45.544 11:38:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:45.544 11:38:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:45.544 11:38:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:45.544 11:38:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:45.544 11:38:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:45.544 11:38:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:45.803 11:38:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:45.803 "name": "raid_bdev1", 00:27:45.803 "uuid": "08e31d09-ef91-4656-bebd-14a9474273f4", 00:27:45.803 "strip_size_kb": 0, 00:27:45.803 "state": "online", 00:27:45.803 "raid_level": "raid1", 00:27:45.803 "superblock": true, 00:27:45.803 "num_base_bdevs": 2, 00:27:45.803 "num_base_bdevs_discovered": 2, 00:27:45.803 "num_base_bdevs_operational": 2, 00:27:45.803 "base_bdevs_list": [ 00:27:45.803 { 00:27:45.803 "name": "pt1", 00:27:45.803 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:45.803 "is_configured": true, 00:27:45.803 "data_offset": 256, 00:27:45.803 "data_size": 7936 00:27:45.803 }, 00:27:45.803 { 00:27:45.803 "name": "pt2", 00:27:45.803 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:45.803 "is_configured": true, 00:27:45.803 "data_offset": 256, 00:27:45.803 "data_size": 7936 00:27:45.803 } 00:27:45.803 ] 00:27:45.803 }' 00:27:45.803 11:38:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:45.803 11:38:29 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:27:46.370 11:38:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_properties raid_bdev1 00:27:46.370 11:38:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:27:46.370 11:38:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:27:46.370 11:38:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:27:46.370 11:38:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:27:46.370 11:38:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@198 -- # local name 00:27:46.370 11:38:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:27:46.370 11:38:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:27:46.629 [2024-07-15 11:38:29.993041] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:46.629 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:27:46.629 "name": "raid_bdev1", 00:27:46.629 "aliases": [ 00:27:46.629 "08e31d09-ef91-4656-bebd-14a9474273f4" 00:27:46.629 ], 00:27:46.629 "product_name": "Raid Volume", 00:27:46.629 "block_size": 4096, 00:27:46.629 "num_blocks": 7936, 00:27:46.629 "uuid": "08e31d09-ef91-4656-bebd-14a9474273f4", 00:27:46.629 "md_size": 32, 00:27:46.629 "md_interleave": false, 00:27:46.629 "dif_type": 0, 00:27:46.629 "assigned_rate_limits": { 00:27:46.629 "rw_ios_per_sec": 0, 00:27:46.629 "rw_mbytes_per_sec": 0, 00:27:46.629 "r_mbytes_per_sec": 0, 00:27:46.629 "w_mbytes_per_sec": 0 00:27:46.629 }, 00:27:46.629 "claimed": false, 00:27:46.629 "zoned": false, 00:27:46.629 "supported_io_types": { 00:27:46.629 "read": true, 00:27:46.629 "write": true, 00:27:46.629 "unmap": false, 00:27:46.629 "flush": false, 00:27:46.629 "reset": true, 00:27:46.629 "nvme_admin": false, 00:27:46.629 "nvme_io": false, 00:27:46.629 "nvme_io_md": false, 00:27:46.629 "write_zeroes": true, 00:27:46.629 "zcopy": false, 00:27:46.629 "get_zone_info": false, 00:27:46.629 "zone_management": false, 00:27:46.629 "zone_append": false, 00:27:46.629 "compare": false, 00:27:46.629 "compare_and_write": false, 00:27:46.629 "abort": false, 00:27:46.629 "seek_hole": false, 00:27:46.629 "seek_data": false, 00:27:46.629 "copy": false, 00:27:46.629 "nvme_iov_md": false 00:27:46.629 }, 00:27:46.629 "memory_domains": [ 00:27:46.629 { 00:27:46.629 "dma_device_id": "system", 00:27:46.629 "dma_device_type": 1 00:27:46.629 }, 00:27:46.629 { 00:27:46.629 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:46.629 "dma_device_type": 2 00:27:46.629 }, 00:27:46.629 { 00:27:46.629 "dma_device_id": "system", 00:27:46.629 "dma_device_type": 1 00:27:46.629 }, 00:27:46.629 { 00:27:46.629 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:46.629 "dma_device_type": 2 00:27:46.629 } 00:27:46.629 ], 00:27:46.629 "driver_specific": { 00:27:46.629 "raid": { 00:27:46.629 "uuid": "08e31d09-ef91-4656-bebd-14a9474273f4", 00:27:46.629 "strip_size_kb": 0, 00:27:46.629 "state": "online", 00:27:46.629 "raid_level": "raid1", 00:27:46.629 "superblock": true, 00:27:46.629 "num_base_bdevs": 2, 00:27:46.629 "num_base_bdevs_discovered": 2, 00:27:46.629 "num_base_bdevs_operational": 2, 00:27:46.629 "base_bdevs_list": [ 00:27:46.629 { 00:27:46.629 "name": "pt1", 00:27:46.629 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:46.629 "is_configured": true, 00:27:46.629 "data_offset": 256, 00:27:46.629 "data_size": 7936 00:27:46.629 }, 00:27:46.629 { 00:27:46.629 "name": "pt2", 00:27:46.629 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:46.629 "is_configured": true, 00:27:46.629 "data_offset": 256, 00:27:46.629 "data_size": 7936 00:27:46.629 } 00:27:46.629 ] 00:27:46.629 } 00:27:46.629 } 00:27:46.629 }' 00:27:46.629 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:27:46.629 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:27:46.629 pt2' 00:27:46.629 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:27:46.629 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:27:46.629 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:27:46.912 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:27:46.912 "name": "pt1", 00:27:46.912 "aliases": [ 00:27:46.912 "00000000-0000-0000-0000-000000000001" 00:27:46.912 ], 00:27:46.912 "product_name": "passthru", 00:27:46.912 "block_size": 4096, 00:27:46.912 "num_blocks": 8192, 00:27:46.912 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:46.912 "md_size": 32, 00:27:46.912 "md_interleave": false, 00:27:46.912 "dif_type": 0, 00:27:46.912 "assigned_rate_limits": { 00:27:46.912 "rw_ios_per_sec": 0, 00:27:46.912 "rw_mbytes_per_sec": 0, 00:27:46.912 "r_mbytes_per_sec": 0, 00:27:46.912 "w_mbytes_per_sec": 0 00:27:46.912 }, 00:27:46.912 "claimed": true, 00:27:46.912 "claim_type": "exclusive_write", 00:27:46.912 "zoned": false, 00:27:46.912 "supported_io_types": { 00:27:46.912 "read": true, 00:27:46.912 "write": true, 00:27:46.912 "unmap": true, 00:27:46.912 "flush": true, 00:27:46.912 "reset": true, 00:27:46.912 "nvme_admin": false, 00:27:46.912 "nvme_io": false, 00:27:46.912 "nvme_io_md": false, 00:27:46.912 "write_zeroes": true, 00:27:46.912 "zcopy": true, 00:27:46.912 "get_zone_info": false, 00:27:46.912 "zone_management": false, 00:27:46.912 "zone_append": false, 00:27:46.912 "compare": false, 00:27:46.912 "compare_and_write": false, 00:27:46.912 "abort": true, 00:27:46.912 "seek_hole": false, 00:27:46.912 "seek_data": false, 00:27:46.912 "copy": true, 00:27:46.912 "nvme_iov_md": false 00:27:46.912 }, 00:27:46.912 "memory_domains": [ 00:27:46.912 { 00:27:46.912 "dma_device_id": "system", 00:27:46.912 "dma_device_type": 1 00:27:46.912 }, 00:27:46.912 { 00:27:46.912 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:46.912 "dma_device_type": 2 00:27:46.912 } 00:27:46.912 ], 00:27:46.912 "driver_specific": { 00:27:46.912 "passthru": { 00:27:46.912 "name": "pt1", 00:27:46.912 "base_bdev_name": "malloc1" 00:27:46.912 } 00:27:46.912 } 00:27:46.912 }' 00:27:46.912 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:27:46.912 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:27:46.912 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@205 -- # [[ 4096 == 4096 ]] 00:27:46.912 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:27:46.912 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:27:46.912 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@206 -- # [[ 32 == 32 ]] 00:27:46.912 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:27:47.171 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:27:47.171 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@207 -- # [[ false == false ]] 00:27:47.171 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:27:47.171 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:27:47.171 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@208 -- # [[ 0 == 0 ]] 00:27:47.171 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:27:47.171 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:27:47.171 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:27:47.429 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:27:47.429 "name": "pt2", 00:27:47.429 "aliases": [ 00:27:47.429 "00000000-0000-0000-0000-000000000002" 00:27:47.429 ], 00:27:47.429 "product_name": "passthru", 00:27:47.429 "block_size": 4096, 00:27:47.429 "num_blocks": 8192, 00:27:47.429 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:47.429 "md_size": 32, 00:27:47.429 "md_interleave": false, 00:27:47.429 "dif_type": 0, 00:27:47.429 "assigned_rate_limits": { 00:27:47.429 "rw_ios_per_sec": 0, 00:27:47.429 "rw_mbytes_per_sec": 0, 00:27:47.429 "r_mbytes_per_sec": 0, 00:27:47.429 "w_mbytes_per_sec": 0 00:27:47.429 }, 00:27:47.429 "claimed": true, 00:27:47.429 "claim_type": "exclusive_write", 00:27:47.429 "zoned": false, 00:27:47.429 "supported_io_types": { 00:27:47.429 "read": true, 00:27:47.429 "write": true, 00:27:47.429 "unmap": true, 00:27:47.429 "flush": true, 00:27:47.429 "reset": true, 00:27:47.429 "nvme_admin": false, 00:27:47.429 "nvme_io": false, 00:27:47.429 "nvme_io_md": false, 00:27:47.429 "write_zeroes": true, 00:27:47.429 "zcopy": true, 00:27:47.429 "get_zone_info": false, 00:27:47.429 "zone_management": false, 00:27:47.429 "zone_append": false, 00:27:47.429 "compare": false, 00:27:47.429 "compare_and_write": false, 00:27:47.429 "abort": true, 00:27:47.429 "seek_hole": false, 00:27:47.429 "seek_data": false, 00:27:47.429 "copy": true, 00:27:47.429 "nvme_iov_md": false 00:27:47.429 }, 00:27:47.429 "memory_domains": [ 00:27:47.429 { 00:27:47.429 "dma_device_id": "system", 00:27:47.429 "dma_device_type": 1 00:27:47.429 }, 00:27:47.429 { 00:27:47.429 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:47.429 "dma_device_type": 2 00:27:47.429 } 00:27:47.429 ], 00:27:47.429 "driver_specific": { 00:27:47.429 "passthru": { 00:27:47.429 "name": "pt2", 00:27:47.429 "base_bdev_name": "malloc2" 00:27:47.429 } 00:27:47.429 } 00:27:47.429 }' 00:27:47.429 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:27:47.429 11:38:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:27:47.429 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@205 -- # [[ 4096 == 4096 ]] 00:27:47.429 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:27:47.685 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:27:47.685 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@206 -- # [[ 32 == 32 ]] 00:27:47.685 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:27:47.685 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:27:47.685 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@207 -- # [[ false == false ]] 00:27:47.685 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:27:47.685 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:27:47.685 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@208 -- # [[ 0 == 0 ]] 00:27:47.685 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@486 -- # jq -r '.[] | .uuid' 00:27:47.685 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@486 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:27:47.942 [2024-07-15 11:38:31.485000] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:47.942 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@486 -- # '[' 08e31d09-ef91-4656-bebd-14a9474273f4 '!=' 08e31d09-ef91-4656-bebd-14a9474273f4 ']' 00:27:47.942 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@490 -- # has_redundancy raid1 00:27:47.942 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@213 -- # case $1 in 00:27:47.942 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@214 -- # return 0 00:27:47.942 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@492 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt1 00:27:48.199 [2024-07-15 11:38:31.733415] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:27:48.199 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@495 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:48.199 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:27:48.199 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:27:48.199 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:48.199 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:48.199 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:27:48.199 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:48.199 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:48.199 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:48.199 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:48.199 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:48.199 11:38:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:48.456 11:38:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:48.456 "name": "raid_bdev1", 00:27:48.456 "uuid": "08e31d09-ef91-4656-bebd-14a9474273f4", 00:27:48.456 "strip_size_kb": 0, 00:27:48.456 "state": "online", 00:27:48.456 "raid_level": "raid1", 00:27:48.456 "superblock": true, 00:27:48.456 "num_base_bdevs": 2, 00:27:48.456 "num_base_bdevs_discovered": 1, 00:27:48.456 "num_base_bdevs_operational": 1, 00:27:48.456 "base_bdevs_list": [ 00:27:48.456 { 00:27:48.456 "name": null, 00:27:48.456 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:48.456 "is_configured": false, 00:27:48.456 "data_offset": 256, 00:27:48.456 "data_size": 7936 00:27:48.456 }, 00:27:48.456 { 00:27:48.456 "name": "pt2", 00:27:48.456 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:48.456 "is_configured": true, 00:27:48.456 "data_offset": 256, 00:27:48.456 "data_size": 7936 00:27:48.456 } 00:27:48.456 ] 00:27:48.456 }' 00:27:48.456 11:38:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:48.456 11:38:32 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:27:49.020 11:38:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@498 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:27:49.276 [2024-07-15 11:38:32.816261] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:49.276 [2024-07-15 11:38:32.816288] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:49.276 [2024-07-15 11:38:32.816339] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:49.276 [2024-07-15 11:38:32.816384] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:49.276 [2024-07-15 11:38:32.816396] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x184f5d0 name raid_bdev1, state offline 00:27:49.276 11:38:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@499 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:49.276 11:38:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@499 -- # jq -r '.[]' 00:27:49.533 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@499 -- # raid_bdev= 00:27:49.533 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # '[' -n '' ']' 00:27:49.533 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@505 -- # (( i = 1 )) 00:27:49.533 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@505 -- # (( i < num_base_bdevs )) 00:27:49.533 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:27:49.789 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@505 -- # (( i++ )) 00:27:49.789 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@505 -- # (( i < num_base_bdevs )) 00:27:49.789 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@510 -- # (( i = 1 )) 00:27:49.789 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@510 -- # (( i < num_base_bdevs - 1 )) 00:27:49.789 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@518 -- # i=1 00:27:49.789 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@519 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:27:50.045 [2024-07-15 11:38:33.546162] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:27:50.045 [2024-07-15 11:38:33.546206] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:50.045 [2024-07-15 11:38:33.546224] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x184d660 00:27:50.045 [2024-07-15 11:38:33.546236] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:50.045 [2024-07-15 11:38:33.547665] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:50.045 [2024-07-15 11:38:33.547693] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:27:50.045 [2024-07-15 11:38:33.547739] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:27:50.045 [2024-07-15 11:38:33.547765] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:27:50.045 [2024-07-15 11:38:33.547842] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x184fd10 00:27:50.045 [2024-07-15 11:38:33.547852] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:27:50.045 [2024-07-15 11:38:33.547922] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1850560 00:27:50.045 [2024-07-15 11:38:33.548029] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x184fd10 00:27:50.045 [2024-07-15 11:38:33.548039] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x184fd10 00:27:50.045 [2024-07-15 11:38:33.548106] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:50.045 pt2 00:27:50.045 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@522 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:50.045 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:27:50.045 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:27:50.045 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:50.045 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:50.045 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:27:50.045 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:50.045 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:50.045 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:50.045 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:50.045 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:50.045 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:50.301 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:50.301 "name": "raid_bdev1", 00:27:50.301 "uuid": "08e31d09-ef91-4656-bebd-14a9474273f4", 00:27:50.301 "strip_size_kb": 0, 00:27:50.301 "state": "online", 00:27:50.301 "raid_level": "raid1", 00:27:50.301 "superblock": true, 00:27:50.301 "num_base_bdevs": 2, 00:27:50.301 "num_base_bdevs_discovered": 1, 00:27:50.301 "num_base_bdevs_operational": 1, 00:27:50.301 "base_bdevs_list": [ 00:27:50.301 { 00:27:50.301 "name": null, 00:27:50.301 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:50.301 "is_configured": false, 00:27:50.302 "data_offset": 256, 00:27:50.302 "data_size": 7936 00:27:50.302 }, 00:27:50.302 { 00:27:50.302 "name": "pt2", 00:27:50.302 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:50.302 "is_configured": true, 00:27:50.302 "data_offset": 256, 00:27:50.302 "data_size": 7936 00:27:50.302 } 00:27:50.302 ] 00:27:50.302 }' 00:27:50.302 11:38:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:50.302 11:38:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:27:50.865 11:38:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@525 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:27:51.122 [2024-07-15 11:38:34.633040] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:51.122 [2024-07-15 11:38:34.633065] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:51.122 [2024-07-15 11:38:34.633115] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:51.122 [2024-07-15 11:38:34.633156] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:51.122 [2024-07-15 11:38:34.633168] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x184fd10 name raid_bdev1, state offline 00:27:51.122 11:38:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@526 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:51.122 11:38:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@526 -- # jq -r '.[]' 00:27:51.379 11:38:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@526 -- # raid_bdev= 00:27:51.379 11:38:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # '[' -n '' ']' 00:27:51.379 11:38:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@531 -- # '[' 2 -gt 2 ']' 00:27:51.379 11:38:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@539 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:27:51.638 [2024-07-15 11:38:35.130335] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:27:51.638 [2024-07-15 11:38:35.130374] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:51.638 [2024-07-15 11:38:35.130391] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x184e760 00:27:51.638 [2024-07-15 11:38:35.130403] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:51.638 [2024-07-15 11:38:35.131797] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:51.638 [2024-07-15 11:38:35.131822] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:27:51.638 [2024-07-15 11:38:35.131866] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:27:51.638 [2024-07-15 11:38:35.131891] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:27:51.638 [2024-07-15 11:38:35.131985] bdev_raid.c:3547:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:27:51.638 [2024-07-15 11:38:35.131998] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:51.638 [2024-07-15 11:38:35.132011] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1850850 name raid_bdev1, state configuring 00:27:51.638 [2024-07-15 11:38:35.132034] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:27:51.638 [2024-07-15 11:38:35.132083] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x184f850 00:27:51.638 [2024-07-15 11:38:35.132093] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:27:51.638 [2024-07-15 11:38:35.132146] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x18503b0 00:27:51.638 [2024-07-15 11:38:35.132243] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x184f850 00:27:51.638 [2024-07-15 11:38:35.132252] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x184f850 00:27:51.638 [2024-07-15 11:38:35.132323] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:51.638 pt1 00:27:51.639 11:38:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@541 -- # '[' 2 -gt 2 ']' 00:27:51.639 11:38:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@553 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:51.639 11:38:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:27:51.639 11:38:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:27:51.639 11:38:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:51.639 11:38:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:51.639 11:38:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:27:51.639 11:38:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:51.639 11:38:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:51.639 11:38:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:51.639 11:38:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:51.639 11:38:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:51.639 11:38:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:51.914 11:38:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:51.914 "name": "raid_bdev1", 00:27:51.914 "uuid": "08e31d09-ef91-4656-bebd-14a9474273f4", 00:27:51.914 "strip_size_kb": 0, 00:27:51.914 "state": "online", 00:27:51.914 "raid_level": "raid1", 00:27:51.914 "superblock": true, 00:27:51.914 "num_base_bdevs": 2, 00:27:51.915 "num_base_bdevs_discovered": 1, 00:27:51.915 "num_base_bdevs_operational": 1, 00:27:51.915 "base_bdevs_list": [ 00:27:51.915 { 00:27:51.915 "name": null, 00:27:51.915 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:51.915 "is_configured": false, 00:27:51.915 "data_offset": 256, 00:27:51.915 "data_size": 7936 00:27:51.915 }, 00:27:51.915 { 00:27:51.915 "name": "pt2", 00:27:51.915 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:51.915 "is_configured": true, 00:27:51.915 "data_offset": 256, 00:27:51.915 "data_size": 7936 00:27:51.915 } 00:27:51.915 ] 00:27:51.915 }' 00:27:51.915 11:38:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:51.915 11:38:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:27:52.495 11:38:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@554 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs online 00:27:52.495 11:38:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@554 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:27:52.753 11:38:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@554 -- # [[ false == \f\a\l\s\e ]] 00:27:52.753 11:38:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@557 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:27:52.753 11:38:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@557 -- # jq -r '.[] | .uuid' 00:27:53.012 [2024-07-15 11:38:36.470214] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:53.012 11:38:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@557 -- # '[' 08e31d09-ef91-4656-bebd-14a9474273f4 '!=' 08e31d09-ef91-4656-bebd-14a9474273f4 ']' 00:27:53.012 11:38:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@562 -- # killprocess 1010655 00:27:53.012 11:38:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@948 -- # '[' -z 1010655 ']' 00:27:53.012 11:38:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@952 -- # kill -0 1010655 00:27:53.012 11:38:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@953 -- # uname 00:27:53.012 11:38:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:27:53.012 11:38:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1010655 00:27:53.012 11:38:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:27:53.012 11:38:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:27:53.012 11:38:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1010655' 00:27:53.012 killing process with pid 1010655 00:27:53.012 11:38:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@967 -- # kill 1010655 00:27:53.012 [2024-07-15 11:38:36.539192] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:27:53.012 [2024-07-15 11:38:36.539245] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:53.012 [2024-07-15 11:38:36.539290] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:53.012 [2024-07-15 11:38:36.539302] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x184f850 name raid_bdev1, state offline 00:27:53.012 11:38:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@972 -- # wait 1010655 00:27:53.012 [2024-07-15 11:38:36.565732] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:27:53.270 11:38:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@564 -- # return 0 00:27:53.270 00:27:53.270 real 0m15.750s 00:27:53.270 user 0m28.489s 00:27:53.270 sys 0m2.936s 00:27:53.270 11:38:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1124 -- # xtrace_disable 00:27:53.270 11:38:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:27:53.270 ************************************ 00:27:53.270 END TEST raid_superblock_test_md_separate 00:27:53.270 ************************************ 00:27:53.270 11:38:36 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:27:53.270 11:38:36 bdev_raid -- bdev/bdev_raid.sh@907 -- # '[' true = true ']' 00:27:53.270 11:38:36 bdev_raid -- bdev/bdev_raid.sh@908 -- # run_test raid_rebuild_test_sb_md_separate raid_rebuild_test raid1 2 true false true 00:27:53.270 11:38:36 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:27:53.270 11:38:36 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:27:53.270 11:38:36 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:27:53.270 ************************************ 00:27:53.270 START TEST raid_rebuild_test_sb_md_separate 00:27:53.270 ************************************ 00:27:53.270 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1123 -- # raid_rebuild_test raid1 2 true false true 00:27:53.528 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@568 -- # local raid_level=raid1 00:27:53.528 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@569 -- # local num_base_bdevs=2 00:27:53.528 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@570 -- # local superblock=true 00:27:53.528 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@571 -- # local background_io=false 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@572 -- # local verify=true 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@573 -- # (( i = 1 )) 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@575 -- # echo BaseBdev1 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@575 -- # echo BaseBdev2 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@573 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@573 -- # local base_bdevs 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # local raid_bdev_name=raid_bdev1 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@575 -- # local strip_size 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # local create_arg 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@577 -- # local raid_bdev_size 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@578 -- # local data_offset 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@580 -- # '[' raid1 '!=' raid1 ']' 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@588 -- # strip_size=0 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@591 -- # '[' true = true ']' 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@592 -- # create_arg+=' -s' 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@596 -- # raid_pid=1012969 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@597 -- # waitforlisten 1012969 /var/tmp/spdk-raid.sock 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@595 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@829 -- # '[' -z 1012969 ']' 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@834 -- # local max_retries=100 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:27:53.529 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@838 -- # xtrace_disable 00:27:53.529 11:38:36 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:27:53.529 [2024-07-15 11:38:36.933890] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:27:53.529 [2024-07-15 11:38:36.933962] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1012969 ] 00:27:53.529 I/O size of 3145728 is greater than zero copy threshold (65536). 00:27:53.529 Zero copy mechanism will not be used. 00:27:53.529 [2024-07-15 11:38:37.065104] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:53.786 [2024-07-15 11:38:37.169359] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:27:53.787 [2024-07-15 11:38:37.224443] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:53.787 [2024-07-15 11:38:37.224476] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:54.352 11:38:37 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:27:54.352 11:38:37 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@862 -- # return 0 00:27:54.352 11:38:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:27:54.352 11:38:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -m 32 -b BaseBdev1_malloc 00:27:54.610 BaseBdev1_malloc 00:27:54.610 11:38:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:27:54.868 [2024-07-15 11:38:38.339209] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:27:54.868 [2024-07-15 11:38:38.339259] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:54.868 [2024-07-15 11:38:38.339285] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x12e06d0 00:27:54.868 [2024-07-15 11:38:38.339298] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:54.868 [2024-07-15 11:38:38.340842] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:54.868 [2024-07-15 11:38:38.340869] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:27:54.868 BaseBdev1 00:27:54.868 11:38:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:27:54.868 11:38:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -m 32 -b BaseBdev2_malloc 00:27:55.126 BaseBdev2_malloc 00:27:55.126 11:38:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:27:55.384 [2024-07-15 11:38:38.835284] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:27:55.384 [2024-07-15 11:38:38.835329] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:55.384 [2024-07-15 11:38:38.835352] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x14381f0 00:27:55.384 [2024-07-15 11:38:38.835364] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:55.384 [2024-07-15 11:38:38.836811] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:55.384 [2024-07-15 11:38:38.836839] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:27:55.384 BaseBdev2 00:27:55.384 11:38:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@606 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -m 32 -b spare_malloc 00:27:55.642 spare_malloc 00:27:55.642 11:38:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@607 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:27:55.900 spare_delay 00:27:55.900 11:38:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@608 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:27:56.158 [2024-07-15 11:38:39.563890] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:27:56.158 [2024-07-15 11:38:39.563939] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:56.158 [2024-07-15 11:38:39.563964] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x14347a0 00:27:56.158 [2024-07-15 11:38:39.563977] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:56.158 [2024-07-15 11:38:39.565358] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:56.158 [2024-07-15 11:38:39.565385] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:27:56.158 spare 00:27:56.158 11:38:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@611 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2' -n raid_bdev1 00:27:56.417 [2024-07-15 11:38:39.796528] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:56.417 [2024-07-15 11:38:39.797814] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:56.417 [2024-07-15 11:38:39.797988] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x14351c0 00:27:56.417 [2024-07-15 11:38:39.798002] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:27:56.417 [2024-07-15 11:38:39.798075] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1346360 00:27:56.417 [2024-07-15 11:38:39.798187] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x14351c0 00:27:56.417 [2024-07-15 11:38:39.798197] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x14351c0 00:27:56.417 [2024-07-15 11:38:39.798267] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:56.417 11:38:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@612 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:27:56.417 11:38:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:27:56.417 11:38:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:27:56.417 11:38:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:56.417 11:38:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:56.417 11:38:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:27:56.417 11:38:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:56.417 11:38:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:56.417 11:38:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:56.417 11:38:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:56.417 11:38:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:56.417 11:38:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:56.675 11:38:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:56.675 "name": "raid_bdev1", 00:27:56.675 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:27:56.675 "strip_size_kb": 0, 00:27:56.675 "state": "online", 00:27:56.675 "raid_level": "raid1", 00:27:56.675 "superblock": true, 00:27:56.675 "num_base_bdevs": 2, 00:27:56.675 "num_base_bdevs_discovered": 2, 00:27:56.675 "num_base_bdevs_operational": 2, 00:27:56.675 "base_bdevs_list": [ 00:27:56.675 { 00:27:56.675 "name": "BaseBdev1", 00:27:56.675 "uuid": "17b520e3-3215-5905-a1ab-88f488e188c5", 00:27:56.675 "is_configured": true, 00:27:56.675 "data_offset": 256, 00:27:56.675 "data_size": 7936 00:27:56.675 }, 00:27:56.675 { 00:27:56.675 "name": "BaseBdev2", 00:27:56.675 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:27:56.675 "is_configured": true, 00:27:56.675 "data_offset": 256, 00:27:56.675 "data_size": 7936 00:27:56.675 } 00:27:56.675 ] 00:27:56.675 }' 00:27:56.675 11:38:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:56.675 11:38:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:27:57.240 11:38:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@615 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:27:57.240 11:38:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@615 -- # jq -r '.[].num_blocks' 00:27:57.498 [2024-07-15 11:38:40.875609] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:57.498 11:38:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@615 -- # raid_bdev_size=7936 00:27:57.498 11:38:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@618 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:57.498 11:38:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@618 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:27:57.756 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@618 -- # data_offset=256 00:27:57.756 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@620 -- # '[' false = true ']' 00:27:57.756 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@623 -- # '[' true = true ']' 00:27:57.756 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@624 -- # local write_unit_size 00:27:57.756 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@627 -- # nbd_start_disks /var/tmp/spdk-raid.sock raid_bdev1 /dev/nbd0 00:27:57.756 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:27:57.756 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:27:57.756 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:27:57.756 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:27:57.756 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:27:57.756 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:27:57.756 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:27:57.756 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:27:57.756 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:27:58.013 [2024-07-15 11:38:41.360683] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1346360 00:27:58.013 /dev/nbd0 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@867 -- # local i 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # break 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:27:58.013 1+0 records in 00:27:58.013 1+0 records out 00:27:58.013 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000164214 s, 24.9 MB/s 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # size=4096 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@887 -- # return 0 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@628 -- # '[' raid1 = raid5f ']' 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@632 -- # write_unit_size=1 00:27:58.013 11:38:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@634 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:27:58.579 7936+0 records in 00:27:58.579 7936+0 records out 00:27:58.579 32505856 bytes (33 MB, 31 MiB) copied, 0.741531 s, 43.8 MB/s 00:27:58.579 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@635 -- # nbd_stop_disks /var/tmp/spdk-raid.sock /dev/nbd0 00:27:58.579 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:27:58.579 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:27:58.579 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:27:58.579 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:27:58.579 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:27:58.579 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd0 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:27:59.145 [2024-07-15 11:38:42.436135] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@639 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev1 00:27:59.145 [2024-07-15 11:38:42.672800] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@642 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:27:59.145 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:59.403 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:27:59.403 "name": "raid_bdev1", 00:27:59.403 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:27:59.403 "strip_size_kb": 0, 00:27:59.403 "state": "online", 00:27:59.403 "raid_level": "raid1", 00:27:59.403 "superblock": true, 00:27:59.403 "num_base_bdevs": 2, 00:27:59.403 "num_base_bdevs_discovered": 1, 00:27:59.403 "num_base_bdevs_operational": 1, 00:27:59.403 "base_bdevs_list": [ 00:27:59.403 { 00:27:59.403 "name": null, 00:27:59.403 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:59.403 "is_configured": false, 00:27:59.404 "data_offset": 256, 00:27:59.404 "data_size": 7936 00:27:59.404 }, 00:27:59.404 { 00:27:59.404 "name": "BaseBdev2", 00:27:59.404 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:27:59.404 "is_configured": true, 00:27:59.404 "data_offset": 256, 00:27:59.404 "data_size": 7936 00:27:59.404 } 00:27:59.404 ] 00:27:59.404 }' 00:27:59.404 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:27:59.404 11:38:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:27:59.969 11:38:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@645 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:28:00.228 [2024-07-15 11:38:43.763687] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:28:00.228 [2024-07-15 11:38:43.765998] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x12df350 00:28:00.228 [2024-07-15 11:38:43.768290] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:28:00.228 11:38:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@646 -- # sleep 1 00:28:01.601 11:38:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@649 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:01.601 11:38:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:28:01.601 11:38:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:28:01.601 11:38:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local target=spare 00:28:01.601 11:38:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:28:01.601 11:38:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:01.601 11:38:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:01.601 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:01.601 "name": "raid_bdev1", 00:28:01.601 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:01.601 "strip_size_kb": 0, 00:28:01.601 "state": "online", 00:28:01.601 "raid_level": "raid1", 00:28:01.601 "superblock": true, 00:28:01.601 "num_base_bdevs": 2, 00:28:01.601 "num_base_bdevs_discovered": 2, 00:28:01.601 "num_base_bdevs_operational": 2, 00:28:01.601 "process": { 00:28:01.601 "type": "rebuild", 00:28:01.601 "target": "spare", 00:28:01.601 "progress": { 00:28:01.601 "blocks": 3072, 00:28:01.601 "percent": 38 00:28:01.601 } 00:28:01.601 }, 00:28:01.601 "base_bdevs_list": [ 00:28:01.601 { 00:28:01.601 "name": "spare", 00:28:01.601 "uuid": "ba5cd540-c78b-5828-9bb2-cc10493afd25", 00:28:01.601 "is_configured": true, 00:28:01.601 "data_offset": 256, 00:28:01.601 "data_size": 7936 00:28:01.601 }, 00:28:01.601 { 00:28:01.601 "name": "BaseBdev2", 00:28:01.601 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:01.601 "is_configured": true, 00:28:01.601 "data_offset": 256, 00:28:01.601 "data_size": 7936 00:28:01.601 } 00:28:01.601 ] 00:28:01.601 }' 00:28:01.601 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:28:01.601 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:01.601 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:28:01.601 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:28:01.601 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@652 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev spare 00:28:01.859 [2024-07-15 11:38:45.341328] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:01.859 [2024-07-15 11:38:45.380842] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:28:01.859 [2024-07-15 11:38:45.380889] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:01.859 [2024-07-15 11:38:45.380904] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:01.859 [2024-07-15 11:38:45.380913] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:28:01.859 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@655 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:28:01.859 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:28:01.859 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:28:01.859 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:01.859 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:01.859 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:28:01.859 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:01.859 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:01.859 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:01.859 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:01.859 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:01.859 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:02.117 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:02.117 "name": "raid_bdev1", 00:28:02.117 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:02.117 "strip_size_kb": 0, 00:28:02.117 "state": "online", 00:28:02.117 "raid_level": "raid1", 00:28:02.117 "superblock": true, 00:28:02.117 "num_base_bdevs": 2, 00:28:02.117 "num_base_bdevs_discovered": 1, 00:28:02.117 "num_base_bdevs_operational": 1, 00:28:02.117 "base_bdevs_list": [ 00:28:02.117 { 00:28:02.117 "name": null, 00:28:02.117 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:02.117 "is_configured": false, 00:28:02.117 "data_offset": 256, 00:28:02.117 "data_size": 7936 00:28:02.117 }, 00:28:02.117 { 00:28:02.117 "name": "BaseBdev2", 00:28:02.117 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:02.117 "is_configured": true, 00:28:02.117 "data_offset": 256, 00:28:02.117 "data_size": 7936 00:28:02.117 } 00:28:02.117 ] 00:28:02.117 }' 00:28:02.117 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:02.117 11:38:45 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:28:02.682 11:38:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@658 -- # verify_raid_bdev_process raid_bdev1 none none 00:28:02.682 11:38:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:28:02.682 11:38:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:28:02.682 11:38:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local target=none 00:28:02.682 11:38:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:28:02.682 11:38:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:02.682 11:38:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:02.939 11:38:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:02.939 "name": "raid_bdev1", 00:28:02.939 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:02.939 "strip_size_kb": 0, 00:28:02.939 "state": "online", 00:28:02.939 "raid_level": "raid1", 00:28:02.939 "superblock": true, 00:28:02.939 "num_base_bdevs": 2, 00:28:02.939 "num_base_bdevs_discovered": 1, 00:28:02.939 "num_base_bdevs_operational": 1, 00:28:02.939 "base_bdevs_list": [ 00:28:02.939 { 00:28:02.939 "name": null, 00:28:02.939 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:02.939 "is_configured": false, 00:28:02.939 "data_offset": 256, 00:28:02.939 "data_size": 7936 00:28:02.940 }, 00:28:02.940 { 00:28:02.940 "name": "BaseBdev2", 00:28:02.940 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:02.940 "is_configured": true, 00:28:02.940 "data_offset": 256, 00:28:02.940 "data_size": 7936 00:28:02.940 } 00:28:02.940 ] 00:28:02.940 }' 00:28:02.940 11:38:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:28:02.940 11:38:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:28:02.940 11:38:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:28:02.940 11:38:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:28:02.940 11:38:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@661 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:28:03.199 [2024-07-15 11:38:46.716035] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:28:03.199 [2024-07-15 11:38:46.718317] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x12e0280 00:28:03.199 [2024-07-15 11:38:46.719848] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:28:03.199 11:38:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@662 -- # sleep 1 00:28:04.576 11:38:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@663 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:04.576 11:38:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:28:04.576 11:38:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:28:04.576 11:38:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local target=spare 00:28:04.576 11:38:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:28:04.576 11:38:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:04.576 11:38:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:04.576 11:38:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:04.576 "name": "raid_bdev1", 00:28:04.576 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:04.576 "strip_size_kb": 0, 00:28:04.576 "state": "online", 00:28:04.576 "raid_level": "raid1", 00:28:04.576 "superblock": true, 00:28:04.576 "num_base_bdevs": 2, 00:28:04.576 "num_base_bdevs_discovered": 2, 00:28:04.576 "num_base_bdevs_operational": 2, 00:28:04.576 "process": { 00:28:04.576 "type": "rebuild", 00:28:04.576 "target": "spare", 00:28:04.576 "progress": { 00:28:04.576 "blocks": 3072, 00:28:04.576 "percent": 38 00:28:04.576 } 00:28:04.576 }, 00:28:04.576 "base_bdevs_list": [ 00:28:04.576 { 00:28:04.576 "name": "spare", 00:28:04.576 "uuid": "ba5cd540-c78b-5828-9bb2-cc10493afd25", 00:28:04.576 "is_configured": true, 00:28:04.576 "data_offset": 256, 00:28:04.576 "data_size": 7936 00:28:04.576 }, 00:28:04.576 { 00:28:04.576 "name": "BaseBdev2", 00:28:04.576 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:04.576 "is_configured": true, 00:28:04.576 "data_offset": 256, 00:28:04.576 "data_size": 7936 00:28:04.576 } 00:28:04.576 ] 00:28:04.576 }' 00:28:04.576 11:38:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:28:04.576 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:04.576 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:28:04.576 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:28:04.576 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@665 -- # '[' true = true ']' 00:28:04.576 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@665 -- # '[' = false ']' 00:28:04.576 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev_raid.sh: line 665: [: =: unary operator expected 00:28:04.576 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@690 -- # local num_base_bdevs_operational=2 00:28:04.576 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@692 -- # '[' raid1 = raid1 ']' 00:28:04.576 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@692 -- # '[' 2 -gt 2 ']' 00:28:04.576 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@705 -- # local timeout=1056 00:28:04.576 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:28:04.576 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:04.576 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:28:04.576 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:28:04.576 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local target=spare 00:28:04.576 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:28:04.576 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:04.576 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:04.835 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:04.835 "name": "raid_bdev1", 00:28:04.835 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:04.835 "strip_size_kb": 0, 00:28:04.835 "state": "online", 00:28:04.835 "raid_level": "raid1", 00:28:04.835 "superblock": true, 00:28:04.835 "num_base_bdevs": 2, 00:28:04.835 "num_base_bdevs_discovered": 2, 00:28:04.835 "num_base_bdevs_operational": 2, 00:28:04.835 "process": { 00:28:04.835 "type": "rebuild", 00:28:04.835 "target": "spare", 00:28:04.835 "progress": { 00:28:04.835 "blocks": 3840, 00:28:04.835 "percent": 48 00:28:04.835 } 00:28:04.835 }, 00:28:04.835 "base_bdevs_list": [ 00:28:04.835 { 00:28:04.835 "name": "spare", 00:28:04.835 "uuid": "ba5cd540-c78b-5828-9bb2-cc10493afd25", 00:28:04.835 "is_configured": true, 00:28:04.835 "data_offset": 256, 00:28:04.835 "data_size": 7936 00:28:04.835 }, 00:28:04.835 { 00:28:04.835 "name": "BaseBdev2", 00:28:04.835 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:04.835 "is_configured": true, 00:28:04.835 "data_offset": 256, 00:28:04.835 "data_size": 7936 00:28:04.835 } 00:28:04.835 ] 00:28:04.835 }' 00:28:04.835 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:28:04.835 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:04.835 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:28:04.835 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:28:04.835 11:38:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@710 -- # sleep 1 00:28:06.208 11:38:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:28:06.208 11:38:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:06.208 11:38:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:28:06.208 11:38:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:28:06.208 11:38:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local target=spare 00:28:06.208 11:38:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:28:06.208 11:38:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:06.208 11:38:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:06.208 11:38:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:06.208 "name": "raid_bdev1", 00:28:06.208 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:06.208 "strip_size_kb": 0, 00:28:06.208 "state": "online", 00:28:06.208 "raid_level": "raid1", 00:28:06.208 "superblock": true, 00:28:06.208 "num_base_bdevs": 2, 00:28:06.208 "num_base_bdevs_discovered": 2, 00:28:06.208 "num_base_bdevs_operational": 2, 00:28:06.208 "process": { 00:28:06.208 "type": "rebuild", 00:28:06.208 "target": "spare", 00:28:06.208 "progress": { 00:28:06.208 "blocks": 7424, 00:28:06.208 "percent": 93 00:28:06.208 } 00:28:06.208 }, 00:28:06.208 "base_bdevs_list": [ 00:28:06.208 { 00:28:06.208 "name": "spare", 00:28:06.208 "uuid": "ba5cd540-c78b-5828-9bb2-cc10493afd25", 00:28:06.208 "is_configured": true, 00:28:06.208 "data_offset": 256, 00:28:06.208 "data_size": 7936 00:28:06.208 }, 00:28:06.208 { 00:28:06.208 "name": "BaseBdev2", 00:28:06.208 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:06.208 "is_configured": true, 00:28:06.208 "data_offset": 256, 00:28:06.208 "data_size": 7936 00:28:06.208 } 00:28:06.208 ] 00:28:06.208 }' 00:28:06.208 11:38:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:28:06.208 11:38:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:06.208 11:38:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:28:06.208 11:38:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:28:06.208 11:38:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@710 -- # sleep 1 00:28:06.466 [2024-07-15 11:38:49.844477] bdev_raid.c:2789:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:28:06.466 [2024-07-15 11:38:49.844534] bdev_raid.c:2504:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:28:06.466 [2024-07-15 11:38:49.844613] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:07.426 11:38:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:28:07.426 11:38:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:07.426 11:38:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:28:07.426 11:38:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:28:07.426 11:38:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local target=spare 00:28:07.426 11:38:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:28:07.426 11:38:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:07.426 11:38:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:07.426 11:38:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:07.426 "name": "raid_bdev1", 00:28:07.426 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:07.426 "strip_size_kb": 0, 00:28:07.426 "state": "online", 00:28:07.426 "raid_level": "raid1", 00:28:07.426 "superblock": true, 00:28:07.426 "num_base_bdevs": 2, 00:28:07.426 "num_base_bdevs_discovered": 2, 00:28:07.426 "num_base_bdevs_operational": 2, 00:28:07.426 "base_bdevs_list": [ 00:28:07.426 { 00:28:07.426 "name": "spare", 00:28:07.426 "uuid": "ba5cd540-c78b-5828-9bb2-cc10493afd25", 00:28:07.426 "is_configured": true, 00:28:07.426 "data_offset": 256, 00:28:07.426 "data_size": 7936 00:28:07.426 }, 00:28:07.426 { 00:28:07.426 "name": "BaseBdev2", 00:28:07.426 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:07.426 "is_configured": true, 00:28:07.426 "data_offset": 256, 00:28:07.426 "data_size": 7936 00:28:07.426 } 00:28:07.426 ] 00:28:07.426 }' 00:28:07.426 11:38:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:28:07.684 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # [[ none == \r\e\b\u\i\l\d ]] 00:28:07.684 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:28:07.684 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # [[ none == \s\p\a\r\e ]] 00:28:07.684 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # break 00:28:07.684 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@714 -- # verify_raid_bdev_process raid_bdev1 none none 00:28:07.684 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:28:07.684 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:28:07.684 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local target=none 00:28:07.684 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:28:07.684 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:07.684 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:07.942 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:07.942 "name": "raid_bdev1", 00:28:07.942 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:07.942 "strip_size_kb": 0, 00:28:07.942 "state": "online", 00:28:07.942 "raid_level": "raid1", 00:28:07.942 "superblock": true, 00:28:07.942 "num_base_bdevs": 2, 00:28:07.942 "num_base_bdevs_discovered": 2, 00:28:07.942 "num_base_bdevs_operational": 2, 00:28:07.942 "base_bdevs_list": [ 00:28:07.942 { 00:28:07.942 "name": "spare", 00:28:07.942 "uuid": "ba5cd540-c78b-5828-9bb2-cc10493afd25", 00:28:07.942 "is_configured": true, 00:28:07.942 "data_offset": 256, 00:28:07.942 "data_size": 7936 00:28:07.942 }, 00:28:07.942 { 00:28:07.942 "name": "BaseBdev2", 00:28:07.942 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:07.942 "is_configured": true, 00:28:07.942 "data_offset": 256, 00:28:07.942 "data_size": 7936 00:28:07.942 } 00:28:07.942 ] 00:28:07.942 }' 00:28:07.942 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:28:07.942 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:28:07.942 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:28:07.942 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:28:07.942 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:28:07.942 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:28:07.942 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:28:07.942 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:07.942 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:07.942 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:28:07.942 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:07.942 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:07.942 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:07.942 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:07.942 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:07.942 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:08.201 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:08.201 "name": "raid_bdev1", 00:28:08.201 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:08.201 "strip_size_kb": 0, 00:28:08.201 "state": "online", 00:28:08.201 "raid_level": "raid1", 00:28:08.201 "superblock": true, 00:28:08.201 "num_base_bdevs": 2, 00:28:08.201 "num_base_bdevs_discovered": 2, 00:28:08.201 "num_base_bdevs_operational": 2, 00:28:08.201 "base_bdevs_list": [ 00:28:08.201 { 00:28:08.201 "name": "spare", 00:28:08.201 "uuid": "ba5cd540-c78b-5828-9bb2-cc10493afd25", 00:28:08.201 "is_configured": true, 00:28:08.201 "data_offset": 256, 00:28:08.201 "data_size": 7936 00:28:08.201 }, 00:28:08.201 { 00:28:08.201 "name": "BaseBdev2", 00:28:08.201 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:08.201 "is_configured": true, 00:28:08.201 "data_offset": 256, 00:28:08.201 "data_size": 7936 00:28:08.201 } 00:28:08.201 ] 00:28:08.201 }' 00:28:08.201 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:08.201 11:38:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:28:08.769 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@718 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:28:09.028 [2024-07-15 11:38:52.455247] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:28:09.028 [2024-07-15 11:38:52.455275] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:09.028 [2024-07-15 11:38:52.455331] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:09.028 [2024-07-15 11:38:52.455386] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:09.028 [2024-07-15 11:38:52.455398] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x14351c0 name raid_bdev1, state offline 00:28:09.028 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@719 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:09.028 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@719 -- # jq length 00:28:09.287 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@719 -- # [[ 0 == 0 ]] 00:28:09.287 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@721 -- # '[' true = true ']' 00:28:09.287 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@722 -- # '[' false = true ']' 00:28:09.287 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@736 -- # nbd_start_disks /var/tmp/spdk-raid.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:28:09.287 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:28:09.287 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:28:09.287 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:28:09.287 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:28:09.287 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:28:09.287 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:28:09.287 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:28:09.287 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:28:09.287 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:28:09.547 /dev/nbd0 00:28:09.547 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:28:09.547 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:28:09.547 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:28:09.547 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@867 -- # local i 00:28:09.547 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:28:09.547 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:28:09.547 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:28:09.547 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # break 00:28:09.547 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:28:09.547 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:28:09.547 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:28:09.547 1+0 records in 00:28:09.547 1+0 records out 00:28:09.547 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000241507 s, 17.0 MB/s 00:28:09.547 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:28:09.547 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # size=4096 00:28:09.547 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:28:09.547 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:28:09.547 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@887 -- # return 0 00:28:09.547 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:28:09.547 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:28:09.547 11:38:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_start_disk spare /dev/nbd1 00:28:09.820 /dev/nbd1 00:28:09.820 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:28:09.820 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:28:09.820 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:28:09.820 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@867 -- # local i 00:28:09.820 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:28:09.820 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:28:09.820 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:28:09.820 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # break 00:28:09.820 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:28:09.821 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:28:09.821 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:28:09.821 1+0 records in 00:28:09.821 1+0 records out 00:28:09.821 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000251413 s, 16.3 MB/s 00:28:09.821 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:28:09.821 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # size=4096 00:28:09.821 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:28:09.821 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:28:09.821 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@887 -- # return 0 00:28:09.821 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:28:09.821 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:28:09.821 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@737 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:28:09.821 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@738 -- # nbd_stop_disks /var/tmp/spdk-raid.sock '/dev/nbd0 /dev/nbd1' 00:28:09.821 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-raid.sock 00:28:09.821 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:28:09.821 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:28:09.821 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:28:09.821 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:28:09.821 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd0 00:28:10.083 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:28:10.083 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:28:10.083 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:28:10.083 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:28:10.083 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:28:10.083 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:28:10.083 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:28:10.083 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:28:10.083 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:28:10.083 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock nbd_stop_disk /dev/nbd1 00:28:10.342 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:28:10.342 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:28:10.342 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:28:10.342 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:28:10.342 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:28:10.342 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:28:10.342 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:28:10.342 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:28:10.342 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@742 -- # '[' true = true ']' 00:28:10.342 11:38:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@744 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:28:10.600 11:38:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@745 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:28:10.859 [2024-07-15 11:38:54.391072] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:28:10.859 [2024-07-15 11:38:54.391116] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:10.859 [2024-07-15 11:38:54.391138] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1434f60 00:28:10.859 [2024-07-15 11:38:54.391151] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:10.859 [2024-07-15 11:38:54.392609] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:10.859 [2024-07-15 11:38:54.392636] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:28:10.859 [2024-07-15 11:38:54.392691] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:28:10.859 [2024-07-15 11:38:54.392723] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:28:10.859 [2024-07-15 11:38:54.392817] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:10.859 spare 00:28:10.859 11:38:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@747 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:28:10.859 11:38:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:28:10.859 11:38:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:28:10.859 11:38:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:10.859 11:38:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:10.859 11:38:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:28:10.859 11:38:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:10.859 11:38:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:10.859 11:38:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:10.859 11:38:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:10.859 11:38:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:10.859 11:38:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:11.118 [2024-07-15 11:38:54.493123] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x13467c0 00:28:11.118 [2024-07-15 11:38:54.493141] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:28:11.118 [2024-07-15 11:38:54.493209] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1346480 00:28:11.118 [2024-07-15 11:38:54.493329] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x13467c0 00:28:11.118 [2024-07-15 11:38:54.493339] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x13467c0 00:28:11.118 [2024-07-15 11:38:54.493414] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:11.118 11:38:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:11.118 "name": "raid_bdev1", 00:28:11.118 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:11.118 "strip_size_kb": 0, 00:28:11.118 "state": "online", 00:28:11.118 "raid_level": "raid1", 00:28:11.118 "superblock": true, 00:28:11.118 "num_base_bdevs": 2, 00:28:11.118 "num_base_bdevs_discovered": 2, 00:28:11.118 "num_base_bdevs_operational": 2, 00:28:11.118 "base_bdevs_list": [ 00:28:11.118 { 00:28:11.118 "name": "spare", 00:28:11.118 "uuid": "ba5cd540-c78b-5828-9bb2-cc10493afd25", 00:28:11.118 "is_configured": true, 00:28:11.118 "data_offset": 256, 00:28:11.118 "data_size": 7936 00:28:11.118 }, 00:28:11.118 { 00:28:11.118 "name": "BaseBdev2", 00:28:11.118 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:11.118 "is_configured": true, 00:28:11.118 "data_offset": 256, 00:28:11.118 "data_size": 7936 00:28:11.118 } 00:28:11.118 ] 00:28:11.118 }' 00:28:11.118 11:38:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:11.118 11:38:54 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:28:11.685 11:38:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@748 -- # verify_raid_bdev_process raid_bdev1 none none 00:28:11.685 11:38:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:28:11.685 11:38:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:28:11.685 11:38:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local target=none 00:28:11.685 11:38:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:28:11.685 11:38:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:11.685 11:38:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:11.965 11:38:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:11.965 "name": "raid_bdev1", 00:28:11.965 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:11.965 "strip_size_kb": 0, 00:28:11.965 "state": "online", 00:28:11.965 "raid_level": "raid1", 00:28:11.965 "superblock": true, 00:28:11.965 "num_base_bdevs": 2, 00:28:11.965 "num_base_bdevs_discovered": 2, 00:28:11.965 "num_base_bdevs_operational": 2, 00:28:11.965 "base_bdevs_list": [ 00:28:11.965 { 00:28:11.965 "name": "spare", 00:28:11.965 "uuid": "ba5cd540-c78b-5828-9bb2-cc10493afd25", 00:28:11.965 "is_configured": true, 00:28:11.965 "data_offset": 256, 00:28:11.965 "data_size": 7936 00:28:11.965 }, 00:28:11.965 { 00:28:11.965 "name": "BaseBdev2", 00:28:11.965 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:11.965 "is_configured": true, 00:28:11.965 "data_offset": 256, 00:28:11.965 "data_size": 7936 00:28:11.965 } 00:28:11.965 ] 00:28:11.965 }' 00:28:11.965 11:38:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:28:12.222 11:38:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:28:12.222 11:38:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:28:12.222 11:38:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:28:12.222 11:38:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@749 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:12.222 11:38:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@749 -- # jq -r '.[].base_bdevs_list[0].name' 00:28:12.478 11:38:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@749 -- # [[ spare == \s\p\a\r\e ]] 00:28:12.478 11:38:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@752 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev spare 00:28:12.735 [2024-07-15 11:38:56.075642] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:12.735 11:38:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@753 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:28:12.735 11:38:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:28:12.735 11:38:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:28:12.735 11:38:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:12.735 11:38:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:12.735 11:38:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:28:12.735 11:38:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:12.735 11:38:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:12.735 11:38:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:12.735 11:38:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:12.735 11:38:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:12.736 11:38:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:12.992 11:38:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:12.992 "name": "raid_bdev1", 00:28:12.992 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:12.992 "strip_size_kb": 0, 00:28:12.992 "state": "online", 00:28:12.992 "raid_level": "raid1", 00:28:12.992 "superblock": true, 00:28:12.992 "num_base_bdevs": 2, 00:28:12.992 "num_base_bdevs_discovered": 1, 00:28:12.992 "num_base_bdevs_operational": 1, 00:28:12.992 "base_bdevs_list": [ 00:28:12.992 { 00:28:12.992 "name": null, 00:28:12.992 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:12.992 "is_configured": false, 00:28:12.992 "data_offset": 256, 00:28:12.992 "data_size": 7936 00:28:12.992 }, 00:28:12.992 { 00:28:12.992 "name": "BaseBdev2", 00:28:12.992 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:12.992 "is_configured": true, 00:28:12.992 "data_offset": 256, 00:28:12.992 "data_size": 7936 00:28:12.992 } 00:28:12.992 ] 00:28:12.992 }' 00:28:12.992 11:38:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:12.992 11:38:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:28:13.554 11:38:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@754 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:28:13.554 [2024-07-15 11:38:57.142493] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:28:13.554 [2024-07-15 11:38:57.142638] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:28:13.554 [2024-07-15 11:38:57.142655] bdev_raid.c:3620:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:28:13.554 [2024-07-15 11:38:57.142682] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:28:13.554 [2024-07-15 11:38:57.144852] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x12e0280 00:28:13.554 [2024-07-15 11:38:57.146194] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:28:13.811 11:38:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@755 -- # sleep 1 00:28:14.741 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@756 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:14.741 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:28:14.741 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:28:14.741 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local target=spare 00:28:14.741 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:28:14.741 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:14.741 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:14.998 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:14.998 "name": "raid_bdev1", 00:28:14.998 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:14.998 "strip_size_kb": 0, 00:28:14.998 "state": "online", 00:28:14.998 "raid_level": "raid1", 00:28:14.998 "superblock": true, 00:28:14.998 "num_base_bdevs": 2, 00:28:14.998 "num_base_bdevs_discovered": 2, 00:28:14.998 "num_base_bdevs_operational": 2, 00:28:14.998 "process": { 00:28:14.998 "type": "rebuild", 00:28:14.998 "target": "spare", 00:28:14.998 "progress": { 00:28:14.998 "blocks": 3072, 00:28:14.998 "percent": 38 00:28:14.998 } 00:28:14.998 }, 00:28:14.998 "base_bdevs_list": [ 00:28:14.998 { 00:28:14.998 "name": "spare", 00:28:14.998 "uuid": "ba5cd540-c78b-5828-9bb2-cc10493afd25", 00:28:14.998 "is_configured": true, 00:28:14.998 "data_offset": 256, 00:28:14.998 "data_size": 7936 00:28:14.998 }, 00:28:14.998 { 00:28:14.998 "name": "BaseBdev2", 00:28:14.998 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:14.998 "is_configured": true, 00:28:14.998 "data_offset": 256, 00:28:14.998 "data_size": 7936 00:28:14.998 } 00:28:14.998 ] 00:28:14.998 }' 00:28:14.998 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:28:14.998 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:14.998 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:28:14.998 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:28:14.998 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@759 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:28:15.256 [2024-07-15 11:38:58.727750] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:15.256 [2024-07-15 11:38:58.758645] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:28:15.256 [2024-07-15 11:38:58.758692] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:15.256 [2024-07-15 11:38:58.758707] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:15.256 [2024-07-15 11:38:58.758722] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:28:15.256 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@760 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:28:15.256 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:28:15.256 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:28:15.256 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:15.256 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:15.256 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:28:15.256 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:15.256 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:15.256 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:15.256 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:15.256 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:15.256 11:38:58 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:15.513 11:38:59 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:15.514 "name": "raid_bdev1", 00:28:15.514 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:15.514 "strip_size_kb": 0, 00:28:15.514 "state": "online", 00:28:15.514 "raid_level": "raid1", 00:28:15.514 "superblock": true, 00:28:15.514 "num_base_bdevs": 2, 00:28:15.514 "num_base_bdevs_discovered": 1, 00:28:15.514 "num_base_bdevs_operational": 1, 00:28:15.514 "base_bdevs_list": [ 00:28:15.514 { 00:28:15.514 "name": null, 00:28:15.514 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:15.514 "is_configured": false, 00:28:15.514 "data_offset": 256, 00:28:15.514 "data_size": 7936 00:28:15.514 }, 00:28:15.514 { 00:28:15.514 "name": "BaseBdev2", 00:28:15.514 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:15.514 "is_configured": true, 00:28:15.514 "data_offset": 256, 00:28:15.514 "data_size": 7936 00:28:15.514 } 00:28:15.514 ] 00:28:15.514 }' 00:28:15.514 11:38:59 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:15.514 11:38:59 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:28:16.079 11:38:59 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@761 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:28:16.337 [2024-07-15 11:38:59.817092] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:28:16.337 [2024-07-15 11:38:59.817139] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:16.337 [2024-07-15 11:38:59.817164] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1345ca0 00:28:16.337 [2024-07-15 11:38:59.817177] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:16.337 [2024-07-15 11:38:59.817389] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:16.337 [2024-07-15 11:38:59.817405] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:28:16.337 [2024-07-15 11:38:59.817462] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:28:16.337 [2024-07-15 11:38:59.817474] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:28:16.337 [2024-07-15 11:38:59.817485] bdev_raid.c:3620:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:28:16.337 [2024-07-15 11:38:59.817503] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:28:16.337 [2024-07-15 11:38:59.819685] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x12e0280 00:28:16.337 [2024-07-15 11:38:59.821017] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:28:16.337 spare 00:28:16.337 11:38:59 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@762 -- # sleep 1 00:28:17.270 11:39:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@763 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:17.270 11:39:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:28:17.270 11:39:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:28:17.270 11:39:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local target=spare 00:28:17.270 11:39:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:28:17.270 11:39:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:17.270 11:39:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:17.528 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:17.528 "name": "raid_bdev1", 00:28:17.528 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:17.528 "strip_size_kb": 0, 00:28:17.528 "state": "online", 00:28:17.528 "raid_level": "raid1", 00:28:17.528 "superblock": true, 00:28:17.528 "num_base_bdevs": 2, 00:28:17.528 "num_base_bdevs_discovered": 2, 00:28:17.528 "num_base_bdevs_operational": 2, 00:28:17.528 "process": { 00:28:17.528 "type": "rebuild", 00:28:17.528 "target": "spare", 00:28:17.528 "progress": { 00:28:17.528 "blocks": 3072, 00:28:17.528 "percent": 38 00:28:17.528 } 00:28:17.528 }, 00:28:17.528 "base_bdevs_list": [ 00:28:17.528 { 00:28:17.528 "name": "spare", 00:28:17.528 "uuid": "ba5cd540-c78b-5828-9bb2-cc10493afd25", 00:28:17.528 "is_configured": true, 00:28:17.528 "data_offset": 256, 00:28:17.528 "data_size": 7936 00:28:17.528 }, 00:28:17.528 { 00:28:17.528 "name": "BaseBdev2", 00:28:17.528 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:17.528 "is_configured": true, 00:28:17.528 "data_offset": 256, 00:28:17.528 "data_size": 7936 00:28:17.528 } 00:28:17.528 ] 00:28:17.528 }' 00:28:17.529 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:28:17.786 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:17.787 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:28:17.787 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:28:17.787 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@766 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:28:18.045 [2024-07-15 11:39:01.421958] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:18.045 [2024-07-15 11:39:01.433742] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:28:18.045 [2024-07-15 11:39:01.433787] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:18.045 [2024-07-15 11:39:01.433803] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:18.045 [2024-07-15 11:39:01.433811] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:28:18.045 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@767 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:28:18.045 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:28:18.045 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:28:18.045 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:18.045 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:18.045 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:28:18.045 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:18.045 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:18.045 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:18.045 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:18.045 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:18.045 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:18.303 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:18.303 "name": "raid_bdev1", 00:28:18.303 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:18.303 "strip_size_kb": 0, 00:28:18.303 "state": "online", 00:28:18.303 "raid_level": "raid1", 00:28:18.303 "superblock": true, 00:28:18.303 "num_base_bdevs": 2, 00:28:18.303 "num_base_bdevs_discovered": 1, 00:28:18.303 "num_base_bdevs_operational": 1, 00:28:18.303 "base_bdevs_list": [ 00:28:18.303 { 00:28:18.303 "name": null, 00:28:18.303 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:18.303 "is_configured": false, 00:28:18.303 "data_offset": 256, 00:28:18.303 "data_size": 7936 00:28:18.303 }, 00:28:18.303 { 00:28:18.303 "name": "BaseBdev2", 00:28:18.303 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:18.303 "is_configured": true, 00:28:18.303 "data_offset": 256, 00:28:18.303 "data_size": 7936 00:28:18.303 } 00:28:18.303 ] 00:28:18.303 }' 00:28:18.303 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:18.303 11:39:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:28:18.868 11:39:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@768 -- # verify_raid_bdev_process raid_bdev1 none none 00:28:18.868 11:39:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:28:18.868 11:39:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:28:18.868 11:39:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local target=none 00:28:18.868 11:39:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:28:18.868 11:39:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:18.868 11:39:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:19.125 11:39:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:19.125 "name": "raid_bdev1", 00:28:19.125 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:19.125 "strip_size_kb": 0, 00:28:19.125 "state": "online", 00:28:19.125 "raid_level": "raid1", 00:28:19.125 "superblock": true, 00:28:19.125 "num_base_bdevs": 2, 00:28:19.125 "num_base_bdevs_discovered": 1, 00:28:19.125 "num_base_bdevs_operational": 1, 00:28:19.125 "base_bdevs_list": [ 00:28:19.125 { 00:28:19.125 "name": null, 00:28:19.125 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:19.125 "is_configured": false, 00:28:19.125 "data_offset": 256, 00:28:19.125 "data_size": 7936 00:28:19.125 }, 00:28:19.125 { 00:28:19.125 "name": "BaseBdev2", 00:28:19.125 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:19.125 "is_configured": true, 00:28:19.125 "data_offset": 256, 00:28:19.125 "data_size": 7936 00:28:19.125 } 00:28:19.125 ] 00:28:19.125 }' 00:28:19.125 11:39:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:28:19.125 11:39:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:28:19.125 11:39:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:28:19.125 11:39:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:28:19.125 11:39:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@771 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete BaseBdev1 00:28:19.382 11:39:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@772 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:28:19.639 [2024-07-15 11:39:02.989206] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:28:19.639 [2024-07-15 11:39:02.989251] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:19.639 [2024-07-15 11:39:02.989272] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x12e0900 00:28:19.639 [2024-07-15 11:39:02.989291] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:19.639 [2024-07-15 11:39:02.989479] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:19.639 [2024-07-15 11:39:02.989495] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:28:19.639 [2024-07-15 11:39:02.989539] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:28:19.639 [2024-07-15 11:39:02.989550] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:28:19.639 [2024-07-15 11:39:02.989561] bdev_raid.c:3581:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:28:19.639 BaseBdev1 00:28:19.639 11:39:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@773 -- # sleep 1 00:28:20.570 11:39:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@774 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:28:20.570 11:39:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:28:20.570 11:39:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:28:20.570 11:39:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:20.570 11:39:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:20.570 11:39:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:28:20.570 11:39:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:20.570 11:39:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:20.571 11:39:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:20.571 11:39:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:20.571 11:39:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:20.571 11:39:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:21.137 11:39:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:21.137 "name": "raid_bdev1", 00:28:21.137 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:21.137 "strip_size_kb": 0, 00:28:21.137 "state": "online", 00:28:21.137 "raid_level": "raid1", 00:28:21.137 "superblock": true, 00:28:21.137 "num_base_bdevs": 2, 00:28:21.137 "num_base_bdevs_discovered": 1, 00:28:21.137 "num_base_bdevs_operational": 1, 00:28:21.137 "base_bdevs_list": [ 00:28:21.137 { 00:28:21.137 "name": null, 00:28:21.137 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:21.137 "is_configured": false, 00:28:21.137 "data_offset": 256, 00:28:21.137 "data_size": 7936 00:28:21.137 }, 00:28:21.137 { 00:28:21.137 "name": "BaseBdev2", 00:28:21.137 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:21.137 "is_configured": true, 00:28:21.137 "data_offset": 256, 00:28:21.137 "data_size": 7936 00:28:21.137 } 00:28:21.137 ] 00:28:21.137 }' 00:28:21.137 11:39:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:21.137 11:39:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:28:21.765 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@775 -- # verify_raid_bdev_process raid_bdev1 none none 00:28:21.765 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:28:21.765 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:28:21.765 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local target=none 00:28:21.765 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:28:21.765 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:21.765 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:21.765 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:21.765 "name": "raid_bdev1", 00:28:21.765 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:21.765 "strip_size_kb": 0, 00:28:21.765 "state": "online", 00:28:21.765 "raid_level": "raid1", 00:28:21.765 "superblock": true, 00:28:21.765 "num_base_bdevs": 2, 00:28:21.765 "num_base_bdevs_discovered": 1, 00:28:21.765 "num_base_bdevs_operational": 1, 00:28:21.765 "base_bdevs_list": [ 00:28:21.765 { 00:28:21.765 "name": null, 00:28:21.765 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:21.765 "is_configured": false, 00:28:21.765 "data_offset": 256, 00:28:21.765 "data_size": 7936 00:28:21.765 }, 00:28:21.765 { 00:28:21.766 "name": "BaseBdev2", 00:28:21.766 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:21.766 "is_configured": true, 00:28:21.766 "data_offset": 256, 00:28:21.766 "data_size": 7936 00:28:21.766 } 00:28:21.766 ] 00:28:21.766 }' 00:28:21.766 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:28:22.034 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:28:22.034 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:28:22.034 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:28:22.034 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@776 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:28:22.034 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@648 -- # local es=0 00:28:22.034 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:28:22.034 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:28:22.034 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:28:22.034 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:28:22.034 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:28:22.034 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:28:22.034 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:28:22.034 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:28:22.034 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py ]] 00:28:22.034 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:28:22.291 [2024-07-15 11:39:05.636266] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:22.291 [2024-07-15 11:39:05.636388] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:28:22.292 [2024-07-15 11:39:05.636404] bdev_raid.c:3581:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:28:22.292 request: 00:28:22.292 { 00:28:22.292 "base_bdev": "BaseBdev1", 00:28:22.292 "raid_bdev": "raid_bdev1", 00:28:22.292 "method": "bdev_raid_add_base_bdev", 00:28:22.292 "req_id": 1 00:28:22.292 } 00:28:22.292 Got JSON-RPC error response 00:28:22.292 response: 00:28:22.292 { 00:28:22.292 "code": -22, 00:28:22.292 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:28:22.292 } 00:28:22.292 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@651 -- # es=1 00:28:22.292 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:28:22.292 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:28:22.292 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:28:22.292 11:39:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@777 -- # sleep 1 00:28:23.226 11:39:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@778 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:28:23.226 11:39:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:28:23.226 11:39:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:28:23.226 11:39:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:23.226 11:39:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:23.226 11:39:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:28:23.226 11:39:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:23.226 11:39:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:23.226 11:39:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:23.226 11:39:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:23.226 11:39:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:23.226 11:39:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:23.485 11:39:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:23.485 "name": "raid_bdev1", 00:28:23.485 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:23.485 "strip_size_kb": 0, 00:28:23.485 "state": "online", 00:28:23.485 "raid_level": "raid1", 00:28:23.485 "superblock": true, 00:28:23.485 "num_base_bdevs": 2, 00:28:23.485 "num_base_bdevs_discovered": 1, 00:28:23.485 "num_base_bdevs_operational": 1, 00:28:23.485 "base_bdevs_list": [ 00:28:23.485 { 00:28:23.485 "name": null, 00:28:23.485 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:23.485 "is_configured": false, 00:28:23.485 "data_offset": 256, 00:28:23.485 "data_size": 7936 00:28:23.485 }, 00:28:23.485 { 00:28:23.485 "name": "BaseBdev2", 00:28:23.485 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:23.485 "is_configured": true, 00:28:23.485 "data_offset": 256, 00:28:23.485 "data_size": 7936 00:28:23.485 } 00:28:23.485 ] 00:28:23.485 }' 00:28:23.485 11:39:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:23.485 11:39:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:28:24.052 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@779 -- # verify_raid_bdev_process raid_bdev1 none none 00:28:24.052 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:28:24.052 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:28:24.052 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local target=none 00:28:24.052 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:28:24.052 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:24.052 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:24.310 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:24.310 "name": "raid_bdev1", 00:28:24.310 "uuid": "4363cdf2-04f6-4f25-a007-39f86f8f75e9", 00:28:24.310 "strip_size_kb": 0, 00:28:24.310 "state": "online", 00:28:24.310 "raid_level": "raid1", 00:28:24.310 "superblock": true, 00:28:24.310 "num_base_bdevs": 2, 00:28:24.310 "num_base_bdevs_discovered": 1, 00:28:24.310 "num_base_bdevs_operational": 1, 00:28:24.310 "base_bdevs_list": [ 00:28:24.310 { 00:28:24.310 "name": null, 00:28:24.310 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:24.310 "is_configured": false, 00:28:24.310 "data_offset": 256, 00:28:24.310 "data_size": 7936 00:28:24.310 }, 00:28:24.310 { 00:28:24.310 "name": "BaseBdev2", 00:28:24.310 "uuid": "81b2b81e-f301-5622-a2b1-03e2bef0a32c", 00:28:24.310 "is_configured": true, 00:28:24.310 "data_offset": 256, 00:28:24.310 "data_size": 7936 00:28:24.310 } 00:28:24.310 ] 00:28:24.310 }' 00:28:24.310 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:28:24.310 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:28:24.310 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:28:24.310 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:28:24.310 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@782 -- # killprocess 1012969 00:28:24.310 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@948 -- # '[' -z 1012969 ']' 00:28:24.310 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@952 -- # kill -0 1012969 00:28:24.310 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@953 -- # uname 00:28:24.310 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:28:24.310 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1012969 00:28:24.310 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:28:24.310 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:28:24.310 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1012969' 00:28:24.310 killing process with pid 1012969 00:28:24.310 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@967 -- # kill 1012969 00:28:24.310 Received shutdown signal, test time was about 60.000000 seconds 00:28:24.310 00:28:24.310 Latency(us) 00:28:24.310 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:24.310 =================================================================================================================== 00:28:24.310 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:28:24.310 [2024-07-15 11:39:07.901542] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:28:24.310 11:39:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@972 -- # wait 1012969 00:28:24.310 [2024-07-15 11:39:07.901626] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:24.310 [2024-07-15 11:39:07.901670] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:24.310 [2024-07-15 11:39:07.901682] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x13467c0 name raid_bdev1, state offline 00:28:24.568 [2024-07-15 11:39:07.934915] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:28:24.568 11:39:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@784 -- # return 0 00:28:24.568 00:28:24.568 real 0m31.275s 00:28:24.568 user 0m48.812s 00:28:24.568 sys 0m5.048s 00:28:24.568 11:39:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1124 -- # xtrace_disable 00:28:24.568 11:39:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:28:24.568 ************************************ 00:28:24.568 END TEST raid_rebuild_test_sb_md_separate 00:28:24.568 ************************************ 00:28:24.827 11:39:08 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:28:24.827 11:39:08 bdev_raid -- bdev/bdev_raid.sh@911 -- # base_malloc_params='-m 32 -i' 00:28:24.827 11:39:08 bdev_raid -- bdev/bdev_raid.sh@912 -- # run_test raid_state_function_test_sb_md_interleaved raid_state_function_test raid1 2 true 00:28:24.827 11:39:08 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:28:24.827 11:39:08 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:28:24.827 11:39:08 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:28:24.827 ************************************ 00:28:24.827 START TEST raid_state_function_test_sb_md_interleaved 00:28:24.827 ************************************ 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1123 -- # raid_state_function_test raid1 2 true 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@220 -- # local raid_level=raid1 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@221 -- # local num_base_bdevs=2 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@222 -- # local superblock=true 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@223 -- # local raid_bdev 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@224 -- # (( i = 1 )) 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@226 -- # echo BaseBdev1 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@226 -- # echo BaseBdev2 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@224 -- # (( i++ )) 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@224 -- # (( i <= num_base_bdevs )) 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@224 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@224 -- # local base_bdevs 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@225 -- # local raid_bdev_name=Existed_Raid 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@226 -- # local strip_size 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@227 -- # local strip_size_create_arg 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@228 -- # local superblock_create_arg 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@230 -- # '[' raid1 '!=' raid1 ']' 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@234 -- # strip_size=0 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@237 -- # '[' true = true ']' 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@238 -- # superblock_create_arg=-s 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@244 -- # raid_pid=1017951 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@245 -- # echo 'Process raid pid: 1017951' 00:28:24.827 Process raid pid: 1017951 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@246 -- # waitforlisten 1017951 /var/tmp/spdk-raid.sock 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@829 -- # '[' -z 1017951 ']' 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@834 -- # local max_retries=100 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:28:24.827 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@838 -- # xtrace_disable 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@243 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -i 0 -L bdev_raid 00:28:24.827 11:39:08 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:28:24.827 [2024-07-15 11:39:08.285980] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:28:24.827 [2024-07-15 11:39:08.286043] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:28:24.827 [2024-07-15 11:39:08.414219] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:25.086 [2024-07-15 11:39:08.520399] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:28:25.086 [2024-07-15 11:39:08.582704] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:25.086 [2024-07-15 11:39:08.582736] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:25.652 11:39:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:28:25.652 11:39:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@862 -- # return 0 00:28:25.652 11:39:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:28:25.911 [2024-07-15 11:39:09.421348] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:28:25.911 [2024-07-15 11:39:09.421388] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:28:25.911 [2024-07-15 11:39:09.421399] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:28:25.911 [2024-07-15 11:39:09.421411] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:28:25.911 11:39:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:28:25.911 11:39:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:28:25.911 11:39:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:28:25.911 11:39:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:25.911 11:39:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:25.911 11:39:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:28:25.911 11:39:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:25.911 11:39:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:25.911 11:39:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:25.911 11:39:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:25.911 11:39:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:25.911 11:39:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:26.170 11:39:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:26.170 "name": "Existed_Raid", 00:28:26.170 "uuid": "860a47b3-4d33-44cc-9eaa-44fb19038aeb", 00:28:26.170 "strip_size_kb": 0, 00:28:26.170 "state": "configuring", 00:28:26.170 "raid_level": "raid1", 00:28:26.170 "superblock": true, 00:28:26.170 "num_base_bdevs": 2, 00:28:26.170 "num_base_bdevs_discovered": 0, 00:28:26.170 "num_base_bdevs_operational": 2, 00:28:26.170 "base_bdevs_list": [ 00:28:26.170 { 00:28:26.170 "name": "BaseBdev1", 00:28:26.170 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:26.170 "is_configured": false, 00:28:26.170 "data_offset": 0, 00:28:26.170 "data_size": 0 00:28:26.170 }, 00:28:26.170 { 00:28:26.170 "name": "BaseBdev2", 00:28:26.170 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:26.170 "is_configured": false, 00:28:26.170 "data_offset": 0, 00:28:26.170 "data_size": 0 00:28:26.170 } 00:28:26.170 ] 00:28:26.170 }' 00:28:26.170 11:39:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:26.170 11:39:09 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:28:26.736 11:39:10 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@252 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:28:26.994 [2024-07-15 11:39:10.492046] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:28:26.994 [2024-07-15 11:39:10.492080] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1163a80 name Existed_Raid, state configuring 00:28:26.994 11:39:10 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@256 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:28:27.251 [2024-07-15 11:39:10.736708] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:28:27.251 [2024-07-15 11:39:10.736733] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:28:27.251 [2024-07-15 11:39:10.736743] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:28:27.251 [2024-07-15 11:39:10.736754] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:28:27.251 11:39:10 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@257 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1 00:28:27.508 [2024-07-15 11:39:10.987384] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:27.508 BaseBdev1 00:28:27.508 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@258 -- # waitforbdev BaseBdev1 00:28:27.508 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev1 00:28:27.508 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:28:27.508 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@899 -- # local i 00:28:27.508 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:28:27.508 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:28:27.508 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:28:27.766 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 -t 2000 00:28:28.025 [ 00:28:28.025 { 00:28:28.025 "name": "BaseBdev1", 00:28:28.025 "aliases": [ 00:28:28.025 "dfdaa172-5d83-4ee4-b911-57232bd1b841" 00:28:28.025 ], 00:28:28.025 "product_name": "Malloc disk", 00:28:28.025 "block_size": 4128, 00:28:28.025 "num_blocks": 8192, 00:28:28.025 "uuid": "dfdaa172-5d83-4ee4-b911-57232bd1b841", 00:28:28.025 "md_size": 32, 00:28:28.025 "md_interleave": true, 00:28:28.025 "dif_type": 0, 00:28:28.025 "assigned_rate_limits": { 00:28:28.025 "rw_ios_per_sec": 0, 00:28:28.025 "rw_mbytes_per_sec": 0, 00:28:28.025 "r_mbytes_per_sec": 0, 00:28:28.025 "w_mbytes_per_sec": 0 00:28:28.025 }, 00:28:28.025 "claimed": true, 00:28:28.025 "claim_type": "exclusive_write", 00:28:28.025 "zoned": false, 00:28:28.025 "supported_io_types": { 00:28:28.025 "read": true, 00:28:28.025 "write": true, 00:28:28.025 "unmap": true, 00:28:28.025 "flush": true, 00:28:28.025 "reset": true, 00:28:28.025 "nvme_admin": false, 00:28:28.025 "nvme_io": false, 00:28:28.025 "nvme_io_md": false, 00:28:28.025 "write_zeroes": true, 00:28:28.025 "zcopy": true, 00:28:28.025 "get_zone_info": false, 00:28:28.025 "zone_management": false, 00:28:28.025 "zone_append": false, 00:28:28.025 "compare": false, 00:28:28.025 "compare_and_write": false, 00:28:28.025 "abort": true, 00:28:28.025 "seek_hole": false, 00:28:28.025 "seek_data": false, 00:28:28.025 "copy": true, 00:28:28.025 "nvme_iov_md": false 00:28:28.025 }, 00:28:28.025 "memory_domains": [ 00:28:28.025 { 00:28:28.025 "dma_device_id": "system", 00:28:28.025 "dma_device_type": 1 00:28:28.025 }, 00:28:28.025 { 00:28:28.025 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:28.025 "dma_device_type": 2 00:28:28.025 } 00:28:28.025 ], 00:28:28.025 "driver_specific": {} 00:28:28.025 } 00:28:28.025 ] 00:28:28.025 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@905 -- # return 0 00:28:28.025 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@259 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:28:28.025 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:28:28.025 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:28:28.025 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:28.025 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:28.025 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:28:28.025 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:28.025 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:28.025 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:28.025 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:28.025 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:28.025 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:28.284 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:28.284 "name": "Existed_Raid", 00:28:28.284 "uuid": "82c70b52-ec4c-487e-ad98-04136bc039b6", 00:28:28.284 "strip_size_kb": 0, 00:28:28.284 "state": "configuring", 00:28:28.284 "raid_level": "raid1", 00:28:28.284 "superblock": true, 00:28:28.284 "num_base_bdevs": 2, 00:28:28.284 "num_base_bdevs_discovered": 1, 00:28:28.284 "num_base_bdevs_operational": 2, 00:28:28.284 "base_bdevs_list": [ 00:28:28.284 { 00:28:28.284 "name": "BaseBdev1", 00:28:28.284 "uuid": "dfdaa172-5d83-4ee4-b911-57232bd1b841", 00:28:28.284 "is_configured": true, 00:28:28.284 "data_offset": 256, 00:28:28.284 "data_size": 7936 00:28:28.284 }, 00:28:28.284 { 00:28:28.284 "name": "BaseBdev2", 00:28:28.284 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:28.284 "is_configured": false, 00:28:28.284 "data_offset": 0, 00:28:28.284 "data_size": 0 00:28:28.284 } 00:28:28.284 ] 00:28:28.284 }' 00:28:28.284 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:28.284 11:39:11 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:28:28.849 11:39:12 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@260 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete Existed_Raid 00:28:29.107 [2024-07-15 11:39:12.487447] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:28:29.107 [2024-07-15 11:39:12.487486] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1163350 name Existed_Raid, state configuring 00:28:29.107 11:39:12 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@264 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2' -n Existed_Raid 00:28:29.365 [2024-07-15 11:39:12.732126] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:29.365 [2024-07-15 11:39:12.733592] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:28:29.365 [2024-07-15 11:39:12.733622] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:28:29.365 11:39:12 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@265 -- # (( i = 1 )) 00:28:29.365 11:39:12 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:28:29.365 11:39:12 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:28:29.365 11:39:12 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:28:29.365 11:39:12 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:28:29.365 11:39:12 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:29.365 11:39:12 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:29.365 11:39:12 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:28:29.365 11:39:12 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:29.365 11:39:12 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:29.365 11:39:12 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:29.365 11:39:12 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:29.365 11:39:12 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:29.365 11:39:12 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:29.623 11:39:13 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:29.623 "name": "Existed_Raid", 00:28:29.623 "uuid": "d2504b6a-49c0-4e03-a41a-493e9f972d22", 00:28:29.623 "strip_size_kb": 0, 00:28:29.623 "state": "configuring", 00:28:29.623 "raid_level": "raid1", 00:28:29.623 "superblock": true, 00:28:29.623 "num_base_bdevs": 2, 00:28:29.623 "num_base_bdevs_discovered": 1, 00:28:29.623 "num_base_bdevs_operational": 2, 00:28:29.623 "base_bdevs_list": [ 00:28:29.623 { 00:28:29.623 "name": "BaseBdev1", 00:28:29.623 "uuid": "dfdaa172-5d83-4ee4-b911-57232bd1b841", 00:28:29.623 "is_configured": true, 00:28:29.623 "data_offset": 256, 00:28:29.623 "data_size": 7936 00:28:29.623 }, 00:28:29.623 { 00:28:29.623 "name": "BaseBdev2", 00:28:29.623 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:29.623 "is_configured": false, 00:28:29.623 "data_offset": 0, 00:28:29.623 "data_size": 0 00:28:29.623 } 00:28:29.623 ] 00:28:29.623 }' 00:28:29.623 11:39:13 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:29.623 11:39:13 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:28:30.188 11:39:13 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@267 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2 00:28:30.446 [2024-07-15 11:39:13.790459] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:30.446 [2024-07-15 11:39:13.790596] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1165180 00:28:30.446 [2024-07-15 11:39:13.790609] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:28:30.446 [2024-07-15 11:39:13.790672] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1165150 00:28:30.446 [2024-07-15 11:39:13.790748] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1165180 00:28:30.446 [2024-07-15 11:39:13.790758] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x1165180 00:28:30.446 [2024-07-15 11:39:13.790812] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:30.446 BaseBdev2 00:28:30.446 11:39:13 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@268 -- # waitforbdev BaseBdev2 00:28:30.446 11:39:13 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@897 -- # local bdev_name=BaseBdev2 00:28:30.446 11:39:13 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:28:30.446 11:39:13 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@899 -- # local i 00:28:30.446 11:39:13 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:28:30.446 11:39:13 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:28:30.446 11:39:13 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_wait_for_examine 00:28:30.704 11:39:14 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 -t 2000 00:28:30.704 [ 00:28:30.704 { 00:28:30.704 "name": "BaseBdev2", 00:28:30.704 "aliases": [ 00:28:30.704 "0a9ceca4-5a30-4d68-b8ed-aa82a05dacac" 00:28:30.704 ], 00:28:30.704 "product_name": "Malloc disk", 00:28:30.704 "block_size": 4128, 00:28:30.704 "num_blocks": 8192, 00:28:30.704 "uuid": "0a9ceca4-5a30-4d68-b8ed-aa82a05dacac", 00:28:30.704 "md_size": 32, 00:28:30.704 "md_interleave": true, 00:28:30.704 "dif_type": 0, 00:28:30.704 "assigned_rate_limits": { 00:28:30.704 "rw_ios_per_sec": 0, 00:28:30.704 "rw_mbytes_per_sec": 0, 00:28:30.704 "r_mbytes_per_sec": 0, 00:28:30.704 "w_mbytes_per_sec": 0 00:28:30.704 }, 00:28:30.704 "claimed": true, 00:28:30.704 "claim_type": "exclusive_write", 00:28:30.704 "zoned": false, 00:28:30.704 "supported_io_types": { 00:28:30.704 "read": true, 00:28:30.704 "write": true, 00:28:30.704 "unmap": true, 00:28:30.704 "flush": true, 00:28:30.704 "reset": true, 00:28:30.704 "nvme_admin": false, 00:28:30.704 "nvme_io": false, 00:28:30.704 "nvme_io_md": false, 00:28:30.704 "write_zeroes": true, 00:28:30.704 "zcopy": true, 00:28:30.704 "get_zone_info": false, 00:28:30.704 "zone_management": false, 00:28:30.704 "zone_append": false, 00:28:30.704 "compare": false, 00:28:30.704 "compare_and_write": false, 00:28:30.704 "abort": true, 00:28:30.704 "seek_hole": false, 00:28:30.704 "seek_data": false, 00:28:30.704 "copy": true, 00:28:30.704 "nvme_iov_md": false 00:28:30.704 }, 00:28:30.704 "memory_domains": [ 00:28:30.704 { 00:28:30.704 "dma_device_id": "system", 00:28:30.704 "dma_device_type": 1 00:28:30.704 }, 00:28:30.704 { 00:28:30.704 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:30.704 "dma_device_type": 2 00:28:30.704 } 00:28:30.704 ], 00:28:30.704 "driver_specific": {} 00:28:30.704 } 00:28:30.704 ] 00:28:30.961 11:39:14 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@905 -- # return 0 00:28:30.961 11:39:14 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@265 -- # (( i++ )) 00:28:30.961 11:39:14 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@265 -- # (( i < num_base_bdevs )) 00:28:30.961 11:39:14 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:28:30.961 11:39:14 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:28:30.961 11:39:14 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:28:30.961 11:39:14 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:30.961 11:39:14 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:30.961 11:39:14 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:28:30.961 11:39:14 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:30.961 11:39:14 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:30.961 11:39:14 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:30.961 11:39:14 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:30.961 11:39:14 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:30.961 11:39:14 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:30.961 11:39:14 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:30.961 "name": "Existed_Raid", 00:28:30.961 "uuid": "d2504b6a-49c0-4e03-a41a-493e9f972d22", 00:28:30.961 "strip_size_kb": 0, 00:28:30.961 "state": "online", 00:28:30.961 "raid_level": "raid1", 00:28:30.961 "superblock": true, 00:28:30.961 "num_base_bdevs": 2, 00:28:30.961 "num_base_bdevs_discovered": 2, 00:28:30.961 "num_base_bdevs_operational": 2, 00:28:30.961 "base_bdevs_list": [ 00:28:30.961 { 00:28:30.961 "name": "BaseBdev1", 00:28:30.961 "uuid": "dfdaa172-5d83-4ee4-b911-57232bd1b841", 00:28:30.961 "is_configured": true, 00:28:30.961 "data_offset": 256, 00:28:30.961 "data_size": 7936 00:28:30.961 }, 00:28:30.961 { 00:28:30.961 "name": "BaseBdev2", 00:28:30.961 "uuid": "0a9ceca4-5a30-4d68-b8ed-aa82a05dacac", 00:28:30.961 "is_configured": true, 00:28:30.961 "data_offset": 256, 00:28:30.961 "data_size": 7936 00:28:30.961 } 00:28:30.961 ] 00:28:30.961 }' 00:28:30.961 11:39:14 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:30.961 11:39:14 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:28:31.893 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # verify_raid_bdev_properties Existed_Raid 00:28:31.893 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=Existed_Raid 00:28:31.893 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:28:31.893 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:28:31.893 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:28:31.893 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@198 -- # local name 00:28:31.893 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b Existed_Raid 00:28:31.893 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:28:31.893 [2024-07-15 11:39:15.362919] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:31.893 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:28:31.893 "name": "Existed_Raid", 00:28:31.893 "aliases": [ 00:28:31.893 "d2504b6a-49c0-4e03-a41a-493e9f972d22" 00:28:31.893 ], 00:28:31.893 "product_name": "Raid Volume", 00:28:31.893 "block_size": 4128, 00:28:31.893 "num_blocks": 7936, 00:28:31.893 "uuid": "d2504b6a-49c0-4e03-a41a-493e9f972d22", 00:28:31.893 "md_size": 32, 00:28:31.893 "md_interleave": true, 00:28:31.893 "dif_type": 0, 00:28:31.893 "assigned_rate_limits": { 00:28:31.893 "rw_ios_per_sec": 0, 00:28:31.893 "rw_mbytes_per_sec": 0, 00:28:31.893 "r_mbytes_per_sec": 0, 00:28:31.893 "w_mbytes_per_sec": 0 00:28:31.893 }, 00:28:31.893 "claimed": false, 00:28:31.893 "zoned": false, 00:28:31.893 "supported_io_types": { 00:28:31.893 "read": true, 00:28:31.893 "write": true, 00:28:31.893 "unmap": false, 00:28:31.893 "flush": false, 00:28:31.893 "reset": true, 00:28:31.893 "nvme_admin": false, 00:28:31.893 "nvme_io": false, 00:28:31.893 "nvme_io_md": false, 00:28:31.893 "write_zeroes": true, 00:28:31.893 "zcopy": false, 00:28:31.893 "get_zone_info": false, 00:28:31.893 "zone_management": false, 00:28:31.893 "zone_append": false, 00:28:31.893 "compare": false, 00:28:31.893 "compare_and_write": false, 00:28:31.893 "abort": false, 00:28:31.893 "seek_hole": false, 00:28:31.893 "seek_data": false, 00:28:31.893 "copy": false, 00:28:31.893 "nvme_iov_md": false 00:28:31.893 }, 00:28:31.893 "memory_domains": [ 00:28:31.893 { 00:28:31.893 "dma_device_id": "system", 00:28:31.893 "dma_device_type": 1 00:28:31.893 }, 00:28:31.893 { 00:28:31.893 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:31.893 "dma_device_type": 2 00:28:31.893 }, 00:28:31.893 { 00:28:31.893 "dma_device_id": "system", 00:28:31.893 "dma_device_type": 1 00:28:31.893 }, 00:28:31.893 { 00:28:31.893 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:31.893 "dma_device_type": 2 00:28:31.893 } 00:28:31.893 ], 00:28:31.893 "driver_specific": { 00:28:31.893 "raid": { 00:28:31.893 "uuid": "d2504b6a-49c0-4e03-a41a-493e9f972d22", 00:28:31.893 "strip_size_kb": 0, 00:28:31.893 "state": "online", 00:28:31.893 "raid_level": "raid1", 00:28:31.893 "superblock": true, 00:28:31.893 "num_base_bdevs": 2, 00:28:31.893 "num_base_bdevs_discovered": 2, 00:28:31.893 "num_base_bdevs_operational": 2, 00:28:31.893 "base_bdevs_list": [ 00:28:31.893 { 00:28:31.893 "name": "BaseBdev1", 00:28:31.893 "uuid": "dfdaa172-5d83-4ee4-b911-57232bd1b841", 00:28:31.893 "is_configured": true, 00:28:31.893 "data_offset": 256, 00:28:31.893 "data_size": 7936 00:28:31.893 }, 00:28:31.893 { 00:28:31.893 "name": "BaseBdev2", 00:28:31.893 "uuid": "0a9ceca4-5a30-4d68-b8ed-aa82a05dacac", 00:28:31.893 "is_configured": true, 00:28:31.893 "data_offset": 256, 00:28:31.893 "data_size": 7936 00:28:31.893 } 00:28:31.893 ] 00:28:31.893 } 00:28:31.893 } 00:28:31.893 }' 00:28:31.893 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:28:31.893 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@201 -- # base_bdev_names='BaseBdev1 00:28:31.893 BaseBdev2' 00:28:31.893 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:28:31.893 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev1 00:28:31.893 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:28:32.151 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:28:32.151 "name": "BaseBdev1", 00:28:32.151 "aliases": [ 00:28:32.151 "dfdaa172-5d83-4ee4-b911-57232bd1b841" 00:28:32.151 ], 00:28:32.151 "product_name": "Malloc disk", 00:28:32.151 "block_size": 4128, 00:28:32.151 "num_blocks": 8192, 00:28:32.151 "uuid": "dfdaa172-5d83-4ee4-b911-57232bd1b841", 00:28:32.151 "md_size": 32, 00:28:32.151 "md_interleave": true, 00:28:32.151 "dif_type": 0, 00:28:32.151 "assigned_rate_limits": { 00:28:32.151 "rw_ios_per_sec": 0, 00:28:32.151 "rw_mbytes_per_sec": 0, 00:28:32.151 "r_mbytes_per_sec": 0, 00:28:32.151 "w_mbytes_per_sec": 0 00:28:32.151 }, 00:28:32.151 "claimed": true, 00:28:32.151 "claim_type": "exclusive_write", 00:28:32.151 "zoned": false, 00:28:32.151 "supported_io_types": { 00:28:32.151 "read": true, 00:28:32.151 "write": true, 00:28:32.151 "unmap": true, 00:28:32.151 "flush": true, 00:28:32.151 "reset": true, 00:28:32.151 "nvme_admin": false, 00:28:32.151 "nvme_io": false, 00:28:32.151 "nvme_io_md": false, 00:28:32.151 "write_zeroes": true, 00:28:32.151 "zcopy": true, 00:28:32.151 "get_zone_info": false, 00:28:32.151 "zone_management": false, 00:28:32.151 "zone_append": false, 00:28:32.151 "compare": false, 00:28:32.151 "compare_and_write": false, 00:28:32.151 "abort": true, 00:28:32.151 "seek_hole": false, 00:28:32.151 "seek_data": false, 00:28:32.151 "copy": true, 00:28:32.151 "nvme_iov_md": false 00:28:32.151 }, 00:28:32.151 "memory_domains": [ 00:28:32.151 { 00:28:32.151 "dma_device_id": "system", 00:28:32.151 "dma_device_type": 1 00:28:32.151 }, 00:28:32.151 { 00:28:32.151 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:32.151 "dma_device_type": 2 00:28:32.151 } 00:28:32.151 ], 00:28:32.151 "driver_specific": {} 00:28:32.151 }' 00:28:32.151 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:28:32.151 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:28:32.408 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@205 -- # [[ 4128 == 4128 ]] 00:28:32.408 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:28:32.408 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:28:32.408 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@206 -- # [[ 32 == 32 ]] 00:28:32.408 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:28:32.408 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:28:32.408 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@207 -- # [[ true == true ]] 00:28:32.408 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:28:32.408 11:39:15 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:28:32.665 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@208 -- # [[ 0 == 0 ]] 00:28:32.665 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:28:32.665 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b BaseBdev2 00:28:32.665 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:28:32.665 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:28:32.665 "name": "BaseBdev2", 00:28:32.665 "aliases": [ 00:28:32.665 "0a9ceca4-5a30-4d68-b8ed-aa82a05dacac" 00:28:32.665 ], 00:28:32.665 "product_name": "Malloc disk", 00:28:32.665 "block_size": 4128, 00:28:32.665 "num_blocks": 8192, 00:28:32.665 "uuid": "0a9ceca4-5a30-4d68-b8ed-aa82a05dacac", 00:28:32.665 "md_size": 32, 00:28:32.665 "md_interleave": true, 00:28:32.665 "dif_type": 0, 00:28:32.665 "assigned_rate_limits": { 00:28:32.665 "rw_ios_per_sec": 0, 00:28:32.665 "rw_mbytes_per_sec": 0, 00:28:32.665 "r_mbytes_per_sec": 0, 00:28:32.665 "w_mbytes_per_sec": 0 00:28:32.665 }, 00:28:32.665 "claimed": true, 00:28:32.665 "claim_type": "exclusive_write", 00:28:32.665 "zoned": false, 00:28:32.665 "supported_io_types": { 00:28:32.666 "read": true, 00:28:32.666 "write": true, 00:28:32.666 "unmap": true, 00:28:32.666 "flush": true, 00:28:32.666 "reset": true, 00:28:32.666 "nvme_admin": false, 00:28:32.666 "nvme_io": false, 00:28:32.666 "nvme_io_md": false, 00:28:32.666 "write_zeroes": true, 00:28:32.666 "zcopy": true, 00:28:32.666 "get_zone_info": false, 00:28:32.666 "zone_management": false, 00:28:32.666 "zone_append": false, 00:28:32.666 "compare": false, 00:28:32.666 "compare_and_write": false, 00:28:32.666 "abort": true, 00:28:32.666 "seek_hole": false, 00:28:32.666 "seek_data": false, 00:28:32.666 "copy": true, 00:28:32.666 "nvme_iov_md": false 00:28:32.666 }, 00:28:32.666 "memory_domains": [ 00:28:32.666 { 00:28:32.666 "dma_device_id": "system", 00:28:32.666 "dma_device_type": 1 00:28:32.666 }, 00:28:32.666 { 00:28:32.666 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:32.666 "dma_device_type": 2 00:28:32.666 } 00:28:32.666 ], 00:28:32.666 "driver_specific": {} 00:28:32.666 }' 00:28:32.666 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:28:32.923 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:28:32.923 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@205 -- # [[ 4128 == 4128 ]] 00:28:32.923 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:28:32.923 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:28:32.923 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@206 -- # [[ 32 == 32 ]] 00:28:32.923 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:28:32.923 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:28:32.923 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@207 -- # [[ true == true ]] 00:28:33.181 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:28:33.181 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:28:33.181 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@208 -- # [[ 0 == 0 ]] 00:28:33.181 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@274 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev1 00:28:33.438 [2024-07-15 11:39:16.830596] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:28:33.438 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@275 -- # local expected_state 00:28:33.438 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@276 -- # has_redundancy raid1 00:28:33.438 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@213 -- # case $1 in 00:28:33.438 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@214 -- # return 0 00:28:33.438 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@279 -- # expected_state=online 00:28:33.438 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@281 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:28:33.438 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=Existed_Raid 00:28:33.438 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:28:33.438 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:33.438 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:33.438 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:28:33.438 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:33.438 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:33.438 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:33.438 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:33.438 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:33.438 11:39:16 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:33.696 11:39:17 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:33.696 "name": "Existed_Raid", 00:28:33.696 "uuid": "d2504b6a-49c0-4e03-a41a-493e9f972d22", 00:28:33.696 "strip_size_kb": 0, 00:28:33.696 "state": "online", 00:28:33.696 "raid_level": "raid1", 00:28:33.696 "superblock": true, 00:28:33.696 "num_base_bdevs": 2, 00:28:33.696 "num_base_bdevs_discovered": 1, 00:28:33.696 "num_base_bdevs_operational": 1, 00:28:33.696 "base_bdevs_list": [ 00:28:33.696 { 00:28:33.696 "name": null, 00:28:33.696 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:33.696 "is_configured": false, 00:28:33.696 "data_offset": 256, 00:28:33.696 "data_size": 7936 00:28:33.696 }, 00:28:33.696 { 00:28:33.696 "name": "BaseBdev2", 00:28:33.696 "uuid": "0a9ceca4-5a30-4d68-b8ed-aa82a05dacac", 00:28:33.696 "is_configured": true, 00:28:33.696 "data_offset": 256, 00:28:33.696 "data_size": 7936 00:28:33.696 } 00:28:33.696 ] 00:28:33.696 }' 00:28:33.696 11:39:17 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:33.696 11:39:17 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:28:34.260 11:39:17 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@285 -- # (( i = 1 )) 00:28:34.260 11:39:17 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:28:34.260 11:39:17 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@286 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:34.260 11:39:17 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@286 -- # jq -r '.[0]["name"]' 00:28:34.518 11:39:17 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@286 -- # raid_bdev=Existed_Raid 00:28:34.518 11:39:17 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@287 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:28:34.518 11:39:17 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@291 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_delete BaseBdev2 00:28:34.518 [2024-07-15 11:39:18.099061] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:28:34.518 [2024-07-15 11:39:18.099147] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:34.518 [2024-07-15 11:39:18.110964] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:34.518 [2024-07-15 11:39:18.110998] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:34.518 [2024-07-15 11:39:18.111010] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1165180 name Existed_Raid, state offline 00:28:34.775 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@285 -- # (( i++ )) 00:28:34.775 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@285 -- # (( i < num_base_bdevs )) 00:28:34.775 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@293 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:34.775 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@293 -- # jq -r '.[0]["name"] | select(.)' 00:28:35.033 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@293 -- # raid_bdev= 00:28:35.033 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@294 -- # '[' -n '' ']' 00:28:35.033 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@299 -- # '[' 2 -gt 2 ']' 00:28:35.033 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@341 -- # killprocess 1017951 00:28:35.033 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@948 -- # '[' -z 1017951 ']' 00:28:35.033 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@952 -- # kill -0 1017951 00:28:35.033 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@953 -- # uname 00:28:35.033 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:28:35.033 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1017951 00:28:35.033 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:28:35.033 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:28:35.033 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1017951' 00:28:35.033 killing process with pid 1017951 00:28:35.033 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@967 -- # kill 1017951 00:28:35.033 [2024-07-15 11:39:18.428191] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:28:35.033 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@972 -- # wait 1017951 00:28:35.033 [2024-07-15 11:39:18.429166] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:28:35.291 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@343 -- # return 0 00:28:35.291 00:28:35.291 real 0m10.438s 00:28:35.291 user 0m18.475s 00:28:35.291 sys 0m2.036s 00:28:35.291 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1124 -- # xtrace_disable 00:28:35.291 11:39:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:28:35.291 ************************************ 00:28:35.291 END TEST raid_state_function_test_sb_md_interleaved 00:28:35.291 ************************************ 00:28:35.291 11:39:18 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:28:35.291 11:39:18 bdev_raid -- bdev/bdev_raid.sh@913 -- # run_test raid_superblock_test_md_interleaved raid_superblock_test raid1 2 00:28:35.291 11:39:18 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:28:35.291 11:39:18 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:28:35.291 11:39:18 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:28:35.291 ************************************ 00:28:35.291 START TEST raid_superblock_test_md_interleaved 00:28:35.291 ************************************ 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1123 -- # raid_superblock_test raid1 2 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@392 -- # local raid_level=raid1 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@393 -- # local num_base_bdevs=2 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@394 -- # base_bdevs_malloc=() 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@394 -- # local base_bdevs_malloc 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # base_bdevs_pt=() 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # local base_bdevs_pt 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt_uuid=() 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt_uuid 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # local raid_bdev_name=raid_bdev1 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@398 -- # local strip_size 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@399 -- # local strip_size_create_arg 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@400 -- # local raid_bdev_uuid 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@401 -- # local raid_bdev 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@403 -- # '[' raid1 '!=' raid1 ']' 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@407 -- # strip_size=0 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@411 -- # raid_pid=1019539 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@412 -- # waitforlisten 1019539 /var/tmp/spdk-raid.sock 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@410 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-raid.sock -L bdev_raid 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@829 -- # '[' -z 1019539 ']' 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@834 -- # local max_retries=100 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:28:35.291 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@838 -- # xtrace_disable 00:28:35.291 11:39:18 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:28:35.291 [2024-07-15 11:39:18.808771] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:28:35.291 [2024-07-15 11:39:18.808836] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1019539 ] 00:28:35.549 [2024-07-15 11:39:18.937413] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:35.549 [2024-07-15 11:39:19.042933] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:28:35.549 [2024-07-15 11:39:19.110334] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:35.549 [2024-07-15 11:39:19.110370] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:36.205 11:39:19 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:28:36.205 11:39:19 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@862 -- # return 0 00:28:36.205 11:39:19 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@415 -- # (( i = 1 )) 00:28:36.205 11:39:19 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:28:36.205 11:39:19 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc1 00:28:36.205 11:39:19 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt1 00:28:36.205 11:39:19 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:28:36.205 11:39:19 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:28:36.205 11:39:19 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:28:36.205 11:39:19 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:28:36.205 11:39:19 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -m 32 -i -b malloc1 00:28:36.460 malloc1 00:28:36.460 11:39:19 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:28:36.758 [2024-07-15 11:39:20.168141] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:28:36.758 [2024-07-15 11:39:20.168195] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:36.758 [2024-07-15 11:39:20.168217] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xd5e4e0 00:28:36.758 [2024-07-15 11:39:20.168230] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:36.758 [2024-07-15 11:39:20.169765] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:36.758 [2024-07-15 11:39:20.169792] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:28:36.758 pt1 00:28:36.759 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:28:36.759 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:28:36.759 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # local bdev_malloc=malloc2 00:28:36.759 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_pt=pt2 00:28:36.759 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:28:36.759 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@420 -- # base_bdevs_malloc+=($bdev_malloc) 00:28:36.759 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_pt+=($bdev_pt) 00:28:36.759 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:28:36.759 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@424 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -m 32 -i -b malloc2 00:28:37.017 malloc2 00:28:37.017 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:28:37.017 [2024-07-15 11:39:20.610469] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:28:37.017 [2024-07-15 11:39:20.610517] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:37.017 [2024-07-15 11:39:20.610537] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xd43570 00:28:37.017 [2024-07-15 11:39:20.610549] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:37.274 [2024-07-15 11:39:20.612047] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:37.274 [2024-07-15 11:39:20.612074] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:28:37.274 pt2 00:28:37.274 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@415 -- # (( i++ )) 00:28:37.274 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@415 -- # (( i <= num_base_bdevs )) 00:28:37.274 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@429 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'pt1 pt2' -n raid_bdev1 -s 00:28:37.274 [2024-07-15 11:39:20.843099] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:28:37.274 [2024-07-15 11:39:20.844569] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:28:37.274 [2024-07-15 11:39:20.844721] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xd44f20 00:28:37.274 [2024-07-15 11:39:20.844735] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:28:37.274 [2024-07-15 11:39:20.844806] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xbc1050 00:28:37.274 [2024-07-15 11:39:20.844889] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xd44f20 00:28:37.274 [2024-07-15 11:39:20.844899] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0xd44f20 00:28:37.274 [2024-07-15 11:39:20.844965] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:37.274 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@430 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:28:37.274 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:28:37.274 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:28:37.274 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:37.274 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:37.274 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:28:37.274 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:37.274 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:37.274 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:37.274 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:37.274 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:37.274 11:39:20 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:37.531 11:39:21 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:37.531 "name": "raid_bdev1", 00:28:37.531 "uuid": "5877095e-6236-42ff-b2ba-4fb5b7a6add2", 00:28:37.531 "strip_size_kb": 0, 00:28:37.531 "state": "online", 00:28:37.531 "raid_level": "raid1", 00:28:37.531 "superblock": true, 00:28:37.531 "num_base_bdevs": 2, 00:28:37.531 "num_base_bdevs_discovered": 2, 00:28:37.531 "num_base_bdevs_operational": 2, 00:28:37.531 "base_bdevs_list": [ 00:28:37.531 { 00:28:37.531 "name": "pt1", 00:28:37.531 "uuid": "00000000-0000-0000-0000-000000000001", 00:28:37.531 "is_configured": true, 00:28:37.531 "data_offset": 256, 00:28:37.531 "data_size": 7936 00:28:37.531 }, 00:28:37.531 { 00:28:37.531 "name": "pt2", 00:28:37.531 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:37.531 "is_configured": true, 00:28:37.531 "data_offset": 256, 00:28:37.531 "data_size": 7936 00:28:37.531 } 00:28:37.531 ] 00:28:37.531 }' 00:28:37.531 11:39:21 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:37.531 11:39:21 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:28:38.094 11:39:21 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_properties raid_bdev1 00:28:38.094 11:39:21 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:28:38.094 11:39:21 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:28:38.094 11:39:21 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:28:38.094 11:39:21 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:28:38.094 11:39:21 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@198 -- # local name 00:28:38.094 11:39:21 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:28:38.351 11:39:21 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:28:38.351 [2024-07-15 11:39:21.914163] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:38.351 11:39:21 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:28:38.351 "name": "raid_bdev1", 00:28:38.351 "aliases": [ 00:28:38.351 "5877095e-6236-42ff-b2ba-4fb5b7a6add2" 00:28:38.351 ], 00:28:38.351 "product_name": "Raid Volume", 00:28:38.351 "block_size": 4128, 00:28:38.351 "num_blocks": 7936, 00:28:38.351 "uuid": "5877095e-6236-42ff-b2ba-4fb5b7a6add2", 00:28:38.351 "md_size": 32, 00:28:38.351 "md_interleave": true, 00:28:38.351 "dif_type": 0, 00:28:38.351 "assigned_rate_limits": { 00:28:38.351 "rw_ios_per_sec": 0, 00:28:38.351 "rw_mbytes_per_sec": 0, 00:28:38.352 "r_mbytes_per_sec": 0, 00:28:38.352 "w_mbytes_per_sec": 0 00:28:38.352 }, 00:28:38.352 "claimed": false, 00:28:38.352 "zoned": false, 00:28:38.352 "supported_io_types": { 00:28:38.352 "read": true, 00:28:38.352 "write": true, 00:28:38.352 "unmap": false, 00:28:38.352 "flush": false, 00:28:38.352 "reset": true, 00:28:38.352 "nvme_admin": false, 00:28:38.352 "nvme_io": false, 00:28:38.352 "nvme_io_md": false, 00:28:38.352 "write_zeroes": true, 00:28:38.352 "zcopy": false, 00:28:38.352 "get_zone_info": false, 00:28:38.352 "zone_management": false, 00:28:38.352 "zone_append": false, 00:28:38.352 "compare": false, 00:28:38.352 "compare_and_write": false, 00:28:38.352 "abort": false, 00:28:38.352 "seek_hole": false, 00:28:38.352 "seek_data": false, 00:28:38.352 "copy": false, 00:28:38.352 "nvme_iov_md": false 00:28:38.352 }, 00:28:38.352 "memory_domains": [ 00:28:38.352 { 00:28:38.352 "dma_device_id": "system", 00:28:38.352 "dma_device_type": 1 00:28:38.352 }, 00:28:38.352 { 00:28:38.352 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:38.352 "dma_device_type": 2 00:28:38.352 }, 00:28:38.352 { 00:28:38.352 "dma_device_id": "system", 00:28:38.352 "dma_device_type": 1 00:28:38.352 }, 00:28:38.352 { 00:28:38.352 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:38.352 "dma_device_type": 2 00:28:38.352 } 00:28:38.352 ], 00:28:38.352 "driver_specific": { 00:28:38.352 "raid": { 00:28:38.352 "uuid": "5877095e-6236-42ff-b2ba-4fb5b7a6add2", 00:28:38.352 "strip_size_kb": 0, 00:28:38.352 "state": "online", 00:28:38.352 "raid_level": "raid1", 00:28:38.352 "superblock": true, 00:28:38.352 "num_base_bdevs": 2, 00:28:38.352 "num_base_bdevs_discovered": 2, 00:28:38.352 "num_base_bdevs_operational": 2, 00:28:38.352 "base_bdevs_list": [ 00:28:38.352 { 00:28:38.352 "name": "pt1", 00:28:38.352 "uuid": "00000000-0000-0000-0000-000000000001", 00:28:38.352 "is_configured": true, 00:28:38.352 "data_offset": 256, 00:28:38.352 "data_size": 7936 00:28:38.352 }, 00:28:38.352 { 00:28:38.352 "name": "pt2", 00:28:38.352 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:38.352 "is_configured": true, 00:28:38.352 "data_offset": 256, 00:28:38.352 "data_size": 7936 00:28:38.352 } 00:28:38.352 ] 00:28:38.352 } 00:28:38.352 } 00:28:38.352 }' 00:28:38.352 11:39:21 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:28:38.609 11:39:21 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:28:38.609 pt2' 00:28:38.609 11:39:21 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:28:38.609 11:39:21 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:28:38.609 11:39:21 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:28:38.867 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:28:38.867 "name": "pt1", 00:28:38.867 "aliases": [ 00:28:38.867 "00000000-0000-0000-0000-000000000001" 00:28:38.867 ], 00:28:38.867 "product_name": "passthru", 00:28:38.867 "block_size": 4128, 00:28:38.867 "num_blocks": 8192, 00:28:38.867 "uuid": "00000000-0000-0000-0000-000000000001", 00:28:38.867 "md_size": 32, 00:28:38.867 "md_interleave": true, 00:28:38.867 "dif_type": 0, 00:28:38.867 "assigned_rate_limits": { 00:28:38.867 "rw_ios_per_sec": 0, 00:28:38.867 "rw_mbytes_per_sec": 0, 00:28:38.867 "r_mbytes_per_sec": 0, 00:28:38.867 "w_mbytes_per_sec": 0 00:28:38.867 }, 00:28:38.867 "claimed": true, 00:28:38.867 "claim_type": "exclusive_write", 00:28:38.867 "zoned": false, 00:28:38.867 "supported_io_types": { 00:28:38.867 "read": true, 00:28:38.867 "write": true, 00:28:38.867 "unmap": true, 00:28:38.867 "flush": true, 00:28:38.867 "reset": true, 00:28:38.867 "nvme_admin": false, 00:28:38.867 "nvme_io": false, 00:28:38.867 "nvme_io_md": false, 00:28:38.867 "write_zeroes": true, 00:28:38.867 "zcopy": true, 00:28:38.867 "get_zone_info": false, 00:28:38.867 "zone_management": false, 00:28:38.867 "zone_append": false, 00:28:38.867 "compare": false, 00:28:38.867 "compare_and_write": false, 00:28:38.867 "abort": true, 00:28:38.867 "seek_hole": false, 00:28:38.867 "seek_data": false, 00:28:38.867 "copy": true, 00:28:38.867 "nvme_iov_md": false 00:28:38.867 }, 00:28:38.867 "memory_domains": [ 00:28:38.867 { 00:28:38.867 "dma_device_id": "system", 00:28:38.867 "dma_device_type": 1 00:28:38.867 }, 00:28:38.867 { 00:28:38.867 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:38.867 "dma_device_type": 2 00:28:38.867 } 00:28:38.867 ], 00:28:38.867 "driver_specific": { 00:28:38.867 "passthru": { 00:28:38.867 "name": "pt1", 00:28:38.868 "base_bdev_name": "malloc1" 00:28:38.868 } 00:28:38.868 } 00:28:38.868 }' 00:28:38.868 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:28:38.868 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:28:38.868 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@205 -- # [[ 4128 == 4128 ]] 00:28:38.868 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:28:38.868 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:28:38.868 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@206 -- # [[ 32 == 32 ]] 00:28:38.868 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:28:38.868 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:28:39.126 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@207 -- # [[ true == true ]] 00:28:39.126 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:28:39.126 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:28:39.126 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@208 -- # [[ 0 == 0 ]] 00:28:39.126 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:28:39.126 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:28:39.126 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:28:39.384 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:28:39.384 "name": "pt2", 00:28:39.384 "aliases": [ 00:28:39.384 "00000000-0000-0000-0000-000000000002" 00:28:39.384 ], 00:28:39.384 "product_name": "passthru", 00:28:39.384 "block_size": 4128, 00:28:39.384 "num_blocks": 8192, 00:28:39.384 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:39.384 "md_size": 32, 00:28:39.384 "md_interleave": true, 00:28:39.384 "dif_type": 0, 00:28:39.384 "assigned_rate_limits": { 00:28:39.384 "rw_ios_per_sec": 0, 00:28:39.384 "rw_mbytes_per_sec": 0, 00:28:39.384 "r_mbytes_per_sec": 0, 00:28:39.384 "w_mbytes_per_sec": 0 00:28:39.384 }, 00:28:39.384 "claimed": true, 00:28:39.384 "claim_type": "exclusive_write", 00:28:39.384 "zoned": false, 00:28:39.384 "supported_io_types": { 00:28:39.384 "read": true, 00:28:39.384 "write": true, 00:28:39.384 "unmap": true, 00:28:39.384 "flush": true, 00:28:39.384 "reset": true, 00:28:39.384 "nvme_admin": false, 00:28:39.384 "nvme_io": false, 00:28:39.384 "nvme_io_md": false, 00:28:39.384 "write_zeroes": true, 00:28:39.384 "zcopy": true, 00:28:39.384 "get_zone_info": false, 00:28:39.384 "zone_management": false, 00:28:39.384 "zone_append": false, 00:28:39.384 "compare": false, 00:28:39.384 "compare_and_write": false, 00:28:39.384 "abort": true, 00:28:39.384 "seek_hole": false, 00:28:39.384 "seek_data": false, 00:28:39.384 "copy": true, 00:28:39.384 "nvme_iov_md": false 00:28:39.384 }, 00:28:39.384 "memory_domains": [ 00:28:39.384 { 00:28:39.384 "dma_device_id": "system", 00:28:39.384 "dma_device_type": 1 00:28:39.384 }, 00:28:39.384 { 00:28:39.384 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:39.384 "dma_device_type": 2 00:28:39.384 } 00:28:39.384 ], 00:28:39.384 "driver_specific": { 00:28:39.384 "passthru": { 00:28:39.384 "name": "pt2", 00:28:39.384 "base_bdev_name": "malloc2" 00:28:39.384 } 00:28:39.384 } 00:28:39.384 }' 00:28:39.384 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:28:39.384 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:28:39.384 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@205 -- # [[ 4128 == 4128 ]] 00:28:39.384 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:28:39.384 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:28:39.384 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@206 -- # [[ 32 == 32 ]] 00:28:39.384 11:39:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:28:39.642 11:39:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:28:39.642 11:39:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@207 -- # [[ true == true ]] 00:28:39.642 11:39:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:28:39.642 11:39:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:28:39.642 11:39:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@208 -- # [[ 0 == 0 ]] 00:28:39.642 11:39:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@434 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:28:39.642 11:39:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@434 -- # jq -r '.[] | .uuid' 00:28:39.899 [2024-07-15 11:39:23.366000] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:39.899 11:39:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@434 -- # raid_bdev_uuid=5877095e-6236-42ff-b2ba-4fb5b7a6add2 00:28:39.899 11:39:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # '[' -z 5877095e-6236-42ff-b2ba-4fb5b7a6add2 ']' 00:28:39.899 11:39:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@440 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:28:40.157 [2024-07-15 11:39:23.602371] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:28:40.157 [2024-07-15 11:39:23.602391] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:40.157 [2024-07-15 11:39:23.602449] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:40.157 [2024-07-15 11:39:23.602501] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:40.157 [2024-07-15 11:39:23.602513] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xd44f20 name raid_bdev1, state offline 00:28:40.157 11:39:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@441 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:40.157 11:39:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@441 -- # jq -r '.[]' 00:28:40.415 11:39:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@441 -- # raid_bdev= 00:28:40.415 11:39:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # '[' -n '' ']' 00:28:40.415 11:39:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:28:40.415 11:39:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt1 00:28:40.672 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@447 -- # for i in "${base_bdevs_pt[@]}" 00:28:40.672 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:28:40.930 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@450 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs 00:28:40.930 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@450 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:28:41.189 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@450 -- # '[' false == true ']' 00:28:41.189 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@456 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'malloc1 malloc2' -n raid_bdev1 00:28:41.189 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@648 -- # local es=0 00:28:41.189 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'malloc1 malloc2' -n raid_bdev1 00:28:41.189 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:28:41.189 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:28:41.189 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:28:41.189 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:28:41.189 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:28:41.189 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:28:41.189 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:28:41.189 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py ]] 00:28:41.189 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -r raid1 -b 'malloc1 malloc2' -n raid_bdev1 00:28:41.447 [2024-07-15 11:39:24.821551] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:28:41.447 [2024-07-15 11:39:24.822950] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:28:41.447 [2024-07-15 11:39:24.823006] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:28:41.447 [2024-07-15 11:39:24.823044] bdev_raid.c:3106:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:28:41.447 [2024-07-15 11:39:24.823063] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:28:41.447 [2024-07-15 11:39:24.823073] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xd4f260 name raid_bdev1, state configuring 00:28:41.447 request: 00:28:41.447 { 00:28:41.447 "name": "raid_bdev1", 00:28:41.447 "raid_level": "raid1", 00:28:41.447 "base_bdevs": [ 00:28:41.447 "malloc1", 00:28:41.447 "malloc2" 00:28:41.447 ], 00:28:41.447 "superblock": false, 00:28:41.447 "method": "bdev_raid_create", 00:28:41.447 "req_id": 1 00:28:41.447 } 00:28:41.447 Got JSON-RPC error response 00:28:41.447 response: 00:28:41.447 { 00:28:41.447 "code": -17, 00:28:41.447 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:28:41.447 } 00:28:41.447 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@651 -- # es=1 00:28:41.447 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:28:41.447 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:28:41.447 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:28:41.447 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@458 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:41.447 11:39:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@458 -- # jq -r '.[]' 00:28:41.705 11:39:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@458 -- # raid_bdev= 00:28:41.705 11:39:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # '[' -n '' ']' 00:28:41.705 11:39:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@464 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:28:41.963 [2024-07-15 11:39:25.314797] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:28:41.963 [2024-07-15 11:39:25.314836] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:41.963 [2024-07-15 11:39:25.314853] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xd46000 00:28:41.963 [2024-07-15 11:39:25.314865] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:41.963 [2024-07-15 11:39:25.316276] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:41.963 [2024-07-15 11:39:25.316303] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:28:41.963 [2024-07-15 11:39:25.316351] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:28:41.963 [2024-07-15 11:39:25.316375] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:28:41.963 pt1 00:28:41.963 11:39:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@467 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:28:41.963 11:39:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:28:41.963 11:39:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=configuring 00:28:41.963 11:39:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:41.963 11:39:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:41.963 11:39:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:28:41.963 11:39:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:41.963 11:39:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:41.963 11:39:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:41.963 11:39:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:41.963 11:39:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:41.963 11:39:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:42.221 11:39:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:42.221 "name": "raid_bdev1", 00:28:42.221 "uuid": "5877095e-6236-42ff-b2ba-4fb5b7a6add2", 00:28:42.221 "strip_size_kb": 0, 00:28:42.221 "state": "configuring", 00:28:42.221 "raid_level": "raid1", 00:28:42.221 "superblock": true, 00:28:42.221 "num_base_bdevs": 2, 00:28:42.221 "num_base_bdevs_discovered": 1, 00:28:42.221 "num_base_bdevs_operational": 2, 00:28:42.221 "base_bdevs_list": [ 00:28:42.221 { 00:28:42.221 "name": "pt1", 00:28:42.221 "uuid": "00000000-0000-0000-0000-000000000001", 00:28:42.221 "is_configured": true, 00:28:42.221 "data_offset": 256, 00:28:42.221 "data_size": 7936 00:28:42.221 }, 00:28:42.221 { 00:28:42.221 "name": null, 00:28:42.221 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:42.221 "is_configured": false, 00:28:42.221 "data_offset": 256, 00:28:42.221 "data_size": 7936 00:28:42.221 } 00:28:42.221 ] 00:28:42.221 }' 00:28:42.221 11:39:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:42.221 11:39:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:28:42.786 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@469 -- # '[' 2 -gt 2 ']' 00:28:42.786 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@477 -- # (( i = 1 )) 00:28:42.786 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:28:42.786 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:28:42.786 [2024-07-15 11:39:26.377620] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:28:42.786 [2024-07-15 11:39:26.377667] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:42.786 [2024-07-15 11:39:26.377688] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xd48270 00:28:42.786 [2024-07-15 11:39:26.377700] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:42.786 [2024-07-15 11:39:26.377858] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:42.786 [2024-07-15 11:39:26.377874] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:28:42.786 [2024-07-15 11:39:26.377914] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:28:42.786 [2024-07-15 11:39:26.377939] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:28:42.786 [2024-07-15 11:39:26.378023] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xbc1c10 00:28:42.786 [2024-07-15 11:39:26.378033] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:28:42.786 [2024-07-15 11:39:26.378089] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xd43d40 00:28:42.786 [2024-07-15 11:39:26.378162] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xbc1c10 00:28:42.786 [2024-07-15 11:39:26.378171] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0xbc1c10 00:28:42.786 [2024-07-15 11:39:26.378226] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:43.043 pt2 00:28:43.043 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@477 -- # (( i++ )) 00:28:43.043 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@477 -- # (( i < num_base_bdevs )) 00:28:43.043 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@482 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:28:43.043 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:28:43.043 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:28:43.043 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:43.043 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:43.043 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:28:43.043 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:43.043 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:43.043 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:43.043 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:43.043 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:43.043 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:43.301 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:43.301 "name": "raid_bdev1", 00:28:43.301 "uuid": "5877095e-6236-42ff-b2ba-4fb5b7a6add2", 00:28:43.301 "strip_size_kb": 0, 00:28:43.301 "state": "online", 00:28:43.301 "raid_level": "raid1", 00:28:43.301 "superblock": true, 00:28:43.301 "num_base_bdevs": 2, 00:28:43.301 "num_base_bdevs_discovered": 2, 00:28:43.301 "num_base_bdevs_operational": 2, 00:28:43.301 "base_bdevs_list": [ 00:28:43.301 { 00:28:43.301 "name": "pt1", 00:28:43.301 "uuid": "00000000-0000-0000-0000-000000000001", 00:28:43.301 "is_configured": true, 00:28:43.301 "data_offset": 256, 00:28:43.301 "data_size": 7936 00:28:43.301 }, 00:28:43.301 { 00:28:43.301 "name": "pt2", 00:28:43.301 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:43.301 "is_configured": true, 00:28:43.301 "data_offset": 256, 00:28:43.301 "data_size": 7936 00:28:43.301 } 00:28:43.301 ] 00:28:43.301 }' 00:28:43.301 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:43.301 11:39:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:28:43.866 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_properties raid_bdev1 00:28:43.866 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@194 -- # local raid_bdev_name=raid_bdev1 00:28:43.866 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@195 -- # local raid_bdev_info 00:28:43.866 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@196 -- # local base_bdev_info 00:28:43.866 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@197 -- # local base_bdev_names 00:28:43.866 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@198 -- # local name 00:28:43.866 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@200 -- # jq '.[]' 00:28:43.866 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@200 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:28:44.124 [2024-07-15 11:39:27.464734] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:44.124 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@200 -- # raid_bdev_info='{ 00:28:44.124 "name": "raid_bdev1", 00:28:44.124 "aliases": [ 00:28:44.124 "5877095e-6236-42ff-b2ba-4fb5b7a6add2" 00:28:44.124 ], 00:28:44.124 "product_name": "Raid Volume", 00:28:44.124 "block_size": 4128, 00:28:44.124 "num_blocks": 7936, 00:28:44.124 "uuid": "5877095e-6236-42ff-b2ba-4fb5b7a6add2", 00:28:44.124 "md_size": 32, 00:28:44.124 "md_interleave": true, 00:28:44.124 "dif_type": 0, 00:28:44.124 "assigned_rate_limits": { 00:28:44.124 "rw_ios_per_sec": 0, 00:28:44.124 "rw_mbytes_per_sec": 0, 00:28:44.124 "r_mbytes_per_sec": 0, 00:28:44.124 "w_mbytes_per_sec": 0 00:28:44.124 }, 00:28:44.124 "claimed": false, 00:28:44.124 "zoned": false, 00:28:44.124 "supported_io_types": { 00:28:44.124 "read": true, 00:28:44.124 "write": true, 00:28:44.124 "unmap": false, 00:28:44.124 "flush": false, 00:28:44.124 "reset": true, 00:28:44.124 "nvme_admin": false, 00:28:44.124 "nvme_io": false, 00:28:44.124 "nvme_io_md": false, 00:28:44.124 "write_zeroes": true, 00:28:44.124 "zcopy": false, 00:28:44.124 "get_zone_info": false, 00:28:44.124 "zone_management": false, 00:28:44.124 "zone_append": false, 00:28:44.124 "compare": false, 00:28:44.124 "compare_and_write": false, 00:28:44.124 "abort": false, 00:28:44.124 "seek_hole": false, 00:28:44.124 "seek_data": false, 00:28:44.124 "copy": false, 00:28:44.124 "nvme_iov_md": false 00:28:44.124 }, 00:28:44.124 "memory_domains": [ 00:28:44.124 { 00:28:44.124 "dma_device_id": "system", 00:28:44.124 "dma_device_type": 1 00:28:44.124 }, 00:28:44.124 { 00:28:44.124 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:44.124 "dma_device_type": 2 00:28:44.124 }, 00:28:44.124 { 00:28:44.124 "dma_device_id": "system", 00:28:44.124 "dma_device_type": 1 00:28:44.124 }, 00:28:44.124 { 00:28:44.124 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:44.124 "dma_device_type": 2 00:28:44.124 } 00:28:44.124 ], 00:28:44.124 "driver_specific": { 00:28:44.124 "raid": { 00:28:44.125 "uuid": "5877095e-6236-42ff-b2ba-4fb5b7a6add2", 00:28:44.125 "strip_size_kb": 0, 00:28:44.125 "state": "online", 00:28:44.125 "raid_level": "raid1", 00:28:44.125 "superblock": true, 00:28:44.125 "num_base_bdevs": 2, 00:28:44.125 "num_base_bdevs_discovered": 2, 00:28:44.125 "num_base_bdevs_operational": 2, 00:28:44.125 "base_bdevs_list": [ 00:28:44.125 { 00:28:44.125 "name": "pt1", 00:28:44.125 "uuid": "00000000-0000-0000-0000-000000000001", 00:28:44.125 "is_configured": true, 00:28:44.125 "data_offset": 256, 00:28:44.125 "data_size": 7936 00:28:44.125 }, 00:28:44.125 { 00:28:44.125 "name": "pt2", 00:28:44.125 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:44.125 "is_configured": true, 00:28:44.125 "data_offset": 256, 00:28:44.125 "data_size": 7936 00:28:44.125 } 00:28:44.125 ] 00:28:44.125 } 00:28:44.125 } 00:28:44.125 }' 00:28:44.125 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@201 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:28:44.125 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@201 -- # base_bdev_names='pt1 00:28:44.125 pt2' 00:28:44.125 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:28:44.125 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt1 00:28:44.125 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:28:44.382 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:28:44.382 "name": "pt1", 00:28:44.382 "aliases": [ 00:28:44.382 "00000000-0000-0000-0000-000000000001" 00:28:44.382 ], 00:28:44.382 "product_name": "passthru", 00:28:44.382 "block_size": 4128, 00:28:44.382 "num_blocks": 8192, 00:28:44.382 "uuid": "00000000-0000-0000-0000-000000000001", 00:28:44.382 "md_size": 32, 00:28:44.382 "md_interleave": true, 00:28:44.382 "dif_type": 0, 00:28:44.382 "assigned_rate_limits": { 00:28:44.382 "rw_ios_per_sec": 0, 00:28:44.382 "rw_mbytes_per_sec": 0, 00:28:44.382 "r_mbytes_per_sec": 0, 00:28:44.382 "w_mbytes_per_sec": 0 00:28:44.382 }, 00:28:44.382 "claimed": true, 00:28:44.382 "claim_type": "exclusive_write", 00:28:44.382 "zoned": false, 00:28:44.382 "supported_io_types": { 00:28:44.382 "read": true, 00:28:44.382 "write": true, 00:28:44.382 "unmap": true, 00:28:44.382 "flush": true, 00:28:44.382 "reset": true, 00:28:44.382 "nvme_admin": false, 00:28:44.382 "nvme_io": false, 00:28:44.382 "nvme_io_md": false, 00:28:44.382 "write_zeroes": true, 00:28:44.382 "zcopy": true, 00:28:44.382 "get_zone_info": false, 00:28:44.382 "zone_management": false, 00:28:44.382 "zone_append": false, 00:28:44.382 "compare": false, 00:28:44.382 "compare_and_write": false, 00:28:44.382 "abort": true, 00:28:44.382 "seek_hole": false, 00:28:44.382 "seek_data": false, 00:28:44.382 "copy": true, 00:28:44.382 "nvme_iov_md": false 00:28:44.382 }, 00:28:44.382 "memory_domains": [ 00:28:44.382 { 00:28:44.382 "dma_device_id": "system", 00:28:44.382 "dma_device_type": 1 00:28:44.382 }, 00:28:44.382 { 00:28:44.382 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:44.382 "dma_device_type": 2 00:28:44.382 } 00:28:44.382 ], 00:28:44.382 "driver_specific": { 00:28:44.382 "passthru": { 00:28:44.382 "name": "pt1", 00:28:44.382 "base_bdev_name": "malloc1" 00:28:44.382 } 00:28:44.382 } 00:28:44.382 }' 00:28:44.382 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:28:44.382 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:28:44.382 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@205 -- # [[ 4128 == 4128 ]] 00:28:44.382 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:28:44.382 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:28:44.382 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@206 -- # [[ 32 == 32 ]] 00:28:44.382 11:39:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:28:44.640 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:28:44.640 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@207 -- # [[ true == true ]] 00:28:44.640 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:28:44.640 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:28:44.640 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@208 -- # [[ 0 == 0 ]] 00:28:44.640 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@203 -- # for name in $base_bdev_names 00:28:44.640 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@204 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b pt2 00:28:44.640 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@204 -- # jq '.[]' 00:28:44.897 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@204 -- # base_bdev_info='{ 00:28:44.897 "name": "pt2", 00:28:44.897 "aliases": [ 00:28:44.897 "00000000-0000-0000-0000-000000000002" 00:28:44.897 ], 00:28:44.897 "product_name": "passthru", 00:28:44.897 "block_size": 4128, 00:28:44.897 "num_blocks": 8192, 00:28:44.897 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:44.897 "md_size": 32, 00:28:44.897 "md_interleave": true, 00:28:44.897 "dif_type": 0, 00:28:44.897 "assigned_rate_limits": { 00:28:44.897 "rw_ios_per_sec": 0, 00:28:44.897 "rw_mbytes_per_sec": 0, 00:28:44.897 "r_mbytes_per_sec": 0, 00:28:44.897 "w_mbytes_per_sec": 0 00:28:44.897 }, 00:28:44.897 "claimed": true, 00:28:44.897 "claim_type": "exclusive_write", 00:28:44.897 "zoned": false, 00:28:44.897 "supported_io_types": { 00:28:44.898 "read": true, 00:28:44.898 "write": true, 00:28:44.898 "unmap": true, 00:28:44.898 "flush": true, 00:28:44.898 "reset": true, 00:28:44.898 "nvme_admin": false, 00:28:44.898 "nvme_io": false, 00:28:44.898 "nvme_io_md": false, 00:28:44.898 "write_zeroes": true, 00:28:44.898 "zcopy": true, 00:28:44.898 "get_zone_info": false, 00:28:44.898 "zone_management": false, 00:28:44.898 "zone_append": false, 00:28:44.898 "compare": false, 00:28:44.898 "compare_and_write": false, 00:28:44.898 "abort": true, 00:28:44.898 "seek_hole": false, 00:28:44.898 "seek_data": false, 00:28:44.898 "copy": true, 00:28:44.898 "nvme_iov_md": false 00:28:44.898 }, 00:28:44.898 "memory_domains": [ 00:28:44.898 { 00:28:44.898 "dma_device_id": "system", 00:28:44.898 "dma_device_type": 1 00:28:44.898 }, 00:28:44.898 { 00:28:44.898 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:44.898 "dma_device_type": 2 00:28:44.898 } 00:28:44.898 ], 00:28:44.898 "driver_specific": { 00:28:44.898 "passthru": { 00:28:44.898 "name": "pt2", 00:28:44.898 "base_bdev_name": "malloc2" 00:28:44.898 } 00:28:44.898 } 00:28:44.898 }' 00:28:44.898 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:28:44.898 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@205 -- # jq .block_size 00:28:44.898 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@205 -- # [[ 4128 == 4128 ]] 00:28:44.898 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:28:45.156 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@206 -- # jq .md_size 00:28:45.156 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@206 -- # [[ 32 == 32 ]] 00:28:45.156 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:28:45.156 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@207 -- # jq .md_interleave 00:28:45.156 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@207 -- # [[ true == true ]] 00:28:45.156 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:28:45.156 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@208 -- # jq .dif_type 00:28:45.156 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@208 -- # [[ 0 == 0 ]] 00:28:45.156 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@486 -- # jq -r '.[] | .uuid' 00:28:45.156 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@486 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:28:45.414 [2024-07-15 11:39:28.956719] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:45.414 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@486 -- # '[' 5877095e-6236-42ff-b2ba-4fb5b7a6add2 '!=' 5877095e-6236-42ff-b2ba-4fb5b7a6add2 ']' 00:28:45.414 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@490 -- # has_redundancy raid1 00:28:45.414 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@213 -- # case $1 in 00:28:45.414 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@214 -- # return 0 00:28:45.414 11:39:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@492 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt1 00:28:45.673 [2024-07-15 11:39:29.205145] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:28:45.673 11:39:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@495 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:28:45.673 11:39:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:28:45.673 11:39:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:28:45.673 11:39:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:45.673 11:39:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:45.673 11:39:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:28:45.673 11:39:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:45.673 11:39:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:45.673 11:39:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:45.673 11:39:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:45.673 11:39:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:45.673 11:39:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:45.931 11:39:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:45.931 "name": "raid_bdev1", 00:28:45.931 "uuid": "5877095e-6236-42ff-b2ba-4fb5b7a6add2", 00:28:45.931 "strip_size_kb": 0, 00:28:45.931 "state": "online", 00:28:45.931 "raid_level": "raid1", 00:28:45.931 "superblock": true, 00:28:45.931 "num_base_bdevs": 2, 00:28:45.931 "num_base_bdevs_discovered": 1, 00:28:45.931 "num_base_bdevs_operational": 1, 00:28:45.931 "base_bdevs_list": [ 00:28:45.931 { 00:28:45.931 "name": null, 00:28:45.931 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:45.931 "is_configured": false, 00:28:45.931 "data_offset": 256, 00:28:45.931 "data_size": 7936 00:28:45.931 }, 00:28:45.931 { 00:28:45.931 "name": "pt2", 00:28:45.931 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:45.931 "is_configured": true, 00:28:45.931 "data_offset": 256, 00:28:45.931 "data_size": 7936 00:28:45.931 } 00:28:45.931 ] 00:28:45.931 }' 00:28:45.931 11:39:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:45.931 11:39:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:28:46.499 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@498 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:28:46.758 [2024-07-15 11:39:30.304068] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:28:46.758 [2024-07-15 11:39:30.304096] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:46.758 [2024-07-15 11:39:30.304153] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:46.758 [2024-07-15 11:39:30.304198] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:46.758 [2024-07-15 11:39:30.304210] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xbc1c10 name raid_bdev1, state offline 00:28:46.758 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@499 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:46.758 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@499 -- # jq -r '.[]' 00:28:47.017 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@499 -- # raid_bdev= 00:28:47.017 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # '[' -n '' ']' 00:28:47.017 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@505 -- # (( i = 1 )) 00:28:47.017 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@505 -- # (( i < num_base_bdevs )) 00:28:47.017 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete pt2 00:28:47.275 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@505 -- # (( i++ )) 00:28:47.275 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@505 -- # (( i < num_base_bdevs )) 00:28:47.275 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@510 -- # (( i = 1 )) 00:28:47.275 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@510 -- # (( i < num_base_bdevs - 1 )) 00:28:47.275 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@518 -- # i=1 00:28:47.275 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@519 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:28:47.533 [2024-07-15 11:39:30.949739] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:28:47.533 [2024-07-15 11:39:30.949788] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:47.533 [2024-07-15 11:39:30.949808] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xd469f0 00:28:47.533 [2024-07-15 11:39:30.949820] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:47.533 [2024-07-15 11:39:30.951277] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:47.533 [2024-07-15 11:39:30.951303] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:28:47.533 [2024-07-15 11:39:30.951351] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:28:47.533 [2024-07-15 11:39:30.951378] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:28:47.533 [2024-07-15 11:39:30.951450] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xd47ea0 00:28:47.533 [2024-07-15 11:39:30.951460] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:28:47.533 [2024-07-15 11:39:30.951520] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xd45bc0 00:28:47.533 [2024-07-15 11:39:30.951592] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xd47ea0 00:28:47.533 [2024-07-15 11:39:30.951601] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0xd47ea0 00:28:47.533 [2024-07-15 11:39:30.951655] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:47.533 pt2 00:28:47.533 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@522 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:28:47.533 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:28:47.533 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:28:47.533 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:47.533 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:47.533 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:28:47.533 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:47.533 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:47.533 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:47.533 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:47.533 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:47.533 11:39:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:47.791 11:39:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:47.791 "name": "raid_bdev1", 00:28:47.791 "uuid": "5877095e-6236-42ff-b2ba-4fb5b7a6add2", 00:28:47.791 "strip_size_kb": 0, 00:28:47.791 "state": "online", 00:28:47.791 "raid_level": "raid1", 00:28:47.791 "superblock": true, 00:28:47.791 "num_base_bdevs": 2, 00:28:47.791 "num_base_bdevs_discovered": 1, 00:28:47.791 "num_base_bdevs_operational": 1, 00:28:47.791 "base_bdevs_list": [ 00:28:47.791 { 00:28:47.791 "name": null, 00:28:47.791 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:47.791 "is_configured": false, 00:28:47.791 "data_offset": 256, 00:28:47.791 "data_size": 7936 00:28:47.791 }, 00:28:47.791 { 00:28:47.791 "name": "pt2", 00:28:47.791 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:47.791 "is_configured": true, 00:28:47.791 "data_offset": 256, 00:28:47.791 "data_size": 7936 00:28:47.791 } 00:28:47.791 ] 00:28:47.791 }' 00:28:47.791 11:39:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:47.791 11:39:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:28:48.357 11:39:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@525 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:28:48.357 [2024-07-15 11:39:31.936339] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:28:48.357 [2024-07-15 11:39:31.936365] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:48.357 [2024-07-15 11:39:31.936415] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:48.357 [2024-07-15 11:39:31.936457] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:48.357 [2024-07-15 11:39:31.936469] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xd47ea0 name raid_bdev1, state offline 00:28:48.615 11:39:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@526 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:48.615 11:39:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@526 -- # jq -r '.[]' 00:28:48.615 11:39:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@526 -- # raid_bdev= 00:28:48.615 11:39:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # '[' -n '' ']' 00:28:48.615 11:39:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@531 -- # '[' 2 -gt 2 ']' 00:28:48.615 11:39:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@539 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:28:48.874 [2024-07-15 11:39:32.345416] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:28:48.874 [2024-07-15 11:39:32.345460] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:48.874 [2024-07-15 11:39:32.345479] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xd46620 00:28:48.874 [2024-07-15 11:39:32.345491] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:48.874 [2024-07-15 11:39:32.346951] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:48.874 [2024-07-15 11:39:32.346977] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:28:48.874 [2024-07-15 11:39:32.347025] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:28:48.874 [2024-07-15 11:39:32.347049] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:28:48.874 [2024-07-15 11:39:32.347129] bdev_raid.c:3547:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:28:48.874 [2024-07-15 11:39:32.347141] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:28:48.874 [2024-07-15 11:39:32.347155] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xd48640 name raid_bdev1, state configuring 00:28:48.874 [2024-07-15 11:39:32.347179] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:28:48.874 [2024-07-15 11:39:32.347232] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0xd48640 00:28:48.874 [2024-07-15 11:39:32.347249] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:28:48.874 [2024-07-15 11:39:32.347303] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0xd47810 00:28:48.874 [2024-07-15 11:39:32.347375] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0xd48640 00:28:48.874 [2024-07-15 11:39:32.347384] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0xd48640 00:28:48.874 [2024-07-15 11:39:32.347443] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:48.874 pt1 00:28:48.874 11:39:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@541 -- # '[' 2 -gt 2 ']' 00:28:48.874 11:39:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@553 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:28:48.874 11:39:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:28:48.874 11:39:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:28:48.874 11:39:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:48.874 11:39:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:48.874 11:39:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:28:48.874 11:39:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:48.874 11:39:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:48.874 11:39:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:48.874 11:39:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:48.874 11:39:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:48.874 11:39:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:49.440 11:39:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:49.440 "name": "raid_bdev1", 00:28:49.440 "uuid": "5877095e-6236-42ff-b2ba-4fb5b7a6add2", 00:28:49.440 "strip_size_kb": 0, 00:28:49.440 "state": "online", 00:28:49.440 "raid_level": "raid1", 00:28:49.440 "superblock": true, 00:28:49.440 "num_base_bdevs": 2, 00:28:49.440 "num_base_bdevs_discovered": 1, 00:28:49.440 "num_base_bdevs_operational": 1, 00:28:49.440 "base_bdevs_list": [ 00:28:49.440 { 00:28:49.440 "name": null, 00:28:49.440 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:49.440 "is_configured": false, 00:28:49.440 "data_offset": 256, 00:28:49.440 "data_size": 7936 00:28:49.440 }, 00:28:49.440 { 00:28:49.440 "name": "pt2", 00:28:49.440 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:49.440 "is_configured": true, 00:28:49.440 "data_offset": 256, 00:28:49.440 "data_size": 7936 00:28:49.440 } 00:28:49.440 ] 00:28:49.440 }' 00:28:49.440 11:39:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:49.440 11:39:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:28:50.407 11:39:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@554 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs online 00:28:50.407 11:39:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@554 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:28:50.407 11:39:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@554 -- # [[ false == \f\a\l\s\e ]] 00:28:50.407 11:39:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@557 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:28:50.407 11:39:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@557 -- # jq -r '.[] | .uuid' 00:28:50.664 [2024-07-15 11:39:34.194536] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:50.664 11:39:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@557 -- # '[' 5877095e-6236-42ff-b2ba-4fb5b7a6add2 '!=' 5877095e-6236-42ff-b2ba-4fb5b7a6add2 ']' 00:28:50.664 11:39:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@562 -- # killprocess 1019539 00:28:50.664 11:39:34 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@948 -- # '[' -z 1019539 ']' 00:28:50.664 11:39:34 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@952 -- # kill -0 1019539 00:28:50.664 11:39:34 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@953 -- # uname 00:28:50.664 11:39:34 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:28:50.664 11:39:34 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1019539 00:28:50.922 11:39:34 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:28:50.922 11:39:34 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:28:50.922 11:39:34 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1019539' 00:28:50.922 killing process with pid 1019539 00:28:50.922 11:39:34 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@967 -- # kill 1019539 00:28:50.922 [2024-07-15 11:39:34.275027] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:28:50.922 [2024-07-15 11:39:34.275082] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:50.922 [2024-07-15 11:39:34.275127] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:50.922 [2024-07-15 11:39:34.275140] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0xd48640 name raid_bdev1, state offline 00:28:50.922 11:39:34 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@972 -- # wait 1019539 00:28:50.923 [2024-07-15 11:39:34.293454] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:28:50.923 11:39:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@564 -- # return 0 00:28:50.923 00:28:50.923 real 0m15.770s 00:28:50.923 user 0m28.641s 00:28:50.923 sys 0m2.836s 00:28:50.923 11:39:34 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1124 -- # xtrace_disable 00:28:50.923 11:39:34 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:28:50.923 ************************************ 00:28:50.923 END TEST raid_superblock_test_md_interleaved 00:28:50.923 ************************************ 00:28:51.182 11:39:34 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:28:51.182 11:39:34 bdev_raid -- bdev/bdev_raid.sh@914 -- # run_test raid_rebuild_test_sb_md_interleaved raid_rebuild_test raid1 2 true false false 00:28:51.182 11:39:34 bdev_raid -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:28:51.182 11:39:34 bdev_raid -- common/autotest_common.sh@1105 -- # xtrace_disable 00:28:51.182 11:39:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:28:51.182 ************************************ 00:28:51.182 START TEST raid_rebuild_test_sb_md_interleaved 00:28:51.182 ************************************ 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1123 -- # raid_rebuild_test raid1 2 true false false 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@568 -- # local raid_level=raid1 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@569 -- # local num_base_bdevs=2 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@570 -- # local superblock=true 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@571 -- # local background_io=false 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@572 -- # local verify=false 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@573 -- # (( i = 1 )) 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@575 -- # echo BaseBdev1 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@575 -- # echo BaseBdev2 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@573 -- # (( i++ )) 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@573 -- # (( i <= num_base_bdevs )) 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@573 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@573 -- # local base_bdevs 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # local raid_bdev_name=raid_bdev1 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@575 -- # local strip_size 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # local create_arg 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@577 -- # local raid_bdev_size 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@578 -- # local data_offset 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@580 -- # '[' raid1 '!=' raid1 ']' 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@588 -- # strip_size=0 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@591 -- # '[' true = true ']' 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@592 -- # create_arg+=' -s' 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@596 -- # raid_pid=1021952 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@597 -- # waitforlisten 1021952 /var/tmp/spdk-raid.sock 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@829 -- # '[' -z 1021952 ']' 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@595 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/spdk-raid.sock -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-raid.sock 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@834 -- # local max_retries=100 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock...' 00:28:51.182 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-raid.sock... 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@838 -- # xtrace_disable 00:28:51.182 11:39:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:28:51.182 [2024-07-15 11:39:34.666401] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:28:51.182 [2024-07-15 11:39:34.666467] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1021952 ] 00:28:51.182 I/O size of 3145728 is greater than zero copy threshold (65536). 00:28:51.182 Zero copy mechanism will not be used. 00:28:51.440 [2024-07-15 11:39:34.789383] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:51.440 [2024-07-15 11:39:34.896363] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:28:51.440 [2024-07-15 11:39:34.970633] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:51.440 [2024-07-15 11:39:34.970673] bdev_raid.c:1416:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:52.006 11:39:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:28:52.006 11:39:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@862 -- # return 0 00:28:52.006 11:39:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:28:52.006 11:39:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1_malloc 00:28:52.264 BaseBdev1_malloc 00:28:52.264 11:39:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:28:52.522 [2024-07-15 11:39:36.048954] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:28:52.522 [2024-07-15 11:39:36.049002] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:52.522 [2024-07-15 11:39:36.049027] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1e1ece0 00:28:52.522 [2024-07-15 11:39:36.049040] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:52.522 [2024-07-15 11:39:36.050600] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:52.522 [2024-07-15 11:39:36.050627] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:28:52.522 BaseBdev1 00:28:52.522 11:39:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@600 -- # for bdev in "${base_bdevs[@]}" 00:28:52.522 11:39:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2_malloc 00:28:52.781 BaseBdev2_malloc 00:28:52.781 11:39:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:28:53.039 [2024-07-15 11:39:36.452419] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:28:53.039 [2024-07-15 11:39:36.452464] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:53.039 [2024-07-15 11:39:36.452488] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1e162d0 00:28:53.039 [2024-07-15 11:39:36.452501] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:53.039 [2024-07-15 11:39:36.454279] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:53.039 [2024-07-15 11:39:36.454305] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:28:53.039 BaseBdev2 00:28:53.039 11:39:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@606 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_malloc_create 32 4096 -m 32 -i -b spare_malloc 00:28:53.297 spare_malloc 00:28:53.297 11:39:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@607 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:28:53.555 spare_delay 00:28:53.555 11:39:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@608 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:28:54.121 [2024-07-15 11:39:37.437111] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:28:54.121 [2024-07-15 11:39:37.437155] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:54.121 [2024-07-15 11:39:37.437177] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1e19070 00:28:54.121 [2024-07-15 11:39:37.437190] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:54.121 [2024-07-15 11:39:37.438580] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:54.121 [2024-07-15 11:39:37.438607] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:28:54.121 spare 00:28:54.121 11:39:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@611 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_create -s -r raid1 -b 'BaseBdev1 BaseBdev2' -n raid_bdev1 00:28:54.121 [2024-07-15 11:39:37.693820] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:54.121 [2024-07-15 11:39:37.695142] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:54.121 [2024-07-15 11:39:37.695307] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1e1b370 00:28:54.121 [2024-07-15 11:39:37.695320] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:28:54.121 [2024-07-15 11:39:37.695395] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1c819c0 00:28:54.121 [2024-07-15 11:39:37.695481] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1e1b370 00:28:54.121 [2024-07-15 11:39:37.695491] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x1e1b370 00:28:54.121 [2024-07-15 11:39:37.695548] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:54.121 11:39:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@612 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:28:54.121 11:39:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:28:54.121 11:39:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:28:54.121 11:39:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:54.121 11:39:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:54.379 11:39:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:28:54.379 11:39:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:54.379 11:39:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:54.379 11:39:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:54.379 11:39:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:54.379 11:39:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:54.379 11:39:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:54.945 11:39:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:54.945 "name": "raid_bdev1", 00:28:54.945 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:28:54.945 "strip_size_kb": 0, 00:28:54.945 "state": "online", 00:28:54.945 "raid_level": "raid1", 00:28:54.945 "superblock": true, 00:28:54.945 "num_base_bdevs": 2, 00:28:54.945 "num_base_bdevs_discovered": 2, 00:28:54.945 "num_base_bdevs_operational": 2, 00:28:54.945 "base_bdevs_list": [ 00:28:54.945 { 00:28:54.945 "name": "BaseBdev1", 00:28:54.945 "uuid": "3c300e44-3582-5097-a776-538ba50edf8b", 00:28:54.945 "is_configured": true, 00:28:54.945 "data_offset": 256, 00:28:54.945 "data_size": 7936 00:28:54.945 }, 00:28:54.945 { 00:28:54.945 "name": "BaseBdev2", 00:28:54.945 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:28:54.945 "is_configured": true, 00:28:54.945 "data_offset": 256, 00:28:54.945 "data_size": 7936 00:28:54.945 } 00:28:54.945 ] 00:28:54.945 }' 00:28:54.945 11:39:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:54.945 11:39:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:28:55.512 11:39:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@615 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_get_bdevs -b raid_bdev1 00:28:55.512 11:39:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@615 -- # jq -r '.[].num_blocks' 00:28:55.512 [2024-07-15 11:39:39.049632] bdev_raid.c:1107:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:55.512 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@615 -- # raid_bdev_size=7936 00:28:55.512 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@618 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:55.512 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@618 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:28:55.770 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@618 -- # data_offset=256 00:28:55.770 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@620 -- # '[' false = true ']' 00:28:55.770 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@623 -- # '[' false = true ']' 00:28:55.770 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@639 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev BaseBdev1 00:28:56.028 [2024-07-15 11:39:39.474496] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:28:56.028 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@642 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:28:56.028 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:28:56.028 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:28:56.028 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:56.028 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:56.028 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:28:56.028 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:56.028 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:56.028 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:56.028 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:56.028 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:56.028 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:56.287 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:56.287 "name": "raid_bdev1", 00:28:56.287 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:28:56.287 "strip_size_kb": 0, 00:28:56.287 "state": "online", 00:28:56.287 "raid_level": "raid1", 00:28:56.287 "superblock": true, 00:28:56.287 "num_base_bdevs": 2, 00:28:56.287 "num_base_bdevs_discovered": 1, 00:28:56.287 "num_base_bdevs_operational": 1, 00:28:56.287 "base_bdevs_list": [ 00:28:56.287 { 00:28:56.287 "name": null, 00:28:56.287 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:56.287 "is_configured": false, 00:28:56.287 "data_offset": 256, 00:28:56.287 "data_size": 7936 00:28:56.287 }, 00:28:56.287 { 00:28:56.287 "name": "BaseBdev2", 00:28:56.287 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:28:56.287 "is_configured": true, 00:28:56.287 "data_offset": 256, 00:28:56.287 "data_size": 7936 00:28:56.287 } 00:28:56.287 ] 00:28:56.287 }' 00:28:56.287 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:56.287 11:39:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:28:56.854 11:39:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@645 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:28:56.854 [2024-07-15 11:39:40.445081] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:28:57.112 [2024-07-15 11:39:40.448700] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1e1b250 00:28:57.112 [2024-07-15 11:39:40.450695] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:28:57.112 11:39:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@646 -- # sleep 1 00:28:58.043 11:39:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@649 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:58.043 11:39:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:28:58.043 11:39:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:28:58.043 11:39:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local target=spare 00:28:58.043 11:39:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:28:58.043 11:39:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:58.043 11:39:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:58.300 11:39:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:58.300 "name": "raid_bdev1", 00:28:58.300 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:28:58.300 "strip_size_kb": 0, 00:28:58.300 "state": "online", 00:28:58.300 "raid_level": "raid1", 00:28:58.300 "superblock": true, 00:28:58.300 "num_base_bdevs": 2, 00:28:58.300 "num_base_bdevs_discovered": 2, 00:28:58.300 "num_base_bdevs_operational": 2, 00:28:58.300 "process": { 00:28:58.300 "type": "rebuild", 00:28:58.300 "target": "spare", 00:28:58.300 "progress": { 00:28:58.300 "blocks": 3072, 00:28:58.300 "percent": 38 00:28:58.300 } 00:28:58.300 }, 00:28:58.300 "base_bdevs_list": [ 00:28:58.300 { 00:28:58.300 "name": "spare", 00:28:58.300 "uuid": "db8a8c81-e619-5cec-bd99-d2b5802226a5", 00:28:58.300 "is_configured": true, 00:28:58.300 "data_offset": 256, 00:28:58.300 "data_size": 7936 00:28:58.300 }, 00:28:58.300 { 00:28:58.300 "name": "BaseBdev2", 00:28:58.300 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:28:58.300 "is_configured": true, 00:28:58.300 "data_offset": 256, 00:28:58.300 "data_size": 7936 00:28:58.300 } 00:28:58.300 ] 00:28:58.300 }' 00:28:58.300 11:39:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:28:58.300 11:39:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:58.300 11:39:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:28:58.300 11:39:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:28:58.300 11:39:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@652 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev spare 00:28:58.557 [2024-07-15 11:39:42.043543] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:58.557 [2024-07-15 11:39:42.063146] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:28:58.557 [2024-07-15 11:39:42.063193] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:58.557 [2024-07-15 11:39:42.063209] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:58.557 [2024-07-15 11:39:42.063218] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:28:58.557 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@655 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:28:58.557 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:28:58.557 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:28:58.558 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:28:58.558 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:28:58.558 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:28:58.558 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:28:58.558 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:28:58.558 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:28:58.558 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:28:58.558 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:58.558 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:58.815 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:28:58.815 "name": "raid_bdev1", 00:28:58.815 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:28:58.815 "strip_size_kb": 0, 00:28:58.815 "state": "online", 00:28:58.815 "raid_level": "raid1", 00:28:58.815 "superblock": true, 00:28:58.815 "num_base_bdevs": 2, 00:28:58.815 "num_base_bdevs_discovered": 1, 00:28:58.815 "num_base_bdevs_operational": 1, 00:28:58.815 "base_bdevs_list": [ 00:28:58.815 { 00:28:58.815 "name": null, 00:28:58.815 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:58.815 "is_configured": false, 00:28:58.815 "data_offset": 256, 00:28:58.815 "data_size": 7936 00:28:58.815 }, 00:28:58.815 { 00:28:58.815 "name": "BaseBdev2", 00:28:58.815 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:28:58.815 "is_configured": true, 00:28:58.815 "data_offset": 256, 00:28:58.815 "data_size": 7936 00:28:58.815 } 00:28:58.815 ] 00:28:58.815 }' 00:28:58.815 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:28:58.815 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:28:59.378 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@658 -- # verify_raid_bdev_process raid_bdev1 none none 00:28:59.378 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:28:59.378 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:28:59.378 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local target=none 00:28:59.378 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:28:59.378 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:28:59.378 11:39:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:59.635 11:39:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:59.635 "name": "raid_bdev1", 00:28:59.635 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:28:59.635 "strip_size_kb": 0, 00:28:59.635 "state": "online", 00:28:59.635 "raid_level": "raid1", 00:28:59.635 "superblock": true, 00:28:59.635 "num_base_bdevs": 2, 00:28:59.635 "num_base_bdevs_discovered": 1, 00:28:59.635 "num_base_bdevs_operational": 1, 00:28:59.635 "base_bdevs_list": [ 00:28:59.635 { 00:28:59.635 "name": null, 00:28:59.635 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:59.635 "is_configured": false, 00:28:59.635 "data_offset": 256, 00:28:59.636 "data_size": 7936 00:28:59.636 }, 00:28:59.636 { 00:28:59.636 "name": "BaseBdev2", 00:28:59.636 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:28:59.636 "is_configured": true, 00:28:59.636 "data_offset": 256, 00:28:59.636 "data_size": 7936 00:28:59.636 } 00:28:59.636 ] 00:28:59.636 }' 00:28:59.636 11:39:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:28:59.892 11:39:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:28:59.892 11:39:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:28:59.892 11:39:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:28:59.892 11:39:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@661 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:28:59.892 [2024-07-15 11:39:43.435235] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:28:59.892 [2024-07-15 11:39:43.438843] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1e17270 00:28:59.892 [2024-07-15 11:39:43.440279] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:28:59.892 11:39:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@662 -- # sleep 1 00:29:01.265 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@663 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local target=spare 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:01.266 "name": "raid_bdev1", 00:29:01.266 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:29:01.266 "strip_size_kb": 0, 00:29:01.266 "state": "online", 00:29:01.266 "raid_level": "raid1", 00:29:01.266 "superblock": true, 00:29:01.266 "num_base_bdevs": 2, 00:29:01.266 "num_base_bdevs_discovered": 2, 00:29:01.266 "num_base_bdevs_operational": 2, 00:29:01.266 "process": { 00:29:01.266 "type": "rebuild", 00:29:01.266 "target": "spare", 00:29:01.266 "progress": { 00:29:01.266 "blocks": 3072, 00:29:01.266 "percent": 38 00:29:01.266 } 00:29:01.266 }, 00:29:01.266 "base_bdevs_list": [ 00:29:01.266 { 00:29:01.266 "name": "spare", 00:29:01.266 "uuid": "db8a8c81-e619-5cec-bd99-d2b5802226a5", 00:29:01.266 "is_configured": true, 00:29:01.266 "data_offset": 256, 00:29:01.266 "data_size": 7936 00:29:01.266 }, 00:29:01.266 { 00:29:01.266 "name": "BaseBdev2", 00:29:01.266 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:29:01.266 "is_configured": true, 00:29:01.266 "data_offset": 256, 00:29:01.266 "data_size": 7936 00:29:01.266 } 00:29:01.266 ] 00:29:01.266 }' 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@665 -- # '[' true = true ']' 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@665 -- # '[' = false ']' 00:29:01.266 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev_raid.sh: line 665: [: =: unary operator expected 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@690 -- # local num_base_bdevs_operational=2 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@692 -- # '[' raid1 = raid1 ']' 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@692 -- # '[' 2 -gt 2 ']' 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@705 -- # local timeout=1112 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local target=spare 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:29:01.266 11:39:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:01.525 11:39:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:01.525 "name": "raid_bdev1", 00:29:01.525 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:29:01.525 "strip_size_kb": 0, 00:29:01.525 "state": "online", 00:29:01.525 "raid_level": "raid1", 00:29:01.525 "superblock": true, 00:29:01.525 "num_base_bdevs": 2, 00:29:01.525 "num_base_bdevs_discovered": 2, 00:29:01.525 "num_base_bdevs_operational": 2, 00:29:01.525 "process": { 00:29:01.525 "type": "rebuild", 00:29:01.525 "target": "spare", 00:29:01.525 "progress": { 00:29:01.525 "blocks": 3840, 00:29:01.525 "percent": 48 00:29:01.525 } 00:29:01.525 }, 00:29:01.525 "base_bdevs_list": [ 00:29:01.525 { 00:29:01.525 "name": "spare", 00:29:01.525 "uuid": "db8a8c81-e619-5cec-bd99-d2b5802226a5", 00:29:01.525 "is_configured": true, 00:29:01.525 "data_offset": 256, 00:29:01.525 "data_size": 7936 00:29:01.525 }, 00:29:01.525 { 00:29:01.525 "name": "BaseBdev2", 00:29:01.525 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:29:01.525 "is_configured": true, 00:29:01.525 "data_offset": 256, 00:29:01.525 "data_size": 7936 00:29:01.525 } 00:29:01.525 ] 00:29:01.525 }' 00:29:01.525 11:39:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:29:01.525 11:39:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:29:01.525 11:39:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:29:01.783 11:39:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:29:01.783 11:39:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@710 -- # sleep 1 00:29:02.729 11:39:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:29:02.729 11:39:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:29:02.729 11:39:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:29:02.729 11:39:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:29:02.729 11:39:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local target=spare 00:29:02.729 11:39:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:29:02.729 11:39:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:29:02.729 11:39:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:02.989 11:39:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:02.989 "name": "raid_bdev1", 00:29:02.989 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:29:02.989 "strip_size_kb": 0, 00:29:02.989 "state": "online", 00:29:02.989 "raid_level": "raid1", 00:29:02.989 "superblock": true, 00:29:02.989 "num_base_bdevs": 2, 00:29:02.989 "num_base_bdevs_discovered": 2, 00:29:02.989 "num_base_bdevs_operational": 2, 00:29:02.989 "process": { 00:29:02.989 "type": "rebuild", 00:29:02.989 "target": "spare", 00:29:02.989 "progress": { 00:29:02.989 "blocks": 7424, 00:29:02.989 "percent": 93 00:29:02.989 } 00:29:02.989 }, 00:29:02.989 "base_bdevs_list": [ 00:29:02.989 { 00:29:02.989 "name": "spare", 00:29:02.989 "uuid": "db8a8c81-e619-5cec-bd99-d2b5802226a5", 00:29:02.989 "is_configured": true, 00:29:02.989 "data_offset": 256, 00:29:02.989 "data_size": 7936 00:29:02.989 }, 00:29:02.989 { 00:29:02.989 "name": "BaseBdev2", 00:29:02.989 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:29:02.989 "is_configured": true, 00:29:02.989 "data_offset": 256, 00:29:02.989 "data_size": 7936 00:29:02.989 } 00:29:02.989 ] 00:29:02.989 }' 00:29:02.989 11:39:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:29:02.989 11:39:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:29:02.989 11:39:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:29:02.989 11:39:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:29:02.989 11:39:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@710 -- # sleep 1 00:29:02.989 [2024-07-15 11:39:46.564464] bdev_raid.c:2789:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:29:02.989 [2024-07-15 11:39:46.564522] bdev_raid.c:2504:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:29:02.989 [2024-07-15 11:39:46.564610] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:03.923 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@706 -- # (( SECONDS < timeout )) 00:29:03.923 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:29:03.923 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:29:03.923 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:29:03.923 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local target=spare 00:29:03.923 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:29:03.923 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:29:03.923 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:04.181 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:04.181 "name": "raid_bdev1", 00:29:04.181 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:29:04.181 "strip_size_kb": 0, 00:29:04.181 "state": "online", 00:29:04.181 "raid_level": "raid1", 00:29:04.181 "superblock": true, 00:29:04.181 "num_base_bdevs": 2, 00:29:04.181 "num_base_bdevs_discovered": 2, 00:29:04.181 "num_base_bdevs_operational": 2, 00:29:04.181 "base_bdevs_list": [ 00:29:04.181 { 00:29:04.181 "name": "spare", 00:29:04.181 "uuid": "db8a8c81-e619-5cec-bd99-d2b5802226a5", 00:29:04.181 "is_configured": true, 00:29:04.181 "data_offset": 256, 00:29:04.181 "data_size": 7936 00:29:04.181 }, 00:29:04.181 { 00:29:04.181 "name": "BaseBdev2", 00:29:04.181 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:29:04.181 "is_configured": true, 00:29:04.181 "data_offset": 256, 00:29:04.181 "data_size": 7936 00:29:04.181 } 00:29:04.181 ] 00:29:04.181 }' 00:29:04.181 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:29:04.181 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # [[ none == \r\e\b\u\i\l\d ]] 00:29:04.181 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:29:04.440 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # [[ none == \s\p\a\r\e ]] 00:29:04.440 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # break 00:29:04.440 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@714 -- # verify_raid_bdev_process raid_bdev1 none none 00:29:04.440 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:29:04.440 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:29:04.440 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local target=none 00:29:04.440 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:29:04.440 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:29:04.440 11:39:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:04.698 11:39:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:04.698 "name": "raid_bdev1", 00:29:04.698 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:29:04.698 "strip_size_kb": 0, 00:29:04.698 "state": "online", 00:29:04.698 "raid_level": "raid1", 00:29:04.698 "superblock": true, 00:29:04.698 "num_base_bdevs": 2, 00:29:04.698 "num_base_bdevs_discovered": 2, 00:29:04.698 "num_base_bdevs_operational": 2, 00:29:04.698 "base_bdevs_list": [ 00:29:04.698 { 00:29:04.698 "name": "spare", 00:29:04.698 "uuid": "db8a8c81-e619-5cec-bd99-d2b5802226a5", 00:29:04.698 "is_configured": true, 00:29:04.698 "data_offset": 256, 00:29:04.698 "data_size": 7936 00:29:04.698 }, 00:29:04.698 { 00:29:04.698 "name": "BaseBdev2", 00:29:04.698 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:29:04.698 "is_configured": true, 00:29:04.698 "data_offset": 256, 00:29:04.698 "data_size": 7936 00:29:04.698 } 00:29:04.698 ] 00:29:04.698 }' 00:29:04.698 11:39:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:29:04.698 11:39:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:29:04.698 11:39:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:29:04.698 11:39:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:29:04.698 11:39:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:29:04.698 11:39:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:29:04.698 11:39:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:29:04.699 11:39:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:29:04.699 11:39:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:29:04.699 11:39:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:29:04.699 11:39:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:29:04.699 11:39:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:29:04.699 11:39:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:29:04.699 11:39:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:29:04.699 11:39:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:29:04.699 11:39:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:04.957 11:39:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:29:04.957 "name": "raid_bdev1", 00:29:04.957 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:29:04.957 "strip_size_kb": 0, 00:29:04.957 "state": "online", 00:29:04.957 "raid_level": "raid1", 00:29:04.957 "superblock": true, 00:29:04.957 "num_base_bdevs": 2, 00:29:04.957 "num_base_bdevs_discovered": 2, 00:29:04.957 "num_base_bdevs_operational": 2, 00:29:04.957 "base_bdevs_list": [ 00:29:04.957 { 00:29:04.957 "name": "spare", 00:29:04.957 "uuid": "db8a8c81-e619-5cec-bd99-d2b5802226a5", 00:29:04.957 "is_configured": true, 00:29:04.957 "data_offset": 256, 00:29:04.957 "data_size": 7936 00:29:04.957 }, 00:29:04.957 { 00:29:04.957 "name": "BaseBdev2", 00:29:04.957 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:29:04.957 "is_configured": true, 00:29:04.957 "data_offset": 256, 00:29:04.957 "data_size": 7936 00:29:04.957 } 00:29:04.957 ] 00:29:04.957 }' 00:29:04.957 11:39:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:29:04.957 11:39:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:29:05.587 11:39:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@718 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_delete raid_bdev1 00:29:05.846 [2024-07-15 11:39:49.244283] bdev_raid.c:2356:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:29:05.846 [2024-07-15 11:39:49.244310] bdev_raid.c:1844:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:05.846 [2024-07-15 11:39:49.244365] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:05.846 [2024-07-15 11:39:49.244421] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:05.846 [2024-07-15 11:39:49.244433] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1e1b370 name raid_bdev1, state offline 00:29:05.846 11:39:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@719 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:29:05.846 11:39:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@719 -- # jq length 00:29:06.104 11:39:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@719 -- # [[ 0 == 0 ]] 00:29:06.104 11:39:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@721 -- # '[' false = true ']' 00:29:06.104 11:39:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@742 -- # '[' true = true ']' 00:29:06.104 11:39:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@744 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:29:06.363 11:39:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@745 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:29:06.363 [2024-07-15 11:39:49.926044] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:29:06.363 [2024-07-15 11:39:49.926097] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:06.363 [2024-07-15 11:39:49.926118] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1e1b040 00:29:06.363 [2024-07-15 11:39:49.926130] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:06.363 [2024-07-15 11:39:49.927624] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:06.363 [2024-07-15 11:39:49.927651] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:29:06.363 [2024-07-15 11:39:49.927709] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:29:06.363 [2024-07-15 11:39:49.927737] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:29:06.363 [2024-07-15 11:39:49.927825] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:06.363 spare 00:29:06.363 11:39:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@747 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:29:06.363 11:39:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:29:06.363 11:39:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:29:06.363 11:39:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:29:06.363 11:39:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:29:06.363 11:39:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=2 00:29:06.363 11:39:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:29:06.363 11:39:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:29:06.363 11:39:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:29:06.363 11:39:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:29:06.363 11:39:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:29:06.363 11:39:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:06.622 [2024-07-15 11:39:50.028139] bdev_raid.c:1694:raid_bdev_configure_cont: *DEBUG*: io device register 0x1e1bf60 00:29:06.622 [2024-07-15 11:39:50.028163] bdev_raid.c:1695:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:29:06.622 [2024-07-15 11:39:50.028249] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1e1bde0 00:29:06.622 [2024-07-15 11:39:50.028353] bdev_raid.c:1724:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x1e1bf60 00:29:06.622 [2024-07-15 11:39:50.028363] bdev_raid.c:1725:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x1e1bf60 00:29:06.622 [2024-07-15 11:39:50.028434] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:06.622 11:39:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:29:06.622 "name": "raid_bdev1", 00:29:06.622 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:29:06.622 "strip_size_kb": 0, 00:29:06.622 "state": "online", 00:29:06.622 "raid_level": "raid1", 00:29:06.622 "superblock": true, 00:29:06.622 "num_base_bdevs": 2, 00:29:06.622 "num_base_bdevs_discovered": 2, 00:29:06.622 "num_base_bdevs_operational": 2, 00:29:06.622 "base_bdevs_list": [ 00:29:06.622 { 00:29:06.622 "name": "spare", 00:29:06.622 "uuid": "db8a8c81-e619-5cec-bd99-d2b5802226a5", 00:29:06.622 "is_configured": true, 00:29:06.622 "data_offset": 256, 00:29:06.622 "data_size": 7936 00:29:06.622 }, 00:29:06.622 { 00:29:06.622 "name": "BaseBdev2", 00:29:06.622 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:29:06.622 "is_configured": true, 00:29:06.622 "data_offset": 256, 00:29:06.622 "data_size": 7936 00:29:06.622 } 00:29:06.622 ] 00:29:06.622 }' 00:29:06.622 11:39:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:29:06.622 11:39:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:29:07.555 11:39:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@748 -- # verify_raid_bdev_process raid_bdev1 none none 00:29:07.555 11:39:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:29:07.555 11:39:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:29:07.555 11:39:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local target=none 00:29:07.555 11:39:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:29:07.555 11:39:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:07.555 11:39:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:29:07.555 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:07.555 "name": "raid_bdev1", 00:29:07.555 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:29:07.555 "strip_size_kb": 0, 00:29:07.555 "state": "online", 00:29:07.555 "raid_level": "raid1", 00:29:07.555 "superblock": true, 00:29:07.555 "num_base_bdevs": 2, 00:29:07.555 "num_base_bdevs_discovered": 2, 00:29:07.555 "num_base_bdevs_operational": 2, 00:29:07.555 "base_bdevs_list": [ 00:29:07.555 { 00:29:07.555 "name": "spare", 00:29:07.555 "uuid": "db8a8c81-e619-5cec-bd99-d2b5802226a5", 00:29:07.555 "is_configured": true, 00:29:07.555 "data_offset": 256, 00:29:07.555 "data_size": 7936 00:29:07.555 }, 00:29:07.555 { 00:29:07.555 "name": "BaseBdev2", 00:29:07.555 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:29:07.555 "is_configured": true, 00:29:07.555 "data_offset": 256, 00:29:07.555 "data_size": 7936 00:29:07.555 } 00:29:07.555 ] 00:29:07.555 }' 00:29:07.555 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:29:07.555 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:29:07.555 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:29:07.555 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:29:07.555 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@749 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:29:07.555 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@749 -- # jq -r '.[].base_bdevs_list[0].name' 00:29:07.814 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@749 -- # [[ spare == \s\p\a\r\e ]] 00:29:07.814 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@752 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_remove_base_bdev spare 00:29:08.072 [2024-07-15 11:39:51.622663] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:29:08.072 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@753 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:29:08.072 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:29:08.072 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:29:08.072 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:29:08.072 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:29:08.072 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:29:08.072 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:29:08.072 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:29:08.072 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:29:08.072 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:29:08.072 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:29:08.072 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:08.330 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:29:08.330 "name": "raid_bdev1", 00:29:08.330 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:29:08.330 "strip_size_kb": 0, 00:29:08.330 "state": "online", 00:29:08.330 "raid_level": "raid1", 00:29:08.330 "superblock": true, 00:29:08.330 "num_base_bdevs": 2, 00:29:08.330 "num_base_bdevs_discovered": 1, 00:29:08.330 "num_base_bdevs_operational": 1, 00:29:08.330 "base_bdevs_list": [ 00:29:08.330 { 00:29:08.330 "name": null, 00:29:08.330 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:08.330 "is_configured": false, 00:29:08.331 "data_offset": 256, 00:29:08.331 "data_size": 7936 00:29:08.331 }, 00:29:08.331 { 00:29:08.331 "name": "BaseBdev2", 00:29:08.331 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:29:08.331 "is_configured": true, 00:29:08.331 "data_offset": 256, 00:29:08.331 "data_size": 7936 00:29:08.331 } 00:29:08.331 ] 00:29:08.331 }' 00:29:08.331 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:29:08.331 11:39:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:29:09.265 11:39:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@754 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 spare 00:29:09.265 [2024-07-15 11:39:52.733634] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:29:09.266 [2024-07-15 11:39:52.733776] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:29:09.266 [2024-07-15 11:39:52.733792] bdev_raid.c:3620:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:29:09.266 [2024-07-15 11:39:52.733819] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:29:09.266 [2024-07-15 11:39:52.737344] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1e1d3a0 00:29:09.266 [2024-07-15 11:39:52.738745] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:29:09.266 11:39:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@755 -- # sleep 1 00:29:10.197 11:39:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@756 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:29:10.197 11:39:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:29:10.197 11:39:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:29:10.197 11:39:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local target=spare 00:29:10.197 11:39:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:29:10.197 11:39:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:29:10.197 11:39:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:10.455 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:10.455 "name": "raid_bdev1", 00:29:10.455 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:29:10.455 "strip_size_kb": 0, 00:29:10.455 "state": "online", 00:29:10.455 "raid_level": "raid1", 00:29:10.455 "superblock": true, 00:29:10.455 "num_base_bdevs": 2, 00:29:10.455 "num_base_bdevs_discovered": 2, 00:29:10.455 "num_base_bdevs_operational": 2, 00:29:10.455 "process": { 00:29:10.455 "type": "rebuild", 00:29:10.455 "target": "spare", 00:29:10.455 "progress": { 00:29:10.455 "blocks": 3072, 00:29:10.455 "percent": 38 00:29:10.455 } 00:29:10.455 }, 00:29:10.455 "base_bdevs_list": [ 00:29:10.455 { 00:29:10.455 "name": "spare", 00:29:10.455 "uuid": "db8a8c81-e619-5cec-bd99-d2b5802226a5", 00:29:10.455 "is_configured": true, 00:29:10.455 "data_offset": 256, 00:29:10.455 "data_size": 7936 00:29:10.455 }, 00:29:10.455 { 00:29:10.455 "name": "BaseBdev2", 00:29:10.455 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:29:10.455 "is_configured": true, 00:29:10.455 "data_offset": 256, 00:29:10.455 "data_size": 7936 00:29:10.455 } 00:29:10.455 ] 00:29:10.455 }' 00:29:10.455 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:29:10.713 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:29:10.713 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:29:10.713 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:29:10.713 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@759 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:29:10.990 [2024-07-15 11:39:54.332131] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:29:10.990 [2024-07-15 11:39:54.351373] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:29:10.990 [2024-07-15 11:39:54.351413] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:10.990 [2024-07-15 11:39:54.351429] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:29:10.990 [2024-07-15 11:39:54.351438] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:29:10.990 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@760 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:29:10.990 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:29:10.990 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:29:10.990 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:29:10.990 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:29:10.990 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:29:10.990 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:29:10.990 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:29:10.990 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:29:10.990 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:29:10.990 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:29:10.990 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:11.248 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:29:11.248 "name": "raid_bdev1", 00:29:11.248 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:29:11.248 "strip_size_kb": 0, 00:29:11.248 "state": "online", 00:29:11.248 "raid_level": "raid1", 00:29:11.248 "superblock": true, 00:29:11.248 "num_base_bdevs": 2, 00:29:11.248 "num_base_bdevs_discovered": 1, 00:29:11.248 "num_base_bdevs_operational": 1, 00:29:11.248 "base_bdevs_list": [ 00:29:11.248 { 00:29:11.248 "name": null, 00:29:11.248 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:11.248 "is_configured": false, 00:29:11.248 "data_offset": 256, 00:29:11.248 "data_size": 7936 00:29:11.248 }, 00:29:11.248 { 00:29:11.248 "name": "BaseBdev2", 00:29:11.248 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:29:11.248 "is_configured": true, 00:29:11.248 "data_offset": 256, 00:29:11.248 "data_size": 7936 00:29:11.248 } 00:29:11.248 ] 00:29:11.248 }' 00:29:11.248 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:29:11.248 11:39:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:29:11.814 11:39:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@761 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b spare_delay -p spare 00:29:12.072 [2024-07-15 11:39:55.438054] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:29:12.072 [2024-07-15 11:39:55.438103] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:12.072 [2024-07-15 11:39:55.438126] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1e1ac80 00:29:12.072 [2024-07-15 11:39:55.438138] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:12.072 [2024-07-15 11:39:55.438328] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:12.072 [2024-07-15 11:39:55.438344] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:29:12.072 [2024-07-15 11:39:55.438398] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:29:12.072 [2024-07-15 11:39:55.438409] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:29:12.072 [2024-07-15 11:39:55.438420] bdev_raid.c:3620:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:29:12.072 [2024-07-15 11:39:55.438438] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:29:12.072 [2024-07-15 11:39:55.441953] bdev_raid.c: 251:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x1e1b2d0 00:29:12.072 [2024-07-15 11:39:55.443279] bdev_raid.c:2824:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:29:12.072 spare 00:29:12.072 11:39:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@762 -- # sleep 1 00:29:13.006 11:39:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@763 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:29:13.006 11:39:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:29:13.006 11:39:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local process_type=rebuild 00:29:13.007 11:39:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local target=spare 00:29:13.007 11:39:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:29:13.007 11:39:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:29:13.007 11:39:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:13.265 11:39:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:13.265 "name": "raid_bdev1", 00:29:13.265 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:29:13.265 "strip_size_kb": 0, 00:29:13.265 "state": "online", 00:29:13.265 "raid_level": "raid1", 00:29:13.265 "superblock": true, 00:29:13.265 "num_base_bdevs": 2, 00:29:13.265 "num_base_bdevs_discovered": 2, 00:29:13.265 "num_base_bdevs_operational": 2, 00:29:13.265 "process": { 00:29:13.265 "type": "rebuild", 00:29:13.265 "target": "spare", 00:29:13.265 "progress": { 00:29:13.265 "blocks": 3072, 00:29:13.265 "percent": 38 00:29:13.265 } 00:29:13.265 }, 00:29:13.265 "base_bdevs_list": [ 00:29:13.265 { 00:29:13.265 "name": "spare", 00:29:13.265 "uuid": "db8a8c81-e619-5cec-bd99-d2b5802226a5", 00:29:13.265 "is_configured": true, 00:29:13.265 "data_offset": 256, 00:29:13.265 "data_size": 7936 00:29:13.265 }, 00:29:13.265 { 00:29:13.265 "name": "BaseBdev2", 00:29:13.265 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:29:13.265 "is_configured": true, 00:29:13.265 "data_offset": 256, 00:29:13.265 "data_size": 7936 00:29:13.265 } 00:29:13.265 ] 00:29:13.265 }' 00:29:13.265 11:39:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:29:13.265 11:39:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:29:13.265 11:39:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:29:13.265 11:39:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # [[ spare == \s\p\a\r\e ]] 00:29:13.265 11:39:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@766 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete spare 00:29:13.524 [2024-07-15 11:39:57.016493] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:29:13.524 [2024-07-15 11:39:57.055615] bdev_raid.c:2513:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:29:13.524 [2024-07-15 11:39:57.055656] bdev_raid.c: 331:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:13.524 [2024-07-15 11:39:57.055671] bdev_raid.c:2120:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:29:13.524 [2024-07-15 11:39:57.055680] bdev_raid.c:2451:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:29:13.524 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@767 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:29:13.524 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:29:13.524 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:29:13.524 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:29:13.524 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:29:13.524 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:29:13.524 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:29:13.524 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:29:13.524 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:29:13.524 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:29:13.524 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:29:13.524 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:13.782 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:29:13.782 "name": "raid_bdev1", 00:29:13.782 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:29:13.782 "strip_size_kb": 0, 00:29:13.782 "state": "online", 00:29:13.782 "raid_level": "raid1", 00:29:13.782 "superblock": true, 00:29:13.782 "num_base_bdevs": 2, 00:29:13.782 "num_base_bdevs_discovered": 1, 00:29:13.782 "num_base_bdevs_operational": 1, 00:29:13.782 "base_bdevs_list": [ 00:29:13.782 { 00:29:13.782 "name": null, 00:29:13.782 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:13.782 "is_configured": false, 00:29:13.782 "data_offset": 256, 00:29:13.782 "data_size": 7936 00:29:13.782 }, 00:29:13.782 { 00:29:13.782 "name": "BaseBdev2", 00:29:13.782 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:29:13.782 "is_configured": true, 00:29:13.782 "data_offset": 256, 00:29:13.782 "data_size": 7936 00:29:13.782 } 00:29:13.782 ] 00:29:13.782 }' 00:29:13.782 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:29:13.782 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:29:14.348 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@768 -- # verify_raid_bdev_process raid_bdev1 none none 00:29:14.348 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:29:14.348 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:29:14.348 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local target=none 00:29:14.348 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:29:14.348 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:29:14.348 11:39:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:14.606 11:39:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:14.606 "name": "raid_bdev1", 00:29:14.606 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:29:14.606 "strip_size_kb": 0, 00:29:14.606 "state": "online", 00:29:14.606 "raid_level": "raid1", 00:29:14.606 "superblock": true, 00:29:14.606 "num_base_bdevs": 2, 00:29:14.606 "num_base_bdevs_discovered": 1, 00:29:14.606 "num_base_bdevs_operational": 1, 00:29:14.606 "base_bdevs_list": [ 00:29:14.606 { 00:29:14.606 "name": null, 00:29:14.606 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:14.606 "is_configured": false, 00:29:14.606 "data_offset": 256, 00:29:14.606 "data_size": 7936 00:29:14.606 }, 00:29:14.606 { 00:29:14.606 "name": "BaseBdev2", 00:29:14.606 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:29:14.606 "is_configured": true, 00:29:14.606 "data_offset": 256, 00:29:14.606 "data_size": 7936 00:29:14.606 } 00:29:14.606 ] 00:29:14.606 }' 00:29:14.606 11:39:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:29:14.865 11:39:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:29:14.865 11:39:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:29:14.865 11:39:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:29:14.865 11:39:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@771 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_delete BaseBdev1 00:29:15.123 11:39:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@772 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:29:15.381 [2024-07-15 11:39:58.727821] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:29:15.381 [2024-07-15 11:39:58.727865] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:15.381 [2024-07-15 11:39:58.727889] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1c82fa0 00:29:15.381 [2024-07-15 11:39:58.727901] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:15.381 [2024-07-15 11:39:58.728065] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:15.381 [2024-07-15 11:39:58.728082] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:29:15.381 [2024-07-15 11:39:58.728127] bdev_raid.c:3752:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:29:15.381 [2024-07-15 11:39:58.728138] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:29:15.381 [2024-07-15 11:39:58.728149] bdev_raid.c:3581:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:29:15.381 BaseBdev1 00:29:15.381 11:39:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@773 -- # sleep 1 00:29:16.315 11:39:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@774 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:29:16.315 11:39:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:29:16.315 11:39:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:29:16.315 11:39:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:29:16.315 11:39:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:29:16.315 11:39:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:29:16.315 11:39:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:29:16.315 11:39:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:29:16.315 11:39:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:29:16.315 11:39:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:29:16.315 11:39:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:29:16.315 11:39:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:16.573 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:29:16.573 "name": "raid_bdev1", 00:29:16.573 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:29:16.573 "strip_size_kb": 0, 00:29:16.573 "state": "online", 00:29:16.573 "raid_level": "raid1", 00:29:16.573 "superblock": true, 00:29:16.573 "num_base_bdevs": 2, 00:29:16.573 "num_base_bdevs_discovered": 1, 00:29:16.573 "num_base_bdevs_operational": 1, 00:29:16.573 "base_bdevs_list": [ 00:29:16.573 { 00:29:16.573 "name": null, 00:29:16.573 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:16.573 "is_configured": false, 00:29:16.573 "data_offset": 256, 00:29:16.573 "data_size": 7936 00:29:16.573 }, 00:29:16.573 { 00:29:16.573 "name": "BaseBdev2", 00:29:16.573 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:29:16.573 "is_configured": true, 00:29:16.573 "data_offset": 256, 00:29:16.573 "data_size": 7936 00:29:16.573 } 00:29:16.573 ] 00:29:16.573 }' 00:29:16.573 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:29:16.573 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:29:17.140 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@775 -- # verify_raid_bdev_process raid_bdev1 none none 00:29:17.140 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:29:17.140 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:29:17.140 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local target=none 00:29:17.140 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:29:17.140 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:29:17.140 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:17.399 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:17.399 "name": "raid_bdev1", 00:29:17.399 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:29:17.399 "strip_size_kb": 0, 00:29:17.399 "state": "online", 00:29:17.399 "raid_level": "raid1", 00:29:17.399 "superblock": true, 00:29:17.399 "num_base_bdevs": 2, 00:29:17.399 "num_base_bdevs_discovered": 1, 00:29:17.399 "num_base_bdevs_operational": 1, 00:29:17.399 "base_bdevs_list": [ 00:29:17.399 { 00:29:17.399 "name": null, 00:29:17.399 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:17.399 "is_configured": false, 00:29:17.399 "data_offset": 256, 00:29:17.399 "data_size": 7936 00:29:17.399 }, 00:29:17.399 { 00:29:17.399 "name": "BaseBdev2", 00:29:17.399 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:29:17.399 "is_configured": true, 00:29:17.399 "data_offset": 256, 00:29:17.399 "data_size": 7936 00:29:17.399 } 00:29:17.399 ] 00:29:17.399 }' 00:29:17.399 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:29:17.399 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:29:17.399 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:29:17.399 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:29:17.399 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@776 -- # NOT /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:29:17.399 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@648 -- # local es=0 00:29:17.399 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:29:17.399 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:29:17.399 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:29:17.399 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:29:17.399 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:29:17.399 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:29:17.399 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:29:17.399 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:29:17.399 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py ]] 00:29:17.399 11:40:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:29:17.657 [2024-07-15 11:40:01.178346] bdev_raid.c:3198:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:17.657 [2024-07-15 11:40:01.178471] bdev_raid.c:3562:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:29:17.657 [2024-07-15 11:40:01.178487] bdev_raid.c:3581:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:29:17.657 request: 00:29:17.657 { 00:29:17.657 "base_bdev": "BaseBdev1", 00:29:17.657 "raid_bdev": "raid_bdev1", 00:29:17.657 "method": "bdev_raid_add_base_bdev", 00:29:17.657 "req_id": 1 00:29:17.657 } 00:29:17.657 Got JSON-RPC error response 00:29:17.657 response: 00:29:17.657 { 00:29:17.657 "code": -22, 00:29:17.657 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:29:17.657 } 00:29:17.657 11:40:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@651 -- # es=1 00:29:17.657 11:40:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:29:17.658 11:40:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:29:17.658 11:40:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:29:17.658 11:40:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@777 -- # sleep 1 00:29:19.032 11:40:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@778 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:29:19.032 11:40:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@116 -- # local raid_bdev_name=raid_bdev1 00:29:19.032 11:40:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@117 -- # local expected_state=online 00:29:19.032 11:40:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@118 -- # local raid_level=raid1 00:29:19.032 11:40:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@119 -- # local strip_size=0 00:29:19.032 11:40:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@120 -- # local num_base_bdevs_operational=1 00:29:19.032 11:40:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@121 -- # local raid_bdev_info 00:29:19.032 11:40:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@122 -- # local num_base_bdevs 00:29:19.032 11:40:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@123 -- # local num_base_bdevs_discovered 00:29:19.032 11:40:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@124 -- # local tmp 00:29:19.032 11:40:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:29:19.032 11:40:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:19.032 11:40:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@126 -- # raid_bdev_info='{ 00:29:19.032 "name": "raid_bdev1", 00:29:19.032 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:29:19.032 "strip_size_kb": 0, 00:29:19.032 "state": "online", 00:29:19.032 "raid_level": "raid1", 00:29:19.032 "superblock": true, 00:29:19.032 "num_base_bdevs": 2, 00:29:19.032 "num_base_bdevs_discovered": 1, 00:29:19.033 "num_base_bdevs_operational": 1, 00:29:19.033 "base_bdevs_list": [ 00:29:19.033 { 00:29:19.033 "name": null, 00:29:19.033 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:19.033 "is_configured": false, 00:29:19.033 "data_offset": 256, 00:29:19.033 "data_size": 7936 00:29:19.033 }, 00:29:19.033 { 00:29:19.033 "name": "BaseBdev2", 00:29:19.033 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:29:19.033 "is_configured": true, 00:29:19.033 "data_offset": 256, 00:29:19.033 "data_size": 7936 00:29:19.033 } 00:29:19.033 ] 00:29:19.033 }' 00:29:19.033 11:40:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@128 -- # xtrace_disable 00:29:19.033 11:40:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:29:19.599 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@779 -- # verify_raid_bdev_process raid_bdev1 none none 00:29:19.599 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_name=raid_bdev1 00:29:19.599 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local process_type=none 00:29:19.599 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local target=none 00:29:19.599 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local raid_bdev_info 00:29:19.599 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-raid.sock bdev_raid_get_bdevs all 00:29:19.599 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:19.857 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:19.857 "name": "raid_bdev1", 00:29:19.857 "uuid": "1b49bd31-8520-438e-8f0f-e7fab8ce3145", 00:29:19.857 "strip_size_kb": 0, 00:29:19.857 "state": "online", 00:29:19.857 "raid_level": "raid1", 00:29:19.857 "superblock": true, 00:29:19.857 "num_base_bdevs": 2, 00:29:19.857 "num_base_bdevs_discovered": 1, 00:29:19.857 "num_base_bdevs_operational": 1, 00:29:19.857 "base_bdevs_list": [ 00:29:19.857 { 00:29:19.857 "name": null, 00:29:19.857 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:19.857 "is_configured": false, 00:29:19.857 "data_offset": 256, 00:29:19.857 "data_size": 7936 00:29:19.857 }, 00:29:19.857 { 00:29:19.857 "name": "BaseBdev2", 00:29:19.857 "uuid": "d0820610-563e-5bc0-b409-0a6e87aec6c0", 00:29:19.857 "is_configured": true, 00:29:19.857 "data_offset": 256, 00:29:19.857 "data_size": 7936 00:29:19.857 } 00:29:19.857 ] 00:29:19.857 }' 00:29:19.857 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '.process.type // "none"' 00:29:19.857 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # [[ none == \n\o\n\e ]] 00:29:19.857 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # jq -r '.process.target // "none"' 00:29:19.857 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@190 -- # [[ none == \n\o\n\e ]] 00:29:19.857 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@782 -- # killprocess 1021952 00:29:19.857 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@948 -- # '[' -z 1021952 ']' 00:29:19.857 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@952 -- # kill -0 1021952 00:29:19.857 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@953 -- # uname 00:29:19.857 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:29:19.857 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1021952 00:29:19.857 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:29:19.857 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:29:19.857 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1021952' 00:29:19.857 killing process with pid 1021952 00:29:19.857 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@967 -- # kill 1021952 00:29:19.857 Received shutdown signal, test time was about 60.000000 seconds 00:29:19.857 00:29:19.857 Latency(us) 00:29:19.857 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:19.857 =================================================================================================================== 00:29:19.857 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:29:19.857 [2024-07-15 11:40:03.431857] bdev_raid.c:1358:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:29:19.857 [2024-07-15 11:40:03.431950] bdev_raid.c: 474:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:19.857 [2024-07-15 11:40:03.431997] bdev_raid.c: 451:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:19.857 [2024-07-15 11:40:03.432010] bdev_raid.c: 366:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x1e1bf60 name raid_bdev1, state offline 00:29:19.857 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@972 -- # wait 1021952 00:29:20.114 [2024-07-15 11:40:03.464125] bdev_raid.c:1375:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:29:20.114 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@784 -- # return 0 00:29:20.114 00:29:20.114 real 0m29.088s 00:29:20.114 user 0m46.491s 00:29:20.114 sys 0m3.813s 00:29:20.114 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1124 -- # xtrace_disable 00:29:20.114 11:40:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:29:20.114 ************************************ 00:29:20.114 END TEST raid_rebuild_test_sb_md_interleaved 00:29:20.114 ************************************ 00:29:20.372 11:40:03 bdev_raid -- common/autotest_common.sh@1142 -- # return 0 00:29:20.372 11:40:03 bdev_raid -- bdev/bdev_raid.sh@916 -- # trap - EXIT 00:29:20.372 11:40:03 bdev_raid -- bdev/bdev_raid.sh@917 -- # cleanup 00:29:20.372 11:40:03 bdev_raid -- bdev/bdev_raid.sh@58 -- # '[' -n 1021952 ']' 00:29:20.372 11:40:03 bdev_raid -- bdev/bdev_raid.sh@58 -- # ps -p 1021952 00:29:20.372 11:40:03 bdev_raid -- bdev/bdev_raid.sh@62 -- # rm -rf /raidtest 00:29:20.372 00:29:20.372 real 18m21.592s 00:29:20.372 user 31m2.595s 00:29:20.372 sys 3m21.624s 00:29:20.372 11:40:03 bdev_raid -- common/autotest_common.sh@1124 -- # xtrace_disable 00:29:20.372 11:40:03 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:29:20.372 ************************************ 00:29:20.372 END TEST bdev_raid 00:29:20.372 ************************************ 00:29:20.372 11:40:03 -- common/autotest_common.sh@1142 -- # return 0 00:29:20.372 11:40:03 -- spdk/autotest.sh@191 -- # run_test bdevperf_config /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevperf/test_config.sh 00:29:20.372 11:40:03 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:29:20.372 11:40:03 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:29:20.372 11:40:03 -- common/autotest_common.sh@10 -- # set +x 00:29:20.372 ************************************ 00:29:20.372 START TEST bdevperf_config 00:29:20.372 ************************************ 00:29:20.372 11:40:03 bdevperf_config -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevperf/test_config.sh 00:29:20.372 * Looking for test storage... 00:29:20.373 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevperf 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/test_config.sh@10 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevperf/common.sh 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/common.sh@5 -- # bdevperf=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/test_config.sh@12 -- # jsonconf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevperf/conf.json 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/test_config.sh@13 -- # testconf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevperf/test.conf 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/test_config.sh@15 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/test_config.sh@17 -- # create_job global read Malloc0 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/common.sh@8 -- # local job_section=global 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/common.sh@9 -- # local rw=read 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/common.sh@10 -- # local filename=Malloc0 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/common.sh@12 -- # [[ global == \g\l\o\b\a\l ]] 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/common.sh@13 -- # cat 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/common.sh@18 -- # job='[global]' 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/common.sh@19 -- # echo 00:29:20.373 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/common.sh@20 -- # cat 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/test_config.sh@18 -- # create_job job0 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/common.sh@8 -- # local job_section=job0 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/common.sh@9 -- # local rw= 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/common.sh@10 -- # local filename= 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/common.sh@12 -- # [[ job0 == \g\l\o\b\a\l ]] 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/common.sh@18 -- # job='[job0]' 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/common.sh@19 -- # echo 00:29:20.373 00:29:20.373 11:40:03 bdevperf_config -- bdevperf/common.sh@20 -- # cat 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/test_config.sh@19 -- # create_job job1 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@8 -- # local job_section=job1 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@9 -- # local rw= 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@10 -- # local filename= 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@12 -- # [[ job1 == \g\l\o\b\a\l ]] 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@18 -- # job='[job1]' 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@19 -- # echo 00:29:20.708 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@20 -- # cat 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/test_config.sh@20 -- # create_job job2 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@8 -- # local job_section=job2 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@9 -- # local rw= 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@10 -- # local filename= 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@12 -- # [[ job2 == \g\l\o\b\a\l ]] 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@18 -- # job='[job2]' 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@19 -- # echo 00:29:20.708 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@20 -- # cat 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/test_config.sh@21 -- # create_job job3 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@8 -- # local job_section=job3 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@9 -- # local rw= 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@10 -- # local filename= 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@12 -- # [[ job3 == \g\l\o\b\a\l ]] 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@18 -- # job='[job3]' 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@19 -- # echo 00:29:20.708 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/common.sh@20 -- # cat 00:29:20.708 11:40:03 bdevperf_config -- bdevperf/test_config.sh@22 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -t 2 --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevperf/conf.json -j /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevperf/test.conf 00:29:23.234 11:40:06 bdevperf_config -- bdevperf/test_config.sh@22 -- # bdevperf_output='[2024-07-15 11:40:04.043564] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:29:23.234 [2024-07-15 11:40:04.043619] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1026110 ] 00:29:23.234 Using job config with 4 jobs 00:29:23.234 [2024-07-15 11:40:04.175041] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:23.234 [2024-07-15 11:40:04.295103] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:29:23.234 cpumask for '\''job0'\'' is too big 00:29:23.234 cpumask for '\''job1'\'' is too big 00:29:23.234 cpumask for '\''job2'\'' is too big 00:29:23.234 cpumask for '\''job3'\'' is too big 00:29:23.234 Running I/O for 2 seconds... 00:29:23.234 00:29:23.234 Latency(us) 00:29:23.234 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:23.234 Job: Malloc0 (Core Mask 0xff, workload: read, depth: 256, IO size: 1024) 00:29:23.234 Malloc0 : 2.02 23965.54 23.40 0.00 0.00 10676.35 1866.35 16412.49 00:29:23.234 Job: Malloc0 (Core Mask 0xff, workload: read, depth: 256, IO size: 1024) 00:29:23.234 Malloc0 : 2.02 23943.63 23.38 0.00 0.00 10661.63 1837.86 14702.86 00:29:23.234 Job: Malloc0 (Core Mask 0xff, workload: read, depth: 256, IO size: 1024) 00:29:23.234 Malloc0 : 2.02 23921.85 23.36 0.00 0.00 10647.25 1852.10 12822.26 00:29:23.234 Job: Malloc0 (Core Mask 0xff, workload: read, depth: 256, IO size: 1024) 00:29:23.234 Malloc0 : 2.02 23900.17 23.34 0.00 0.00 10633.04 1837.86 10941.66 00:29:23.234 =================================================================================================================== 00:29:23.234 Total : 95731.19 93.49 0.00 0.00 10654.57 1837.86 16412.49' 00:29:23.234 11:40:06 bdevperf_config -- bdevperf/test_config.sh@23 -- # get_num_jobs '[2024-07-15 11:40:04.043564] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:29:23.234 [2024-07-15 11:40:04.043619] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1026110 ] 00:29:23.234 Using job config with 4 jobs 00:29:23.234 [2024-07-15 11:40:04.175041] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:23.234 [2024-07-15 11:40:04.295103] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:29:23.234 cpumask for '\''job0'\'' is too big 00:29:23.234 cpumask for '\''job1'\'' is too big 00:29:23.234 cpumask for '\''job2'\'' is too big 00:29:23.234 cpumask for '\''job3'\'' is too big 00:29:23.234 Running I/O for 2 seconds... 00:29:23.234 00:29:23.234 Latency(us) 00:29:23.234 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:23.234 Job: Malloc0 (Core Mask 0xff, workload: read, depth: 256, IO size: 1024) 00:29:23.234 Malloc0 : 2.02 23965.54 23.40 0.00 0.00 10676.35 1866.35 16412.49 00:29:23.235 Job: Malloc0 (Core Mask 0xff, workload: read, depth: 256, IO size: 1024) 00:29:23.235 Malloc0 : 2.02 23943.63 23.38 0.00 0.00 10661.63 1837.86 14702.86 00:29:23.235 Job: Malloc0 (Core Mask 0xff, workload: read, depth: 256, IO size: 1024) 00:29:23.235 Malloc0 : 2.02 23921.85 23.36 0.00 0.00 10647.25 1852.10 12822.26 00:29:23.235 Job: Malloc0 (Core Mask 0xff, workload: read, depth: 256, IO size: 1024) 00:29:23.235 Malloc0 : 2.02 23900.17 23.34 0.00 0.00 10633.04 1837.86 10941.66 00:29:23.235 =================================================================================================================== 00:29:23.235 Total : 95731.19 93.49 0.00 0.00 10654.57 1837.86 16412.49' 00:29:23.235 11:40:06 bdevperf_config -- bdevperf/common.sh@32 -- # echo '[2024-07-15 11:40:04.043564] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:29:23.235 [2024-07-15 11:40:04.043619] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1026110 ] 00:29:23.235 Using job config with 4 jobs 00:29:23.235 [2024-07-15 11:40:04.175041] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:23.235 [2024-07-15 11:40:04.295103] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:29:23.235 cpumask for '\''job0'\'' is too big 00:29:23.235 cpumask for '\''job1'\'' is too big 00:29:23.235 cpumask for '\''job2'\'' is too big 00:29:23.235 cpumask for '\''job3'\'' is too big 00:29:23.235 Running I/O for 2 seconds... 00:29:23.235 00:29:23.235 Latency(us) 00:29:23.235 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:23.235 Job: Malloc0 (Core Mask 0xff, workload: read, depth: 256, IO size: 1024) 00:29:23.235 Malloc0 : 2.02 23965.54 23.40 0.00 0.00 10676.35 1866.35 16412.49 00:29:23.235 Job: Malloc0 (Core Mask 0xff, workload: read, depth: 256, IO size: 1024) 00:29:23.235 Malloc0 : 2.02 23943.63 23.38 0.00 0.00 10661.63 1837.86 14702.86 00:29:23.235 Job: Malloc0 (Core Mask 0xff, workload: read, depth: 256, IO size: 1024) 00:29:23.235 Malloc0 : 2.02 23921.85 23.36 0.00 0.00 10647.25 1852.10 12822.26 00:29:23.235 Job: Malloc0 (Core Mask 0xff, workload: read, depth: 256, IO size: 1024) 00:29:23.235 Malloc0 : 2.02 23900.17 23.34 0.00 0.00 10633.04 1837.86 10941.66 00:29:23.235 =================================================================================================================== 00:29:23.235 Total : 95731.19 93.49 0.00 0.00 10654.57 1837.86 16412.49' 00:29:23.235 11:40:06 bdevperf_config -- bdevperf/common.sh@32 -- # grep -oE 'Using job config with [0-9]+ jobs' 00:29:23.235 11:40:06 bdevperf_config -- bdevperf/common.sh@32 -- # grep -oE '[0-9]+' 00:29:23.235 11:40:06 bdevperf_config -- bdevperf/test_config.sh@23 -- # [[ 4 == \4 ]] 00:29:23.235 11:40:06 bdevperf_config -- bdevperf/test_config.sh@25 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -C -t 2 --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevperf/conf.json -j /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevperf/test.conf 00:29:23.235 [2024-07-15 11:40:06.783578] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:29:23.235 [2024-07-15 11:40:06.783644] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1026468 ] 00:29:23.493 [2024-07-15 11:40:06.922397] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:23.493 [2024-07-15 11:40:07.030490] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:29:23.750 cpumask for 'job0' is too big 00:29:23.750 cpumask for 'job1' is too big 00:29:23.750 cpumask for 'job2' is too big 00:29:23.750 cpumask for 'job3' is too big 00:29:26.269 11:40:09 bdevperf_config -- bdevperf/test_config.sh@25 -- # bdevperf_output='Using job config with 4 jobs 00:29:26.269 Running I/O for 2 seconds... 00:29:26.269 00:29:26.269 Latency(us) 00:29:26.269 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:26.269 Job: Malloc0 (Core Mask 0xff, workload: read, depth: 256, IO size: 1024) 00:29:26.269 Malloc0 : 2.01 24051.80 23.49 0.00 0.00 10628.84 1852.10 16298.52 00:29:26.269 Job: Malloc0 (Core Mask 0xff, workload: read, depth: 256, IO size: 1024) 00:29:26.269 Malloc0 : 2.02 24061.01 23.50 0.00 0.00 10600.71 1852.10 14417.92 00:29:26.269 Job: Malloc0 (Core Mask 0xff, workload: read, depth: 256, IO size: 1024) 00:29:26.270 Malloc0 : 2.02 24039.06 23.48 0.00 0.00 10586.66 1837.86 12594.31 00:29:26.270 Job: Malloc0 (Core Mask 0xff, workload: read, depth: 256, IO size: 1024) 00:29:26.270 Malloc0 : 2.03 24017.19 23.45 0.00 0.00 10571.96 1852.10 10884.67 00:29:26.270 =================================================================================================================== 00:29:26.270 Total : 96169.06 93.92 0.00 0.00 10597.00 1837.86 16298.52' 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/test_config.sh@27 -- # cleanup 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@36 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevperf/test.conf 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/test_config.sh@29 -- # create_job job0 write Malloc0 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@8 -- # local job_section=job0 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@9 -- # local rw=write 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@10 -- # local filename=Malloc0 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@12 -- # [[ job0 == \g\l\o\b\a\l ]] 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@18 -- # job='[job0]' 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@19 -- # echo 00:29:26.270 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@20 -- # cat 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/test_config.sh@30 -- # create_job job1 write Malloc0 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@8 -- # local job_section=job1 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@9 -- # local rw=write 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@10 -- # local filename=Malloc0 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@12 -- # [[ job1 == \g\l\o\b\a\l ]] 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@18 -- # job='[job1]' 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@19 -- # echo 00:29:26.270 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@20 -- # cat 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/test_config.sh@31 -- # create_job job2 write Malloc0 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@8 -- # local job_section=job2 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@9 -- # local rw=write 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@10 -- # local filename=Malloc0 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@12 -- # [[ job2 == \g\l\o\b\a\l ]] 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@18 -- # job='[job2]' 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@19 -- # echo 00:29:26.270 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/common.sh@20 -- # cat 00:29:26.270 11:40:09 bdevperf_config -- bdevperf/test_config.sh@32 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -t 2 --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevperf/conf.json -j /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevperf/test.conf 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/test_config.sh@32 -- # bdevperf_output='[2024-07-15 11:40:09.537004] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:29:28.793 [2024-07-15 11:40:09.537076] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1026817 ] 00:29:28.793 Using job config with 3 jobs 00:29:28.793 [2024-07-15 11:40:09.679640] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:28.793 [2024-07-15 11:40:09.800790] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:29:28.793 cpumask for '\''job0'\'' is too big 00:29:28.793 cpumask for '\''job1'\'' is too big 00:29:28.793 cpumask for '\''job2'\'' is too big 00:29:28.793 Running I/O for 2 seconds... 00:29:28.793 00:29:28.793 Latency(us) 00:29:28.793 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:28.793 Job: Malloc0 (Core Mask 0xff, workload: write, depth: 256, IO size: 1024) 00:29:28.793 Malloc0 : 2.02 32512.47 31.75 0.00 0.00 7864.83 1802.24 11568.53 00:29:28.793 Job: Malloc0 (Core Mask 0xff, workload: write, depth: 256, IO size: 1024) 00:29:28.793 Malloc0 : 2.02 32482.59 31.72 0.00 0.00 7854.81 1823.61 9744.92 00:29:28.793 Job: Malloc0 (Core Mask 0xff, workload: write, depth: 256, IO size: 1024) 00:29:28.793 Malloc0 : 2.02 32452.85 31.69 0.00 0.00 7845.27 1780.87 8890.10 00:29:28.793 =================================================================================================================== 00:29:28.793 Total : 97447.91 95.16 0.00 0.00 7854.97 1780.87 11568.53' 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/test_config.sh@33 -- # get_num_jobs '[2024-07-15 11:40:09.537004] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:29:28.793 [2024-07-15 11:40:09.537076] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1026817 ] 00:29:28.793 Using job config with 3 jobs 00:29:28.793 [2024-07-15 11:40:09.679640] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:28.793 [2024-07-15 11:40:09.800790] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:29:28.793 cpumask for '\''job0'\'' is too big 00:29:28.793 cpumask for '\''job1'\'' is too big 00:29:28.793 cpumask for '\''job2'\'' is too big 00:29:28.793 Running I/O for 2 seconds... 00:29:28.793 00:29:28.793 Latency(us) 00:29:28.793 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:28.793 Job: Malloc0 (Core Mask 0xff, workload: write, depth: 256, IO size: 1024) 00:29:28.793 Malloc0 : 2.02 32512.47 31.75 0.00 0.00 7864.83 1802.24 11568.53 00:29:28.793 Job: Malloc0 (Core Mask 0xff, workload: write, depth: 256, IO size: 1024) 00:29:28.793 Malloc0 : 2.02 32482.59 31.72 0.00 0.00 7854.81 1823.61 9744.92 00:29:28.793 Job: Malloc0 (Core Mask 0xff, workload: write, depth: 256, IO size: 1024) 00:29:28.793 Malloc0 : 2.02 32452.85 31.69 0.00 0.00 7845.27 1780.87 8890.10 00:29:28.793 =================================================================================================================== 00:29:28.793 Total : 97447.91 95.16 0.00 0.00 7854.97 1780.87 11568.53' 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@32 -- # echo '[2024-07-15 11:40:09.537004] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:29:28.793 [2024-07-15 11:40:09.537076] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1026817 ] 00:29:28.793 Using job config with 3 jobs 00:29:28.793 [2024-07-15 11:40:09.679640] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:28.793 [2024-07-15 11:40:09.800790] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:29:28.793 cpumask for '\''job0'\'' is too big 00:29:28.793 cpumask for '\''job1'\'' is too big 00:29:28.793 cpumask for '\''job2'\'' is too big 00:29:28.793 Running I/O for 2 seconds... 00:29:28.793 00:29:28.793 Latency(us) 00:29:28.793 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:28.793 Job: Malloc0 (Core Mask 0xff, workload: write, depth: 256, IO size: 1024) 00:29:28.793 Malloc0 : 2.02 32512.47 31.75 0.00 0.00 7864.83 1802.24 11568.53 00:29:28.793 Job: Malloc0 (Core Mask 0xff, workload: write, depth: 256, IO size: 1024) 00:29:28.793 Malloc0 : 2.02 32482.59 31.72 0.00 0.00 7854.81 1823.61 9744.92 00:29:28.793 Job: Malloc0 (Core Mask 0xff, workload: write, depth: 256, IO size: 1024) 00:29:28.793 Malloc0 : 2.02 32452.85 31.69 0.00 0.00 7845.27 1780.87 8890.10 00:29:28.793 =================================================================================================================== 00:29:28.793 Total : 97447.91 95.16 0.00 0.00 7854.97 1780.87 11568.53' 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@32 -- # grep -oE 'Using job config with [0-9]+ jobs' 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@32 -- # grep -oE '[0-9]+' 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/test_config.sh@33 -- # [[ 3 == \3 ]] 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/test_config.sh@35 -- # cleanup 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@36 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevperf/test.conf 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/test_config.sh@37 -- # create_job global rw Malloc0:Malloc1 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@8 -- # local job_section=global 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@9 -- # local rw=rw 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@10 -- # local filename=Malloc0:Malloc1 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@12 -- # [[ global == \g\l\o\b\a\l ]] 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@13 -- # cat 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@18 -- # job='[global]' 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@19 -- # echo 00:29:28.793 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@20 -- # cat 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/test_config.sh@38 -- # create_job job0 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@8 -- # local job_section=job0 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@9 -- # local rw= 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@10 -- # local filename= 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@12 -- # [[ job0 == \g\l\o\b\a\l ]] 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@18 -- # job='[job0]' 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@19 -- # echo 00:29:28.793 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@20 -- # cat 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/test_config.sh@39 -- # create_job job1 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@8 -- # local job_section=job1 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@9 -- # local rw= 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@10 -- # local filename= 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@12 -- # [[ job1 == \g\l\o\b\a\l ]] 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@18 -- # job='[job1]' 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@19 -- # echo 00:29:28.793 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@20 -- # cat 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/test_config.sh@40 -- # create_job job2 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@8 -- # local job_section=job2 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@9 -- # local rw= 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@10 -- # local filename= 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@12 -- # [[ job2 == \g\l\o\b\a\l ]] 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@18 -- # job='[job2]' 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@19 -- # echo 00:29:28.793 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@20 -- # cat 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/test_config.sh@41 -- # create_job job3 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@8 -- # local job_section=job3 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@9 -- # local rw= 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@10 -- # local filename= 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@12 -- # [[ job3 == \g\l\o\b\a\l ]] 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@18 -- # job='[job3]' 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@19 -- # echo 00:29:28.793 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/common.sh@20 -- # cat 00:29:28.793 11:40:12 bdevperf_config -- bdevperf/test_config.sh@42 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -t 2 --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevperf/conf.json -j /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevperf/test.conf 00:29:32.074 11:40:15 bdevperf_config -- bdevperf/test_config.sh@42 -- # bdevperf_output='[2024-07-15 11:40:12.352096] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:29:32.074 [2024-07-15 11:40:12.352165] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1027177 ] 00:29:32.074 Using job config with 4 jobs 00:29:32.074 [2024-07-15 11:40:12.498134] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:32.074 [2024-07-15 11:40:12.614279] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:29:32.074 cpumask for '\''job0'\'' is too big 00:29:32.074 cpumask for '\''job1'\'' is too big 00:29:32.074 cpumask for '\''job2'\'' is too big 00:29:32.074 cpumask for '\''job3'\'' is too big 00:29:32.074 Running I/O for 2 seconds... 00:29:32.074 00:29:32.074 Latency(us) 00:29:32.074 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:32.074 Job: Malloc0 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.074 Malloc0 : 2.04 12035.67 11.75 0.00 0.00 21245.82 3789.69 32824.99 00:29:32.074 Job: Malloc1 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.074 Malloc1 : 2.04 12024.50 11.74 0.00 0.00 21248.90 4616.01 32824.99 00:29:32.074 Job: Malloc0 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.074 Malloc0 : 2.05 12013.70 11.73 0.00 0.00 21192.95 3732.70 28949.82 00:29:32.074 Job: Malloc1 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.074 Malloc1 : 2.05 12002.56 11.72 0.00 0.00 21190.85 4587.52 28949.82 00:29:32.074 Job: Malloc0 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.074 Malloc0 : 2.05 11991.78 11.71 0.00 0.00 21131.94 3732.70 25188.62 00:29:32.074 Job: Malloc1 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.074 Malloc1 : 2.05 11980.79 11.70 0.00 0.00 21131.98 4587.52 25188.62 00:29:32.074 Job: Malloc0 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.074 Malloc0 : 2.05 11969.98 11.69 0.00 0.00 21075.30 3732.70 21541.40 00:29:32.074 Job: Malloc1 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.074 Malloc1 : 2.06 11958.98 11.68 0.00 0.00 21075.34 4587.52 21655.37 00:29:32.074 =================================================================================================================== 00:29:32.074 Total : 95977.95 93.73 0.00 0.00 21161.63 3732.70 32824.99' 00:29:32.074 11:40:15 bdevperf_config -- bdevperf/test_config.sh@43 -- # get_num_jobs '[2024-07-15 11:40:12.352096] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:29:32.074 [2024-07-15 11:40:12.352165] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1027177 ] 00:29:32.074 Using job config with 4 jobs 00:29:32.074 [2024-07-15 11:40:12.498134] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:32.074 [2024-07-15 11:40:12.614279] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:29:32.074 cpumask for '\''job0'\'' is too big 00:29:32.074 cpumask for '\''job1'\'' is too big 00:29:32.074 cpumask for '\''job2'\'' is too big 00:29:32.074 cpumask for '\''job3'\'' is too big 00:29:32.074 Running I/O for 2 seconds... 00:29:32.074 00:29:32.074 Latency(us) 00:29:32.074 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:32.074 Job: Malloc0 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.074 Malloc0 : 2.04 12035.67 11.75 0.00 0.00 21245.82 3789.69 32824.99 00:29:32.074 Job: Malloc1 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.074 Malloc1 : 2.04 12024.50 11.74 0.00 0.00 21248.90 4616.01 32824.99 00:29:32.074 Job: Malloc0 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.074 Malloc0 : 2.05 12013.70 11.73 0.00 0.00 21192.95 3732.70 28949.82 00:29:32.074 Job: Malloc1 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.074 Malloc1 : 2.05 12002.56 11.72 0.00 0.00 21190.85 4587.52 28949.82 00:29:32.074 Job: Malloc0 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.074 Malloc0 : 2.05 11991.78 11.71 0.00 0.00 21131.94 3732.70 25188.62 00:29:32.074 Job: Malloc1 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.075 Malloc1 : 2.05 11980.79 11.70 0.00 0.00 21131.98 4587.52 25188.62 00:29:32.075 Job: Malloc0 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.075 Malloc0 : 2.05 11969.98 11.69 0.00 0.00 21075.30 3732.70 21541.40 00:29:32.075 Job: Malloc1 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.075 Malloc1 : 2.06 11958.98 11.68 0.00 0.00 21075.34 4587.52 21655.37 00:29:32.075 =================================================================================================================== 00:29:32.075 Total : 95977.95 93.73 0.00 0.00 21161.63 3732.70 32824.99' 00:29:32.075 11:40:15 bdevperf_config -- bdevperf/common.sh@32 -- # echo '[2024-07-15 11:40:12.352096] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:29:32.075 [2024-07-15 11:40:12.352165] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1027177 ] 00:29:32.075 Using job config with 4 jobs 00:29:32.075 [2024-07-15 11:40:12.498134] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:32.075 [2024-07-15 11:40:12.614279] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:29:32.075 cpumask for '\''job0'\'' is too big 00:29:32.075 cpumask for '\''job1'\'' is too big 00:29:32.075 cpumask for '\''job2'\'' is too big 00:29:32.075 cpumask for '\''job3'\'' is too big 00:29:32.075 Running I/O for 2 seconds... 00:29:32.075 00:29:32.075 Latency(us) 00:29:32.075 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:32.075 Job: Malloc0 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.075 Malloc0 : 2.04 12035.67 11.75 0.00 0.00 21245.82 3789.69 32824.99 00:29:32.075 Job: Malloc1 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.075 Malloc1 : 2.04 12024.50 11.74 0.00 0.00 21248.90 4616.01 32824.99 00:29:32.075 Job: Malloc0 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.075 Malloc0 : 2.05 12013.70 11.73 0.00 0.00 21192.95 3732.70 28949.82 00:29:32.075 Job: Malloc1 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.075 Malloc1 : 2.05 12002.56 11.72 0.00 0.00 21190.85 4587.52 28949.82 00:29:32.075 Job: Malloc0 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.075 Malloc0 : 2.05 11991.78 11.71 0.00 0.00 21131.94 3732.70 25188.62 00:29:32.075 Job: Malloc1 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.075 Malloc1 : 2.05 11980.79 11.70 0.00 0.00 21131.98 4587.52 25188.62 00:29:32.075 Job: Malloc0 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.075 Malloc0 : 2.05 11969.98 11.69 0.00 0.00 21075.30 3732.70 21541.40 00:29:32.075 Job: Malloc1 (Core Mask 0xff, workload: rw, percentage: 70, depth: 256, IO size: 1024) 00:29:32.075 Malloc1 : 2.06 11958.98 11.68 0.00 0.00 21075.34 4587.52 21655.37 00:29:32.075 =================================================================================================================== 00:29:32.075 Total : 95977.95 93.73 0.00 0.00 21161.63 3732.70 32824.99' 00:29:32.075 11:40:15 bdevperf_config -- bdevperf/common.sh@32 -- # grep -oE 'Using job config with [0-9]+ jobs' 00:29:32.075 11:40:15 bdevperf_config -- bdevperf/common.sh@32 -- # grep -oE '[0-9]+' 00:29:32.075 11:40:15 bdevperf_config -- bdevperf/test_config.sh@43 -- # [[ 4 == \4 ]] 00:29:32.075 11:40:15 bdevperf_config -- bdevperf/test_config.sh@44 -- # cleanup 00:29:32.075 11:40:15 bdevperf_config -- bdevperf/common.sh@36 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevperf/test.conf 00:29:32.075 11:40:15 bdevperf_config -- bdevperf/test_config.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:29:32.075 00:29:32.075 real 0m11.265s 00:29:32.075 user 0m9.909s 00:29:32.075 sys 0m1.211s 00:29:32.075 11:40:15 bdevperf_config -- common/autotest_common.sh@1124 -- # xtrace_disable 00:29:32.075 11:40:15 bdevperf_config -- common/autotest_common.sh@10 -- # set +x 00:29:32.075 ************************************ 00:29:32.075 END TEST bdevperf_config 00:29:32.075 ************************************ 00:29:32.075 11:40:15 -- common/autotest_common.sh@1142 -- # return 0 00:29:32.075 11:40:15 -- spdk/autotest.sh@192 -- # uname -s 00:29:32.075 11:40:15 -- spdk/autotest.sh@192 -- # [[ Linux == Linux ]] 00:29:32.075 11:40:15 -- spdk/autotest.sh@193 -- # run_test reactor_set_interrupt /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/reactor_set_interrupt.sh 00:29:32.075 11:40:15 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:29:32.075 11:40:15 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:29:32.075 11:40:15 -- common/autotest_common.sh@10 -- # set +x 00:29:32.075 ************************************ 00:29:32.075 START TEST reactor_set_interrupt 00:29:32.075 ************************************ 00:29:32.075 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/reactor_set_interrupt.sh 00:29:32.075 * Looking for test storage... 00:29:32.075 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt 00:29:32.075 11:40:15 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@9 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/interrupt_common.sh 00:29:32.075 11:40:15 reactor_set_interrupt -- interrupt/interrupt_common.sh@5 -- # dirname /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/reactor_set_interrupt.sh 00:29:32.075 11:40:15 reactor_set_interrupt -- interrupt/interrupt_common.sh@5 -- # readlink -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt 00:29:32.075 11:40:15 reactor_set_interrupt -- interrupt/interrupt_common.sh@5 -- # testdir=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt 00:29:32.075 11:40:15 reactor_set_interrupt -- interrupt/interrupt_common.sh@6 -- # readlink -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/../.. 00:29:32.075 11:40:15 reactor_set_interrupt -- interrupt/interrupt_common.sh@6 -- # rootdir=/var/jenkins/workspace/crypto-phy-autotest/spdk 00:29:32.075 11:40:15 reactor_set_interrupt -- interrupt/interrupt_common.sh@7 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/common/autotest_common.sh 00:29:32.075 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@7 -- # rpc_py=rpc_cmd 00:29:32.075 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@34 -- # set -e 00:29:32.075 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@35 -- # shopt -s nullglob 00:29:32.075 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@36 -- # shopt -s extglob 00:29:32.075 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@37 -- # shopt -s inherit_errexit 00:29:32.075 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@39 -- # '[' -z /var/jenkins/workspace/crypto-phy-autotest/spdk/../output ']' 00:29:32.075 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@44 -- # [[ -e /var/jenkins/workspace/crypto-phy-autotest/spdk/test/common/build_config.sh ]] 00:29:32.075 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@45 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/common/build_config.sh 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@1 -- # CONFIG_WPDK_DIR= 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@2 -- # CONFIG_ASAN=n 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@3 -- # CONFIG_VBDEV_COMPRESS=y 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@4 -- # CONFIG_HAVE_EXECINFO_H=y 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@5 -- # CONFIG_USDT=n 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@6 -- # CONFIG_CUSTOMOCF=n 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@7 -- # CONFIG_PREFIX=/usr/local 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@8 -- # CONFIG_RBD=n 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@9 -- # CONFIG_LIBDIR= 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@10 -- # CONFIG_IDXD=y 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@11 -- # CONFIG_NVME_CUSE=y 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@12 -- # CONFIG_SMA=n 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@13 -- # CONFIG_VTUNE=n 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@14 -- # CONFIG_TSAN=n 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@15 -- # CONFIG_RDMA_SEND_WITH_INVAL=y 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@16 -- # CONFIG_VFIO_USER_DIR= 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@17 -- # CONFIG_PGO_CAPTURE=n 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@18 -- # CONFIG_HAVE_UUID_GENERATE_SHA1=y 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@19 -- # CONFIG_ENV=/var/jenkins/workspace/crypto-phy-autotest/spdk/lib/env_dpdk 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@20 -- # CONFIG_LTO=n 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@21 -- # CONFIG_ISCSI_INITIATOR=y 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@22 -- # CONFIG_CET=n 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@23 -- # CONFIG_VBDEV_COMPRESS_MLX5=y 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@24 -- # CONFIG_OCF_PATH= 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@25 -- # CONFIG_RDMA_SET_TOS=y 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@26 -- # CONFIG_HAVE_ARC4RANDOM=y 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@27 -- # CONFIG_HAVE_LIBARCHIVE=n 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@28 -- # CONFIG_UBLK=y 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@29 -- # CONFIG_ISAL_CRYPTO=y 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@30 -- # CONFIG_OPENSSL_PATH= 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@31 -- # CONFIG_OCF=n 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@32 -- # CONFIG_FUSE=n 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@33 -- # CONFIG_VTUNE_DIR= 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@34 -- # CONFIG_FUZZER_LIB= 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@35 -- # CONFIG_FUZZER=n 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@36 -- # CONFIG_DPDK_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@37 -- # CONFIG_CRYPTO=y 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@38 -- # CONFIG_PGO_USE=n 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@39 -- # CONFIG_VHOST=y 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@40 -- # CONFIG_DAOS=n 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@41 -- # CONFIG_DPDK_INC_DIR= 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@42 -- # CONFIG_DAOS_DIR= 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@43 -- # CONFIG_UNIT_TESTS=n 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@44 -- # CONFIG_RDMA_SET_ACK_TIMEOUT=y 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@45 -- # CONFIG_VIRTIO=y 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@46 -- # CONFIG_DPDK_UADK=n 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@47 -- # CONFIG_COVERAGE=y 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@48 -- # CONFIG_RDMA=y 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@49 -- # CONFIG_FIO_SOURCE_DIR=/usr/src/fio 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@50 -- # CONFIG_URING_PATH= 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@51 -- # CONFIG_XNVME=n 00:29:32.075 11:40:15 reactor_set_interrupt -- common/build_config.sh@52 -- # CONFIG_VFIO_USER=n 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@53 -- # CONFIG_ARCH=native 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@54 -- # CONFIG_HAVE_EVP_MAC=y 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@55 -- # CONFIG_URING_ZNS=n 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@56 -- # CONFIG_WERROR=y 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@57 -- # CONFIG_HAVE_LIBBSD=n 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@58 -- # CONFIG_UBSAN=y 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@59 -- # CONFIG_IPSEC_MB_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/intel-ipsec-mb/lib 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@60 -- # CONFIG_GOLANG=n 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@61 -- # CONFIG_ISAL=y 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@62 -- # CONFIG_IDXD_KERNEL=y 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@63 -- # CONFIG_DPDK_LIB_DIR= 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@64 -- # CONFIG_RDMA_PROV=verbs 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@65 -- # CONFIG_APPS=y 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@66 -- # CONFIG_SHARED=y 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@67 -- # CONFIG_HAVE_KEYUTILS=y 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@68 -- # CONFIG_FC_PATH= 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@69 -- # CONFIG_DPDK_PKG_CONFIG=n 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@70 -- # CONFIG_FC=n 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@71 -- # CONFIG_AVAHI=n 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@72 -- # CONFIG_FIO_PLUGIN=y 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@73 -- # CONFIG_RAID5F=n 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@74 -- # CONFIG_EXAMPLES=y 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@75 -- # CONFIG_TESTS=y 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@76 -- # CONFIG_CRYPTO_MLX5=y 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@77 -- # CONFIG_MAX_LCORES=128 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@78 -- # CONFIG_IPSEC_MB=y 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@79 -- # CONFIG_PGO_DIR= 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@80 -- # CONFIG_DEBUG=y 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@81 -- # CONFIG_DPDK_COMPRESSDEV=y 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@82 -- # CONFIG_CROSS_PREFIX= 00:29:32.076 11:40:15 reactor_set_interrupt -- common/build_config.sh@83 -- # CONFIG_URING=n 00:29:32.076 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@54 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/common/applications.sh 00:29:32.076 11:40:15 reactor_set_interrupt -- common/applications.sh@8 -- # dirname /var/jenkins/workspace/crypto-phy-autotest/spdk/test/common/applications.sh 00:29:32.076 11:40:15 reactor_set_interrupt -- common/applications.sh@8 -- # readlink -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/common 00:29:32.076 11:40:15 reactor_set_interrupt -- common/applications.sh@8 -- # _root=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/common 00:29:32.076 11:40:15 reactor_set_interrupt -- common/applications.sh@9 -- # _root=/var/jenkins/workspace/crypto-phy-autotest/spdk 00:29:32.076 11:40:15 reactor_set_interrupt -- common/applications.sh@10 -- # _app_dir=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin 00:29:32.076 11:40:15 reactor_set_interrupt -- common/applications.sh@11 -- # _test_app_dir=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/app 00:29:32.076 11:40:15 reactor_set_interrupt -- common/applications.sh@12 -- # _examples_dir=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples 00:29:32.076 11:40:15 reactor_set_interrupt -- common/applications.sh@14 -- # VHOST_FUZZ_APP=("$_test_app_dir/fuzz/vhost_fuzz/vhost_fuzz") 00:29:32.076 11:40:15 reactor_set_interrupt -- common/applications.sh@15 -- # ISCSI_APP=("$_app_dir/iscsi_tgt") 00:29:32.076 11:40:15 reactor_set_interrupt -- common/applications.sh@16 -- # NVMF_APP=("$_app_dir/nvmf_tgt") 00:29:32.076 11:40:15 reactor_set_interrupt -- common/applications.sh@17 -- # VHOST_APP=("$_app_dir/vhost") 00:29:32.076 11:40:15 reactor_set_interrupt -- common/applications.sh@18 -- # DD_APP=("$_app_dir/spdk_dd") 00:29:32.076 11:40:15 reactor_set_interrupt -- common/applications.sh@19 -- # SPDK_APP=("$_app_dir/spdk_tgt") 00:29:32.076 11:40:15 reactor_set_interrupt -- common/applications.sh@22 -- # [[ -e /var/jenkins/workspace/crypto-phy-autotest/spdk/include/spdk/config.h ]] 00:29:32.076 11:40:15 reactor_set_interrupt -- common/applications.sh@23 -- # [[ #ifndef SPDK_CONFIG_H 00:29:32.076 #define SPDK_CONFIG_H 00:29:32.076 #define SPDK_CONFIG_APPS 1 00:29:32.076 #define SPDK_CONFIG_ARCH native 00:29:32.076 #undef SPDK_CONFIG_ASAN 00:29:32.076 #undef SPDK_CONFIG_AVAHI 00:29:32.076 #undef SPDK_CONFIG_CET 00:29:32.076 #define SPDK_CONFIG_COVERAGE 1 00:29:32.076 #define SPDK_CONFIG_CROSS_PREFIX 00:29:32.076 #define SPDK_CONFIG_CRYPTO 1 00:29:32.076 #define SPDK_CONFIG_CRYPTO_MLX5 1 00:29:32.076 #undef SPDK_CONFIG_CUSTOMOCF 00:29:32.076 #undef SPDK_CONFIG_DAOS 00:29:32.076 #define SPDK_CONFIG_DAOS_DIR 00:29:32.076 #define SPDK_CONFIG_DEBUG 1 00:29:32.076 #define SPDK_CONFIG_DPDK_COMPRESSDEV 1 00:29:32.076 #define SPDK_CONFIG_DPDK_DIR /var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build 00:29:32.076 #define SPDK_CONFIG_DPDK_INC_DIR 00:29:32.076 #define SPDK_CONFIG_DPDK_LIB_DIR 00:29:32.076 #undef SPDK_CONFIG_DPDK_PKG_CONFIG 00:29:32.076 #undef SPDK_CONFIG_DPDK_UADK 00:29:32.076 #define SPDK_CONFIG_ENV /var/jenkins/workspace/crypto-phy-autotest/spdk/lib/env_dpdk 00:29:32.076 #define SPDK_CONFIG_EXAMPLES 1 00:29:32.076 #undef SPDK_CONFIG_FC 00:29:32.076 #define SPDK_CONFIG_FC_PATH 00:29:32.076 #define SPDK_CONFIG_FIO_PLUGIN 1 00:29:32.076 #define SPDK_CONFIG_FIO_SOURCE_DIR /usr/src/fio 00:29:32.076 #undef SPDK_CONFIG_FUSE 00:29:32.076 #undef SPDK_CONFIG_FUZZER 00:29:32.076 #define SPDK_CONFIG_FUZZER_LIB 00:29:32.076 #undef SPDK_CONFIG_GOLANG 00:29:32.076 #define SPDK_CONFIG_HAVE_ARC4RANDOM 1 00:29:32.076 #define SPDK_CONFIG_HAVE_EVP_MAC 1 00:29:32.076 #define SPDK_CONFIG_HAVE_EXECINFO_H 1 00:29:32.076 #define SPDK_CONFIG_HAVE_KEYUTILS 1 00:29:32.076 #undef SPDK_CONFIG_HAVE_LIBARCHIVE 00:29:32.076 #undef SPDK_CONFIG_HAVE_LIBBSD 00:29:32.076 #define SPDK_CONFIG_HAVE_UUID_GENERATE_SHA1 1 00:29:32.076 #define SPDK_CONFIG_IDXD 1 00:29:32.076 #define SPDK_CONFIG_IDXD_KERNEL 1 00:29:32.076 #define SPDK_CONFIG_IPSEC_MB 1 00:29:32.076 #define SPDK_CONFIG_IPSEC_MB_DIR /var/jenkins/workspace/crypto-phy-autotest/spdk/intel-ipsec-mb/lib 00:29:32.076 #define SPDK_CONFIG_ISAL 1 00:29:32.076 #define SPDK_CONFIG_ISAL_CRYPTO 1 00:29:32.076 #define SPDK_CONFIG_ISCSI_INITIATOR 1 00:29:32.076 #define SPDK_CONFIG_LIBDIR 00:29:32.076 #undef SPDK_CONFIG_LTO 00:29:32.076 #define SPDK_CONFIG_MAX_LCORES 128 00:29:32.076 #define SPDK_CONFIG_NVME_CUSE 1 00:29:32.076 #undef SPDK_CONFIG_OCF 00:29:32.076 #define SPDK_CONFIG_OCF_PATH 00:29:32.076 #define SPDK_CONFIG_OPENSSL_PATH 00:29:32.076 #undef SPDK_CONFIG_PGO_CAPTURE 00:29:32.076 #define SPDK_CONFIG_PGO_DIR 00:29:32.076 #undef SPDK_CONFIG_PGO_USE 00:29:32.076 #define SPDK_CONFIG_PREFIX /usr/local 00:29:32.076 #undef SPDK_CONFIG_RAID5F 00:29:32.076 #undef SPDK_CONFIG_RBD 00:29:32.076 #define SPDK_CONFIG_RDMA 1 00:29:32.076 #define SPDK_CONFIG_RDMA_PROV verbs 00:29:32.076 #define SPDK_CONFIG_RDMA_SEND_WITH_INVAL 1 00:29:32.076 #define SPDK_CONFIG_RDMA_SET_ACK_TIMEOUT 1 00:29:32.076 #define SPDK_CONFIG_RDMA_SET_TOS 1 00:29:32.076 #define SPDK_CONFIG_SHARED 1 00:29:32.076 #undef SPDK_CONFIG_SMA 00:29:32.076 #define SPDK_CONFIG_TESTS 1 00:29:32.076 #undef SPDK_CONFIG_TSAN 00:29:32.076 #define SPDK_CONFIG_UBLK 1 00:29:32.076 #define SPDK_CONFIG_UBSAN 1 00:29:32.076 #undef SPDK_CONFIG_UNIT_TESTS 00:29:32.076 #undef SPDK_CONFIG_URING 00:29:32.076 #define SPDK_CONFIG_URING_PATH 00:29:32.076 #undef SPDK_CONFIG_URING_ZNS 00:29:32.076 #undef SPDK_CONFIG_USDT 00:29:32.076 #define SPDK_CONFIG_VBDEV_COMPRESS 1 00:29:32.076 #define SPDK_CONFIG_VBDEV_COMPRESS_MLX5 1 00:29:32.076 #undef SPDK_CONFIG_VFIO_USER 00:29:32.076 #define SPDK_CONFIG_VFIO_USER_DIR 00:29:32.076 #define SPDK_CONFIG_VHOST 1 00:29:32.076 #define SPDK_CONFIG_VIRTIO 1 00:29:32.076 #undef SPDK_CONFIG_VTUNE 00:29:32.076 #define SPDK_CONFIG_VTUNE_DIR 00:29:32.076 #define SPDK_CONFIG_WERROR 1 00:29:32.076 #define SPDK_CONFIG_WPDK_DIR 00:29:32.076 #undef SPDK_CONFIG_XNVME 00:29:32.076 #endif /* SPDK_CONFIG_H */ == *\#\d\e\f\i\n\e\ \S\P\D\K\_\C\O\N\F\I\G\_\D\E\B\U\G* ]] 00:29:32.076 11:40:15 reactor_set_interrupt -- common/applications.sh@24 -- # (( SPDK_AUTOTEST_DEBUG_APPS )) 00:29:32.076 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@55 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/common.sh 00:29:32.076 11:40:15 reactor_set_interrupt -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:29:32.076 11:40:15 reactor_set_interrupt -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:29:32.076 11:40:15 reactor_set_interrupt -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:29:32.076 11:40:15 reactor_set_interrupt -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:32.076 11:40:15 reactor_set_interrupt -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:32.076 11:40:15 reactor_set_interrupt -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:32.076 11:40:15 reactor_set_interrupt -- paths/export.sh@5 -- # export PATH 00:29:32.076 11:40:15 reactor_set_interrupt -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:32.076 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@56 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm/common 00:29:32.076 11:40:15 reactor_set_interrupt -- pm/common@6 -- # dirname /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm/common 00:29:32.076 11:40:15 reactor_set_interrupt -- pm/common@6 -- # readlink -f /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm 00:29:32.076 11:40:15 reactor_set_interrupt -- pm/common@6 -- # _pmdir=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm 00:29:32.076 11:40:15 reactor_set_interrupt -- pm/common@7 -- # readlink -f /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm/../../../ 00:29:32.076 11:40:15 reactor_set_interrupt -- pm/common@7 -- # _pmrootdir=/var/jenkins/workspace/crypto-phy-autotest/spdk 00:29:32.076 11:40:15 reactor_set_interrupt -- pm/common@64 -- # TEST_TAG=N/A 00:29:32.076 11:40:15 reactor_set_interrupt -- pm/common@65 -- # TEST_TAG_FILE=/var/jenkins/workspace/crypto-phy-autotest/spdk/.run_test_name 00:29:32.077 11:40:15 reactor_set_interrupt -- pm/common@67 -- # PM_OUTPUTDIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power 00:29:32.077 11:40:15 reactor_set_interrupt -- pm/common@68 -- # uname -s 00:29:32.077 11:40:15 reactor_set_interrupt -- pm/common@68 -- # PM_OS=Linux 00:29:32.077 11:40:15 reactor_set_interrupt -- pm/common@70 -- # MONITOR_RESOURCES_SUDO=() 00:29:32.077 11:40:15 reactor_set_interrupt -- pm/common@70 -- # declare -A MONITOR_RESOURCES_SUDO 00:29:32.077 11:40:15 reactor_set_interrupt -- pm/common@71 -- # MONITOR_RESOURCES_SUDO["collect-bmc-pm"]=1 00:29:32.077 11:40:15 reactor_set_interrupt -- pm/common@72 -- # MONITOR_RESOURCES_SUDO["collect-cpu-load"]=0 00:29:32.077 11:40:15 reactor_set_interrupt -- pm/common@73 -- # MONITOR_RESOURCES_SUDO["collect-cpu-temp"]=0 00:29:32.077 11:40:15 reactor_set_interrupt -- pm/common@74 -- # MONITOR_RESOURCES_SUDO["collect-vmstat"]=0 00:29:32.077 11:40:15 reactor_set_interrupt -- pm/common@76 -- # SUDO[0]= 00:29:32.077 11:40:15 reactor_set_interrupt -- pm/common@76 -- # SUDO[1]='sudo -E' 00:29:32.077 11:40:15 reactor_set_interrupt -- pm/common@78 -- # MONITOR_RESOURCES=(collect-cpu-load collect-vmstat) 00:29:32.077 11:40:15 reactor_set_interrupt -- pm/common@79 -- # [[ Linux == FreeBSD ]] 00:29:32.077 11:40:15 reactor_set_interrupt -- pm/common@81 -- # [[ Linux == Linux ]] 00:29:32.077 11:40:15 reactor_set_interrupt -- pm/common@81 -- # [[ ............................... != QEMU ]] 00:29:32.077 11:40:15 reactor_set_interrupt -- pm/common@81 -- # [[ ! -e /.dockerenv ]] 00:29:32.077 11:40:15 reactor_set_interrupt -- pm/common@84 -- # MONITOR_RESOURCES+=(collect-cpu-temp) 00:29:32.077 11:40:15 reactor_set_interrupt -- pm/common@85 -- # MONITOR_RESOURCES+=(collect-bmc-pm) 00:29:32.077 11:40:15 reactor_set_interrupt -- pm/common@88 -- # [[ ! -d /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power ]] 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@58 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@59 -- # export RUN_NIGHTLY 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@62 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@63 -- # export SPDK_AUTOTEST_DEBUG_APPS 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@64 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@65 -- # export SPDK_RUN_VALGRIND 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@66 -- # : 1 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@67 -- # export SPDK_RUN_FUNCTIONAL_TEST 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@68 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@69 -- # export SPDK_TEST_UNITTEST 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@70 -- # : 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@71 -- # export SPDK_TEST_AUTOBUILD 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@72 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@73 -- # export SPDK_TEST_RELEASE_BUILD 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@74 -- # : 1 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@75 -- # export SPDK_TEST_ISAL 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@76 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@77 -- # export SPDK_TEST_ISCSI 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@78 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@79 -- # export SPDK_TEST_ISCSI_INITIATOR 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@80 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@81 -- # export SPDK_TEST_NVME 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@82 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@83 -- # export SPDK_TEST_NVME_PMR 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@84 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@85 -- # export SPDK_TEST_NVME_BP 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@86 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@87 -- # export SPDK_TEST_NVME_CLI 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@88 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@89 -- # export SPDK_TEST_NVME_CUSE 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@90 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@91 -- # export SPDK_TEST_NVME_FDP 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@92 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@93 -- # export SPDK_TEST_NVMF 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@94 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@95 -- # export SPDK_TEST_VFIOUSER 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@96 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@97 -- # export SPDK_TEST_VFIOUSER_QEMU 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@98 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@99 -- # export SPDK_TEST_FUZZER 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@100 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@101 -- # export SPDK_TEST_FUZZER_SHORT 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@102 -- # : rdma 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@103 -- # export SPDK_TEST_NVMF_TRANSPORT 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@104 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@105 -- # export SPDK_TEST_RBD 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@106 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@107 -- # export SPDK_TEST_VHOST 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@108 -- # : 1 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@109 -- # export SPDK_TEST_BLOCKDEV 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@110 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@111 -- # export SPDK_TEST_IOAT 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@112 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@113 -- # export SPDK_TEST_BLOBFS 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@114 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@115 -- # export SPDK_TEST_VHOST_INIT 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@116 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@117 -- # export SPDK_TEST_LVOL 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@118 -- # : 1 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@119 -- # export SPDK_TEST_VBDEV_COMPRESS 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@120 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@121 -- # export SPDK_RUN_ASAN 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@122 -- # : 1 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@123 -- # export SPDK_RUN_UBSAN 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@124 -- # : 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@125 -- # export SPDK_RUN_EXTERNAL_DPDK 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@126 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@127 -- # export SPDK_RUN_NON_ROOT 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@128 -- # : 1 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@129 -- # export SPDK_TEST_CRYPTO 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@130 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@131 -- # export SPDK_TEST_FTL 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@132 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@133 -- # export SPDK_TEST_OCF 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@134 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@135 -- # export SPDK_TEST_VMD 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@136 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@137 -- # export SPDK_TEST_OPAL 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@138 -- # : 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@139 -- # export SPDK_TEST_NATIVE_DPDK 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@140 -- # : true 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@141 -- # export SPDK_AUTOTEST_X 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@142 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@143 -- # export SPDK_TEST_RAID5 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@144 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@145 -- # export SPDK_TEST_URING 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@146 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@147 -- # export SPDK_TEST_USDT 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@148 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@149 -- # export SPDK_TEST_USE_IGB_UIO 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@150 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@151 -- # export SPDK_TEST_SCHEDULER 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@152 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@153 -- # export SPDK_TEST_SCANBUILD 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@154 -- # : 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@155 -- # export SPDK_TEST_NVMF_NICS 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@156 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@157 -- # export SPDK_TEST_SMA 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@158 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@159 -- # export SPDK_TEST_DAOS 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@160 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@161 -- # export SPDK_TEST_XNVME 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@162 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@163 -- # export SPDK_TEST_ACCEL_DSA 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@164 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@165 -- # export SPDK_TEST_ACCEL_IAA 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@167 -- # : 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@168 -- # export SPDK_TEST_FUZZER_TARGET 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@169 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@170 -- # export SPDK_TEST_NVMF_MDNS 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@171 -- # : 0 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@172 -- # export SPDK_JSONRPC_GO_CLIENT 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@175 -- # export SPDK_LIB_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/lib 00:29:32.077 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@175 -- # SPDK_LIB_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/lib 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@176 -- # export DPDK_LIB_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build/lib 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@176 -- # DPDK_LIB_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build/lib 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@177 -- # export VFIO_LIB_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@177 -- # VFIO_LIB_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@178 -- # export LD_LIBRARY_PATH=:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@178 -- # LD_LIBRARY_PATH=:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@181 -- # export PCI_BLOCK_SYNC_ON_RESET=yes 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@181 -- # PCI_BLOCK_SYNC_ON_RESET=yes 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@185 -- # export PYTHONPATH=:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/python 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@185 -- # PYTHONPATH=:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/python 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@189 -- # export PYTHONDONTWRITEBYTECODE=1 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@189 -- # PYTHONDONTWRITEBYTECODE=1 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@193 -- # export ASAN_OPTIONS=new_delete_type_mismatch=0:disable_coredump=0:abort_on_error=1:use_sigaltstack=0 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@193 -- # ASAN_OPTIONS=new_delete_type_mismatch=0:disable_coredump=0:abort_on_error=1:use_sigaltstack=0 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@194 -- # export UBSAN_OPTIONS=halt_on_error=1:print_stacktrace=1:abort_on_error=1:disable_coredump=0:exitcode=134 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@194 -- # UBSAN_OPTIONS=halt_on_error=1:print_stacktrace=1:abort_on_error=1:disable_coredump=0:exitcode=134 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@198 -- # asan_suppression_file=/var/tmp/asan_suppression_file 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@199 -- # rm -rf /var/tmp/asan_suppression_file 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@200 -- # cat 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@236 -- # echo leak:libfuse3.so 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@238 -- # export LSAN_OPTIONS=suppressions=/var/tmp/asan_suppression_file 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@238 -- # LSAN_OPTIONS=suppressions=/var/tmp/asan_suppression_file 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@240 -- # export DEFAULT_RPC_ADDR=/var/tmp/spdk.sock 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@240 -- # DEFAULT_RPC_ADDR=/var/tmp/spdk.sock 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@242 -- # '[' -z /var/spdk/dependencies ']' 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@245 -- # export DEPENDENCY_DIR 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@249 -- # export SPDK_BIN_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@249 -- # SPDK_BIN_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@250 -- # export SPDK_EXAMPLE_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@250 -- # SPDK_EXAMPLE_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@253 -- # export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@253 -- # QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@254 -- # export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@254 -- # VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@256 -- # export AR_TOOL=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/ar-xnvme-fixer 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@256 -- # AR_TOOL=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/ar-xnvme-fixer 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@259 -- # export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@259 -- # UNBIND_ENTIRE_IOMMU_GROUP=yes 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@262 -- # '[' 0 -eq 0 ']' 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@263 -- # export valgrind= 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@263 -- # valgrind= 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@269 -- # uname -s 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@269 -- # '[' Linux = Linux ']' 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@270 -- # HUGEMEM=4096 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@271 -- # export CLEAR_HUGE=yes 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@271 -- # CLEAR_HUGE=yes 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@272 -- # [[ 1 -eq 1 ]] 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@276 -- # export HUGE_EVEN_ALLOC=yes 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@276 -- # HUGE_EVEN_ALLOC=yes 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@279 -- # MAKE=make 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@280 -- # MAKEFLAGS=-j72 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@296 -- # export HUGEMEM=4096 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@296 -- # HUGEMEM=4096 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@298 -- # NO_HUGE=() 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@299 -- # TEST_MODE= 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@318 -- # [[ -z 1027639 ]] 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@318 -- # kill -0 1027639 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@1680 -- # set_test_storage 2147483648 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@328 -- # [[ -v testdir ]] 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@330 -- # local requested_size=2147483648 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@331 -- # local mount target_dir 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@333 -- # local -A mounts fss sizes avails uses 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@334 -- # local source fs size avail mount use 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@336 -- # local storage_fallback storage_candidates 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@338 -- # mktemp -udt spdk.XXXXXX 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@338 -- # storage_fallback=/tmp/spdk.cxQj94 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@343 -- # storage_candidates=("$testdir" "$storage_fallback/tests/${testdir##*/}" "$storage_fallback") 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@345 -- # [[ -n '' ]] 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@350 -- # [[ -n '' ]] 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@355 -- # mkdir -p /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt /tmp/spdk.cxQj94/tests/interrupt /tmp/spdk.cxQj94 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@358 -- # requested_size=2214592512 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@327 -- # df -T 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@327 -- # grep -v Filesystem 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@361 -- # mounts["$mount"]=spdk_devtmpfs 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@361 -- # fss["$mount"]=devtmpfs 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@362 -- # avails["$mount"]=67108864 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@362 -- # sizes["$mount"]=67108864 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@363 -- # uses["$mount"]=0 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:29:32.078 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@361 -- # mounts["$mount"]=/dev/pmem0 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@361 -- # fss["$mount"]=ext2 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@362 -- # avails["$mount"]=946290688 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@362 -- # sizes["$mount"]=5284429824 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@363 -- # uses["$mount"]=4338139136 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@361 -- # mounts["$mount"]=spdk_root 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@361 -- # fss["$mount"]=overlay 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@362 -- # avails["$mount"]=88761561088 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@362 -- # sizes["$mount"]=94508515328 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@363 -- # uses["$mount"]=5746954240 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@361 -- # mounts["$mount"]=tmpfs 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@361 -- # fss["$mount"]=tmpfs 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@362 -- # avails["$mount"]=47249547264 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@362 -- # sizes["$mount"]=47254257664 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@363 -- # uses["$mount"]=4710400 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@361 -- # mounts["$mount"]=tmpfs 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@361 -- # fss["$mount"]=tmpfs 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@362 -- # avails["$mount"]=18892267520 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@362 -- # sizes["$mount"]=18901704704 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@363 -- # uses["$mount"]=9437184 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@361 -- # mounts["$mount"]=tmpfs 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@361 -- # fss["$mount"]=tmpfs 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@362 -- # avails["$mount"]=47253434368 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@362 -- # sizes["$mount"]=47254257664 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@363 -- # uses["$mount"]=823296 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@361 -- # mounts["$mount"]=tmpfs 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@361 -- # fss["$mount"]=tmpfs 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@362 -- # avails["$mount"]=9450844160 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@362 -- # sizes["$mount"]=9450848256 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@363 -- # uses["$mount"]=4096 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@366 -- # printf '* Looking for test storage...\n' 00:29:32.079 * Looking for test storage... 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@368 -- # local target_space new_size 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@369 -- # for target_dir in "${storage_candidates[@]}" 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@372 -- # df /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@372 -- # awk '$1 !~ /Filesystem/{print $6}' 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@372 -- # mount=/ 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@374 -- # target_space=88761561088 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@375 -- # (( target_space == 0 || target_space < requested_size )) 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@378 -- # (( target_space >= requested_size )) 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@380 -- # [[ overlay == tmpfs ]] 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@380 -- # [[ overlay == ramfs ]] 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@380 -- # [[ / == / ]] 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@381 -- # new_size=7961546752 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@382 -- # (( new_size * 100 / sizes[/] > 95 )) 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@387 -- # export SPDK_TEST_STORAGE=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@387 -- # SPDK_TEST_STORAGE=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@388 -- # printf '* Found test storage at %s\n' /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt 00:29:32.079 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@389 -- # return 0 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@1682 -- # set -o errtrace 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@1683 -- # shopt -s extdebug 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@1684 -- # trap 'trap - ERR; print_backtrace >&2' ERR 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@1686 -- # PS4=' \t ${test_domain:-} -- ${BASH_SOURCE#${BASH_SOURCE%/*/*}/}@${LINENO} -- \$ ' 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@1687 -- # true 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@1689 -- # xtrace_fd 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@25 -- # [[ -n 13 ]] 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@25 -- # [[ -e /proc/self/fd/13 ]] 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@27 -- # exec 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@29 -- # exec 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@31 -- # xtrace_restore 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@16 -- # unset -v 'X_STACK[0 - 1 < 0 ? 0 : 0 - 1]' 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@17 -- # (( 0 == 0 )) 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@18 -- # set -x 00:29:32.079 11:40:15 reactor_set_interrupt -- interrupt/interrupt_common.sh@8 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/common.sh 00:29:32.079 11:40:15 reactor_set_interrupt -- interrupt/interrupt_common.sh@10 -- # rpc_py=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:29:32.079 11:40:15 reactor_set_interrupt -- interrupt/interrupt_common.sh@12 -- # r0_mask=0x1 00:29:32.079 11:40:15 reactor_set_interrupt -- interrupt/interrupt_common.sh@13 -- # r1_mask=0x2 00:29:32.079 11:40:15 reactor_set_interrupt -- interrupt/interrupt_common.sh@14 -- # r2_mask=0x4 00:29:32.079 11:40:15 reactor_set_interrupt -- interrupt/interrupt_common.sh@16 -- # cpu_server_mask=0x07 00:29:32.079 11:40:15 reactor_set_interrupt -- interrupt/interrupt_common.sh@17 -- # rpc_server_addr=/var/tmp/spdk.sock 00:29:32.079 11:40:15 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@11 -- # export PYTHONPATH=:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/examples/interrupt_tgt 00:29:32.079 11:40:15 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@11 -- # PYTHONPATH=:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/examples/interrupt_tgt 00:29:32.079 11:40:15 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@86 -- # start_intr_tgt 00:29:32.079 11:40:15 reactor_set_interrupt -- interrupt/interrupt_common.sh@20 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:32.079 11:40:15 reactor_set_interrupt -- interrupt/interrupt_common.sh@21 -- # local cpu_mask=0x07 00:29:32.079 11:40:15 reactor_set_interrupt -- interrupt/interrupt_common.sh@24 -- # intr_tgt_pid=1027769 00:29:32.079 11:40:15 reactor_set_interrupt -- interrupt/interrupt_common.sh@23 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/interrupt_tgt -m 0x07 -r /var/tmp/spdk.sock -E -g 00:29:32.079 11:40:15 reactor_set_interrupt -- interrupt/interrupt_common.sh@25 -- # trap 'killprocess "$intr_tgt_pid"; cleanup; exit 1' SIGINT SIGTERM EXIT 00:29:32.079 11:40:15 reactor_set_interrupt -- interrupt/interrupt_common.sh@26 -- # waitforlisten 1027769 /var/tmp/spdk.sock 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@829 -- # '[' -z 1027769 ']' 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@834 -- # local max_retries=100 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:32.079 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@838 -- # xtrace_disable 00:29:32.079 11:40:15 reactor_set_interrupt -- common/autotest_common.sh@10 -- # set +x 00:29:32.079 [2024-07-15 11:40:15.512356] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:29:32.079 [2024-07-15 11:40:15.512421] [ DPDK EAL parameters: interrupt_tgt --no-shconf -c 0x07 --single-file-segments --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1027769 ] 00:29:32.079 [2024-07-15 11:40:15.639446] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:29:32.338 [2024-07-15 11:40:15.750591] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:29:32.338 [2024-07-15 11:40:15.750677] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:29:32.338 [2024-07-15 11:40:15.750682] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:29:32.338 [2024-07-15 11:40:15.821877] thread.c:2099:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:29:32.906 11:40:16 reactor_set_interrupt -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:29:32.906 11:40:16 reactor_set_interrupt -- common/autotest_common.sh@862 -- # return 0 00:29:32.906 11:40:16 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@87 -- # setup_bdev_mem 00:29:32.906 11:40:16 reactor_set_interrupt -- interrupt/common.sh@67 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:29:33.164 Malloc0 00:29:33.164 Malloc1 00:29:33.164 Malloc2 00:29:33.164 11:40:16 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@88 -- # setup_bdev_aio 00:29:33.421 11:40:16 reactor_set_interrupt -- interrupt/common.sh@75 -- # uname -s 00:29:33.421 11:40:16 reactor_set_interrupt -- interrupt/common.sh@75 -- # [[ Linux != \F\r\e\e\B\S\D ]] 00:29:33.421 11:40:16 reactor_set_interrupt -- interrupt/common.sh@76 -- # dd if=/dev/zero of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/aiofile bs=2048 count=5000 00:29:33.421 5000+0 records in 00:29:33.421 5000+0 records out 00:29:33.421 10240000 bytes (10 MB, 9.8 MiB) copied, 0.0249188 s, 411 MB/s 00:29:33.421 11:40:16 reactor_set_interrupt -- interrupt/common.sh@77 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/aiofile AIO0 2048 00:29:33.679 AIO0 00:29:33.679 11:40:17 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@90 -- # reactor_set_mode_without_threads 1027769 00:29:33.679 11:40:17 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@76 -- # reactor_set_intr_mode 1027769 without_thd 00:29:33.679 11:40:17 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@14 -- # local spdk_pid=1027769 00:29:33.679 11:40:17 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@15 -- # local without_thd=without_thd 00:29:33.679 11:40:17 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@17 -- # thd0_ids=($(reactor_get_thread_ids $r0_mask)) 00:29:33.679 11:40:17 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@17 -- # reactor_get_thread_ids 0x1 00:29:33.679 11:40:17 reactor_set_interrupt -- interrupt/common.sh@55 -- # local reactor_cpumask=0x1 00:29:33.679 11:40:17 reactor_set_interrupt -- interrupt/common.sh@56 -- # local grep_str 00:29:33.679 11:40:17 reactor_set_interrupt -- interrupt/common.sh@58 -- # reactor_cpumask=1 00:29:33.679 11:40:17 reactor_set_interrupt -- interrupt/common.sh@59 -- # jq_str='.threads|.[]|select(.cpumask == $reactor_cpumask)|.id' 00:29:33.679 11:40:17 reactor_set_interrupt -- interrupt/common.sh@62 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py thread_get_stats 00:29:33.679 11:40:17 reactor_set_interrupt -- interrupt/common.sh@62 -- # jq --arg reactor_cpumask 1 '.threads|.[]|select(.cpumask == $reactor_cpumask)|.id' 00:29:33.936 11:40:17 reactor_set_interrupt -- interrupt/common.sh@62 -- # echo 1 00:29:33.936 11:40:17 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@18 -- # thd2_ids=($(reactor_get_thread_ids $r2_mask)) 00:29:33.936 11:40:17 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@18 -- # reactor_get_thread_ids 0x4 00:29:33.936 11:40:17 reactor_set_interrupt -- interrupt/common.sh@55 -- # local reactor_cpumask=0x4 00:29:33.936 11:40:17 reactor_set_interrupt -- interrupt/common.sh@56 -- # local grep_str 00:29:33.936 11:40:17 reactor_set_interrupt -- interrupt/common.sh@58 -- # reactor_cpumask=4 00:29:33.936 11:40:17 reactor_set_interrupt -- interrupt/common.sh@59 -- # jq_str='.threads|.[]|select(.cpumask == $reactor_cpumask)|.id' 00:29:33.936 11:40:17 reactor_set_interrupt -- interrupt/common.sh@62 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py thread_get_stats 00:29:33.936 11:40:17 reactor_set_interrupt -- interrupt/common.sh@62 -- # jq --arg reactor_cpumask 4 '.threads|.[]|select(.cpumask == $reactor_cpumask)|.id' 00:29:34.194 11:40:17 reactor_set_interrupt -- interrupt/common.sh@62 -- # echo '' 00:29:34.194 11:40:17 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@21 -- # [[ 1 -eq 0 ]] 00:29:34.194 11:40:17 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@25 -- # echo 'spdk_thread ids are 1 on reactor0.' 00:29:34.194 spdk_thread ids are 1 on reactor0. 00:29:34.194 11:40:17 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@29 -- # for i in {0..2} 00:29:34.194 11:40:17 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@30 -- # reactor_is_idle 1027769 0 00:29:34.194 11:40:17 reactor_set_interrupt -- interrupt/common.sh@51 -- # reactor_is_busy_or_idle 1027769 0 idle 00:29:34.194 11:40:17 reactor_set_interrupt -- interrupt/common.sh@10 -- # local pid=1027769 00:29:34.194 11:40:17 reactor_set_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:29:34.194 11:40:17 reactor_set_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:29:34.194 11:40:17 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ idle != \b\u\s\y ]] 00:29:34.194 11:40:17 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ idle != \i\d\l\e ]] 00:29:34.194 11:40:17 reactor_set_interrupt -- interrupt/common.sh@18 -- # hash top 00:29:34.194 11:40:17 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j = 10 )) 00:29:34.194 11:40:17 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j != 0 )) 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@24 -- # grep reactor_0 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@24 -- # top -bHn 1 -p 1027769 -w 256 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@24 -- # top_reactor='1027769 root 20 0 128.2g 36864 23616 S 0.0 0.0 0:00.39 reactor_0' 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@25 -- # awk '{print $9}' 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@25 -- # echo 1027769 root 20 0 128.2g 36864 23616 S 0.0 0.0 0:00.39 reactor_0 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@25 -- # sed -e 's/^\s*//g' 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@25 -- # cpu_rate=0.0 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@26 -- # cpu_rate=0 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@28 -- # [[ idle = \b\u\s\y ]] 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ idle = \i\d\l\e ]] 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ 0 -gt 30 ]] 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@33 -- # return 0 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@29 -- # for i in {0..2} 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@30 -- # reactor_is_idle 1027769 1 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@51 -- # reactor_is_busy_or_idle 1027769 1 idle 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@10 -- # local pid=1027769 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@11 -- # local idx=1 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ idle != \b\u\s\y ]] 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ idle != \i\d\l\e ]] 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@18 -- # hash top 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j = 10 )) 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j != 0 )) 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@24 -- # grep reactor_1 00:29:34.195 11:40:17 reactor_set_interrupt -- interrupt/common.sh@24 -- # top -bHn 1 -p 1027769 -w 256 00:29:34.453 11:40:17 reactor_set_interrupt -- interrupt/common.sh@24 -- # top_reactor='1027772 root 20 0 128.2g 36864 23616 S 0.0 0.0 0:00.00 reactor_1' 00:29:34.453 11:40:17 reactor_set_interrupt -- interrupt/common.sh@25 -- # echo 1027772 root 20 0 128.2g 36864 23616 S 0.0 0.0 0:00.00 reactor_1 00:29:34.453 11:40:17 reactor_set_interrupt -- interrupt/common.sh@25 -- # sed -e 's/^\s*//g' 00:29:34.453 11:40:17 reactor_set_interrupt -- interrupt/common.sh@25 -- # awk '{print $9}' 00:29:34.454 11:40:17 reactor_set_interrupt -- interrupt/common.sh@25 -- # cpu_rate=0.0 00:29:34.454 11:40:17 reactor_set_interrupt -- interrupt/common.sh@26 -- # cpu_rate=0 00:29:34.454 11:40:17 reactor_set_interrupt -- interrupt/common.sh@28 -- # [[ idle = \b\u\s\y ]] 00:29:34.454 11:40:17 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ idle = \i\d\l\e ]] 00:29:34.454 11:40:17 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ 0 -gt 30 ]] 00:29:34.454 11:40:17 reactor_set_interrupt -- interrupt/common.sh@33 -- # return 0 00:29:34.454 11:40:17 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@29 -- # for i in {0..2} 00:29:34.454 11:40:17 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@30 -- # reactor_is_idle 1027769 2 00:29:34.454 11:40:17 reactor_set_interrupt -- interrupt/common.sh@51 -- # reactor_is_busy_or_idle 1027769 2 idle 00:29:34.454 11:40:17 reactor_set_interrupt -- interrupt/common.sh@10 -- # local pid=1027769 00:29:34.454 11:40:17 reactor_set_interrupt -- interrupt/common.sh@11 -- # local idx=2 00:29:34.454 11:40:17 reactor_set_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:29:34.454 11:40:17 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ idle != \b\u\s\y ]] 00:29:34.454 11:40:17 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ idle != \i\d\l\e ]] 00:29:34.454 11:40:17 reactor_set_interrupt -- interrupt/common.sh@18 -- # hash top 00:29:34.454 11:40:17 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j = 10 )) 00:29:34.454 11:40:17 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j != 0 )) 00:29:34.454 11:40:17 reactor_set_interrupt -- interrupt/common.sh@24 -- # top -bHn 1 -p 1027769 -w 256 00:29:34.454 11:40:17 reactor_set_interrupt -- interrupt/common.sh@24 -- # grep reactor_2 00:29:34.712 11:40:18 reactor_set_interrupt -- interrupt/common.sh@24 -- # top_reactor='1027773 root 20 0 128.2g 36864 23616 S 0.0 0.0 0:00.00 reactor_2' 00:29:34.712 11:40:18 reactor_set_interrupt -- interrupt/common.sh@25 -- # echo 1027773 root 20 0 128.2g 36864 23616 S 0.0 0.0 0:00.00 reactor_2 00:29:34.712 11:40:18 reactor_set_interrupt -- interrupt/common.sh@25 -- # sed -e 's/^\s*//g' 00:29:34.712 11:40:18 reactor_set_interrupt -- interrupt/common.sh@25 -- # awk '{print $9}' 00:29:34.712 11:40:18 reactor_set_interrupt -- interrupt/common.sh@25 -- # cpu_rate=0.0 00:29:34.712 11:40:18 reactor_set_interrupt -- interrupt/common.sh@26 -- # cpu_rate=0 00:29:34.712 11:40:18 reactor_set_interrupt -- interrupt/common.sh@28 -- # [[ idle = \b\u\s\y ]] 00:29:34.712 11:40:18 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ idle = \i\d\l\e ]] 00:29:34.712 11:40:18 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ 0 -gt 30 ]] 00:29:34.712 11:40:18 reactor_set_interrupt -- interrupt/common.sh@33 -- # return 0 00:29:34.712 11:40:18 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@33 -- # '[' without_thdx '!=' x ']' 00:29:34.712 11:40:18 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@35 -- # for i in "${thd0_ids[@]}" 00:29:34.712 11:40:18 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@36 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py thread_set_cpumask -i 1 -m 0x2 00:29:34.969 [2024-07-15 11:40:18.355645] thread.c:2099:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:29:34.969 11:40:18 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@43 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py --plugin interrupt_plugin reactor_set_interrupt_mode 0 -d 00:29:35.226 [2024-07-15 11:40:18.603370] interrupt_tgt.c: 99:rpc_reactor_set_interrupt_mode: *NOTICE*: RPC Start to disable interrupt mode on reactor 0. 00:29:35.226 [2024-07-15 11:40:18.603752] interrupt_tgt.c: 36:rpc_reactor_set_interrupt_mode_cb: *NOTICE*: complete reactor switch 00:29:35.226 11:40:18 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@44 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py --plugin interrupt_plugin reactor_set_interrupt_mode 2 -d 00:29:35.486 [2024-07-15 11:40:18.843290] interrupt_tgt.c: 99:rpc_reactor_set_interrupt_mode: *NOTICE*: RPC Start to disable interrupt mode on reactor 2. 00:29:35.486 [2024-07-15 11:40:18.843422] interrupt_tgt.c: 36:rpc_reactor_set_interrupt_mode_cb: *NOTICE*: complete reactor switch 00:29:35.486 11:40:18 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@46 -- # for i in 0 2 00:29:35.486 11:40:18 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@47 -- # reactor_is_busy 1027769 0 00:29:35.486 11:40:18 reactor_set_interrupt -- interrupt/common.sh@47 -- # reactor_is_busy_or_idle 1027769 0 busy 00:29:35.486 11:40:18 reactor_set_interrupt -- interrupt/common.sh@10 -- # local pid=1027769 00:29:35.486 11:40:18 reactor_set_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:29:35.486 11:40:18 reactor_set_interrupt -- interrupt/common.sh@12 -- # local state=busy 00:29:35.486 11:40:18 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ busy != \b\u\s\y ]] 00:29:35.486 11:40:18 reactor_set_interrupt -- interrupt/common.sh@18 -- # hash top 00:29:35.486 11:40:18 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j = 10 )) 00:29:35.486 11:40:18 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j != 0 )) 00:29:35.486 11:40:18 reactor_set_interrupt -- interrupt/common.sh@24 -- # top -bHn 1 -p 1027769 -w 256 00:29:35.486 11:40:18 reactor_set_interrupt -- interrupt/common.sh@24 -- # grep reactor_0 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/common.sh@24 -- # top_reactor='1027769 root 20 0 128.2g 36864 23616 R 99.9 0.0 0:00.83 reactor_0' 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/common.sh@25 -- # echo 1027769 root 20 0 128.2g 36864 23616 R 99.9 0.0 0:00.83 reactor_0 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/common.sh@25 -- # sed -e 's/^\s*//g' 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/common.sh@25 -- # awk '{print $9}' 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/common.sh@25 -- # cpu_rate=99.9 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/common.sh@26 -- # cpu_rate=99 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/common.sh@28 -- # [[ busy = \b\u\s\y ]] 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/common.sh@28 -- # [[ 99 -lt 70 ]] 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ busy = \i\d\l\e ]] 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/common.sh@33 -- # return 0 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@46 -- # for i in 0 2 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@47 -- # reactor_is_busy 1027769 2 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/common.sh@47 -- # reactor_is_busy_or_idle 1027769 2 busy 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/common.sh@10 -- # local pid=1027769 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/common.sh@11 -- # local idx=2 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/common.sh@12 -- # local state=busy 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ busy != \b\u\s\y ]] 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/common.sh@18 -- # hash top 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j = 10 )) 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j != 0 )) 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/common.sh@24 -- # grep reactor_2 00:29:35.486 11:40:19 reactor_set_interrupt -- interrupt/common.sh@24 -- # top -bHn 1 -p 1027769 -w 256 00:29:35.743 11:40:19 reactor_set_interrupt -- interrupt/common.sh@24 -- # top_reactor='1027773 root 20 0 128.2g 36864 23616 R 99.9 0.0 0:00.36 reactor_2' 00:29:35.743 11:40:19 reactor_set_interrupt -- interrupt/common.sh@25 -- # awk '{print $9}' 00:29:35.743 11:40:19 reactor_set_interrupt -- interrupt/common.sh@25 -- # echo 1027773 root 20 0 128.2g 36864 23616 R 99.9 0.0 0:00.36 reactor_2 00:29:35.743 11:40:19 reactor_set_interrupt -- interrupt/common.sh@25 -- # sed -e 's/^\s*//g' 00:29:35.743 11:40:19 reactor_set_interrupt -- interrupt/common.sh@25 -- # cpu_rate=99.9 00:29:35.743 11:40:19 reactor_set_interrupt -- interrupt/common.sh@26 -- # cpu_rate=99 00:29:35.743 11:40:19 reactor_set_interrupt -- interrupt/common.sh@28 -- # [[ busy = \b\u\s\y ]] 00:29:35.743 11:40:19 reactor_set_interrupt -- interrupt/common.sh@28 -- # [[ 99 -lt 70 ]] 00:29:35.743 11:40:19 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ busy = \i\d\l\e ]] 00:29:35.743 11:40:19 reactor_set_interrupt -- interrupt/common.sh@33 -- # return 0 00:29:35.743 11:40:19 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@51 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py --plugin interrupt_plugin reactor_set_interrupt_mode 2 00:29:36.001 [2024-07-15 11:40:19.447280] interrupt_tgt.c: 99:rpc_reactor_set_interrupt_mode: *NOTICE*: RPC Start to enable interrupt mode on reactor 2. 00:29:36.001 [2024-07-15 11:40:19.447389] interrupt_tgt.c: 36:rpc_reactor_set_interrupt_mode_cb: *NOTICE*: complete reactor switch 00:29:36.001 11:40:19 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@52 -- # '[' without_thdx '!=' x ']' 00:29:36.001 11:40:19 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@59 -- # reactor_is_idle 1027769 2 00:29:36.001 11:40:19 reactor_set_interrupt -- interrupt/common.sh@51 -- # reactor_is_busy_or_idle 1027769 2 idle 00:29:36.001 11:40:19 reactor_set_interrupt -- interrupt/common.sh@10 -- # local pid=1027769 00:29:36.001 11:40:19 reactor_set_interrupt -- interrupt/common.sh@11 -- # local idx=2 00:29:36.001 11:40:19 reactor_set_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:29:36.001 11:40:19 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ idle != \b\u\s\y ]] 00:29:36.001 11:40:19 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ idle != \i\d\l\e ]] 00:29:36.001 11:40:19 reactor_set_interrupt -- interrupt/common.sh@18 -- # hash top 00:29:36.001 11:40:19 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j = 10 )) 00:29:36.001 11:40:19 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j != 0 )) 00:29:36.001 11:40:19 reactor_set_interrupt -- interrupt/common.sh@24 -- # top -bHn 1 -p 1027769 -w 256 00:29:36.001 11:40:19 reactor_set_interrupt -- interrupt/common.sh@24 -- # grep reactor_2 00:29:36.259 11:40:19 reactor_set_interrupt -- interrupt/common.sh@24 -- # top_reactor='1027773 root 20 0 128.2g 36864 23616 S 0.0 0.0 0:00.60 reactor_2' 00:29:36.259 11:40:19 reactor_set_interrupt -- interrupt/common.sh@25 -- # echo 1027773 root 20 0 128.2g 36864 23616 S 0.0 0.0 0:00.60 reactor_2 00:29:36.259 11:40:19 reactor_set_interrupt -- interrupt/common.sh@25 -- # sed -e 's/^\s*//g' 00:29:36.259 11:40:19 reactor_set_interrupt -- interrupt/common.sh@25 -- # awk '{print $9}' 00:29:36.259 11:40:19 reactor_set_interrupt -- interrupt/common.sh@25 -- # cpu_rate=0.0 00:29:36.259 11:40:19 reactor_set_interrupt -- interrupt/common.sh@26 -- # cpu_rate=0 00:29:36.259 11:40:19 reactor_set_interrupt -- interrupt/common.sh@28 -- # [[ idle = \b\u\s\y ]] 00:29:36.259 11:40:19 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ idle = \i\d\l\e ]] 00:29:36.259 11:40:19 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ 0 -gt 30 ]] 00:29:36.259 11:40:19 reactor_set_interrupt -- interrupt/common.sh@33 -- # return 0 00:29:36.259 11:40:19 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@62 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py --plugin interrupt_plugin reactor_set_interrupt_mode 0 00:29:36.259 [2024-07-15 11:40:19.803270] interrupt_tgt.c: 99:rpc_reactor_set_interrupt_mode: *NOTICE*: RPC Start to enable interrupt mode on reactor 0. 00:29:36.259 [2024-07-15 11:40:19.803377] interrupt_tgt.c: 36:rpc_reactor_set_interrupt_mode_cb: *NOTICE*: complete reactor switch 00:29:36.259 11:40:19 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@63 -- # '[' without_thdx '!=' x ']' 00:29:36.259 11:40:19 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@65 -- # for i in "${thd0_ids[@]}" 00:29:36.259 11:40:19 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@66 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py thread_set_cpumask -i 1 -m 0x1 00:29:36.517 [2024-07-15 11:40:20.051569] thread.c:2099:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:29:36.517 11:40:20 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@70 -- # reactor_is_idle 1027769 0 00:29:36.517 11:40:20 reactor_set_interrupt -- interrupt/common.sh@51 -- # reactor_is_busy_or_idle 1027769 0 idle 00:29:36.517 11:40:20 reactor_set_interrupt -- interrupt/common.sh@10 -- # local pid=1027769 00:29:36.517 11:40:20 reactor_set_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:29:36.517 11:40:20 reactor_set_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:29:36.517 11:40:20 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ idle != \b\u\s\y ]] 00:29:36.517 11:40:20 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ idle != \i\d\l\e ]] 00:29:36.517 11:40:20 reactor_set_interrupt -- interrupt/common.sh@18 -- # hash top 00:29:36.517 11:40:20 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j = 10 )) 00:29:36.517 11:40:20 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j != 0 )) 00:29:36.517 11:40:20 reactor_set_interrupt -- interrupt/common.sh@24 -- # top -bHn 1 -p 1027769 -w 256 00:29:36.517 11:40:20 reactor_set_interrupt -- interrupt/common.sh@24 -- # grep reactor_0 00:29:36.775 11:40:20 reactor_set_interrupt -- interrupt/common.sh@24 -- # top_reactor='1027769 root 20 0 128.2g 36864 23616 S 0.0 0.0 0:01.60 reactor_0' 00:29:36.776 11:40:20 reactor_set_interrupt -- interrupt/common.sh@25 -- # echo 1027769 root 20 0 128.2g 36864 23616 S 0.0 0.0 0:01.60 reactor_0 00:29:36.776 11:40:20 reactor_set_interrupt -- interrupt/common.sh@25 -- # sed -e 's/^\s*//g' 00:29:36.776 11:40:20 reactor_set_interrupt -- interrupt/common.sh@25 -- # awk '{print $9}' 00:29:36.776 11:40:20 reactor_set_interrupt -- interrupt/common.sh@25 -- # cpu_rate=0.0 00:29:36.776 11:40:20 reactor_set_interrupt -- interrupt/common.sh@26 -- # cpu_rate=0 00:29:36.776 11:40:20 reactor_set_interrupt -- interrupt/common.sh@28 -- # [[ idle = \b\u\s\y ]] 00:29:36.776 11:40:20 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ idle = \i\d\l\e ]] 00:29:36.776 11:40:20 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ 0 -gt 30 ]] 00:29:36.776 11:40:20 reactor_set_interrupt -- interrupt/common.sh@33 -- # return 0 00:29:36.776 11:40:20 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@72 -- # return 0 00:29:36.776 11:40:20 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@77 -- # return 0 00:29:36.776 11:40:20 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@92 -- # trap - SIGINT SIGTERM EXIT 00:29:36.776 11:40:20 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@93 -- # killprocess 1027769 00:29:36.776 11:40:20 reactor_set_interrupt -- common/autotest_common.sh@948 -- # '[' -z 1027769 ']' 00:29:36.776 11:40:20 reactor_set_interrupt -- common/autotest_common.sh@952 -- # kill -0 1027769 00:29:36.776 11:40:20 reactor_set_interrupt -- common/autotest_common.sh@953 -- # uname 00:29:36.776 11:40:20 reactor_set_interrupt -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:29:36.776 11:40:20 reactor_set_interrupt -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1027769 00:29:36.776 11:40:20 reactor_set_interrupt -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:29:36.776 11:40:20 reactor_set_interrupt -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:29:36.776 11:40:20 reactor_set_interrupt -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1027769' 00:29:36.776 killing process with pid 1027769 00:29:36.776 11:40:20 reactor_set_interrupt -- common/autotest_common.sh@967 -- # kill 1027769 00:29:36.776 11:40:20 reactor_set_interrupt -- common/autotest_common.sh@972 -- # wait 1027769 00:29:37.034 11:40:20 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@94 -- # cleanup 00:29:37.034 11:40:20 reactor_set_interrupt -- interrupt/common.sh@6 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/aiofile 00:29:37.035 11:40:20 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@97 -- # start_intr_tgt 00:29:37.035 11:40:20 reactor_set_interrupt -- interrupt/interrupt_common.sh@20 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:37.035 11:40:20 reactor_set_interrupt -- interrupt/interrupt_common.sh@21 -- # local cpu_mask=0x07 00:29:37.035 11:40:20 reactor_set_interrupt -- interrupt/interrupt_common.sh@24 -- # intr_tgt_pid=1028379 00:29:37.035 11:40:20 reactor_set_interrupt -- interrupt/interrupt_common.sh@23 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/interrupt_tgt -m 0x07 -r /var/tmp/spdk.sock -E -g 00:29:37.035 11:40:20 reactor_set_interrupt -- interrupt/interrupt_common.sh@25 -- # trap 'killprocess "$intr_tgt_pid"; cleanup; exit 1' SIGINT SIGTERM EXIT 00:29:37.035 11:40:20 reactor_set_interrupt -- interrupt/interrupt_common.sh@26 -- # waitforlisten 1028379 /var/tmp/spdk.sock 00:29:37.035 11:40:20 reactor_set_interrupt -- common/autotest_common.sh@829 -- # '[' -z 1028379 ']' 00:29:37.035 11:40:20 reactor_set_interrupt -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:37.035 11:40:20 reactor_set_interrupt -- common/autotest_common.sh@834 -- # local max_retries=100 00:29:37.035 11:40:20 reactor_set_interrupt -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:37.035 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:37.035 11:40:20 reactor_set_interrupt -- common/autotest_common.sh@838 -- # xtrace_disable 00:29:37.035 11:40:20 reactor_set_interrupt -- common/autotest_common.sh@10 -- # set +x 00:29:37.292 [2024-07-15 11:40:20.641298] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:29:37.292 [2024-07-15 11:40:20.641374] [ DPDK EAL parameters: interrupt_tgt --no-shconf -c 0x07 --single-file-segments --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1028379 ] 00:29:37.292 [2024-07-15 11:40:20.771976] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:29:37.292 [2024-07-15 11:40:20.881817] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:29:37.292 [2024-07-15 11:40:20.882945] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:29:37.292 [2024-07-15 11:40:20.882951] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:29:37.551 [2024-07-15 11:40:20.958588] thread.c:2099:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:29:38.119 11:40:21 reactor_set_interrupt -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:29:38.119 11:40:21 reactor_set_interrupt -- common/autotest_common.sh@862 -- # return 0 00:29:38.119 11:40:21 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@98 -- # setup_bdev_mem 00:29:38.119 11:40:21 reactor_set_interrupt -- interrupt/common.sh@67 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:29:38.378 Malloc0 00:29:38.378 Malloc1 00:29:38.378 Malloc2 00:29:38.378 11:40:21 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@99 -- # setup_bdev_aio 00:29:38.378 11:40:21 reactor_set_interrupt -- interrupt/common.sh@75 -- # uname -s 00:29:38.378 11:40:21 reactor_set_interrupt -- interrupt/common.sh@75 -- # [[ Linux != \F\r\e\e\B\S\D ]] 00:29:38.378 11:40:21 reactor_set_interrupt -- interrupt/common.sh@76 -- # dd if=/dev/zero of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/aiofile bs=2048 count=5000 00:29:38.378 5000+0 records in 00:29:38.378 5000+0 records out 00:29:38.378 10240000 bytes (10 MB, 9.8 MiB) copied, 0.0255862 s, 400 MB/s 00:29:38.378 11:40:21 reactor_set_interrupt -- interrupt/common.sh@77 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/aiofile AIO0 2048 00:29:38.645 AIO0 00:29:38.645 11:40:22 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@101 -- # reactor_set_mode_with_threads 1028379 00:29:38.645 11:40:22 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@81 -- # reactor_set_intr_mode 1028379 00:29:38.645 11:40:22 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@14 -- # local spdk_pid=1028379 00:29:38.645 11:40:22 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@15 -- # local without_thd= 00:29:38.645 11:40:22 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@17 -- # thd0_ids=($(reactor_get_thread_ids $r0_mask)) 00:29:38.645 11:40:22 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@17 -- # reactor_get_thread_ids 0x1 00:29:38.645 11:40:22 reactor_set_interrupt -- interrupt/common.sh@55 -- # local reactor_cpumask=0x1 00:29:38.645 11:40:22 reactor_set_interrupt -- interrupt/common.sh@56 -- # local grep_str 00:29:38.645 11:40:22 reactor_set_interrupt -- interrupt/common.sh@58 -- # reactor_cpumask=1 00:29:38.645 11:40:22 reactor_set_interrupt -- interrupt/common.sh@59 -- # jq_str='.threads|.[]|select(.cpumask == $reactor_cpumask)|.id' 00:29:38.645 11:40:22 reactor_set_interrupt -- interrupt/common.sh@62 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py thread_get_stats 00:29:38.645 11:40:22 reactor_set_interrupt -- interrupt/common.sh@62 -- # jq --arg reactor_cpumask 1 '.threads|.[]|select(.cpumask == $reactor_cpumask)|.id' 00:29:38.959 11:40:22 reactor_set_interrupt -- interrupt/common.sh@62 -- # echo 1 00:29:38.959 11:40:22 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@18 -- # thd2_ids=($(reactor_get_thread_ids $r2_mask)) 00:29:38.959 11:40:22 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@18 -- # reactor_get_thread_ids 0x4 00:29:38.959 11:40:22 reactor_set_interrupt -- interrupt/common.sh@55 -- # local reactor_cpumask=0x4 00:29:38.959 11:40:22 reactor_set_interrupt -- interrupt/common.sh@56 -- # local grep_str 00:29:38.959 11:40:22 reactor_set_interrupt -- interrupt/common.sh@58 -- # reactor_cpumask=4 00:29:38.959 11:40:22 reactor_set_interrupt -- interrupt/common.sh@59 -- # jq_str='.threads|.[]|select(.cpumask == $reactor_cpumask)|.id' 00:29:38.959 11:40:22 reactor_set_interrupt -- interrupt/common.sh@62 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py thread_get_stats 00:29:38.959 11:40:22 reactor_set_interrupt -- interrupt/common.sh@62 -- # jq --arg reactor_cpumask 4 '.threads|.[]|select(.cpumask == $reactor_cpumask)|.id' 00:29:39.217 11:40:22 reactor_set_interrupt -- interrupt/common.sh@62 -- # echo '' 00:29:39.217 11:40:22 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@21 -- # [[ 1 -eq 0 ]] 00:29:39.218 11:40:22 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@25 -- # echo 'spdk_thread ids are 1 on reactor0.' 00:29:39.218 spdk_thread ids are 1 on reactor0. 00:29:39.218 11:40:22 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@29 -- # for i in {0..2} 00:29:39.218 11:40:22 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@30 -- # reactor_is_idle 1028379 0 00:29:39.218 11:40:22 reactor_set_interrupt -- interrupt/common.sh@51 -- # reactor_is_busy_or_idle 1028379 0 idle 00:29:39.218 11:40:22 reactor_set_interrupt -- interrupt/common.sh@10 -- # local pid=1028379 00:29:39.218 11:40:22 reactor_set_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:29:39.218 11:40:22 reactor_set_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:29:39.218 11:40:22 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ idle != \b\u\s\y ]] 00:29:39.218 11:40:22 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ idle != \i\d\l\e ]] 00:29:39.218 11:40:22 reactor_set_interrupt -- interrupt/common.sh@18 -- # hash top 00:29:39.218 11:40:22 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j = 10 )) 00:29:39.218 11:40:22 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j != 0 )) 00:29:39.218 11:40:22 reactor_set_interrupt -- interrupt/common.sh@24 -- # top -bHn 1 -p 1028379 -w 256 00:29:39.218 11:40:22 reactor_set_interrupt -- interrupt/common.sh@24 -- # grep reactor_0 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@24 -- # top_reactor='1028379 root 20 0 128.2g 36864 23616 S 0.0 0.0 0:00.41 reactor_0' 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@25 -- # sed -e 's/^\s*//g' 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@25 -- # echo 1028379 root 20 0 128.2g 36864 23616 S 0.0 0.0 0:00.41 reactor_0 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@25 -- # awk '{print $9}' 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@25 -- # cpu_rate=0.0 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@26 -- # cpu_rate=0 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@28 -- # [[ idle = \b\u\s\y ]] 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ idle = \i\d\l\e ]] 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ 0 -gt 30 ]] 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@33 -- # return 0 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@29 -- # for i in {0..2} 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@30 -- # reactor_is_idle 1028379 1 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@51 -- # reactor_is_busy_or_idle 1028379 1 idle 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@10 -- # local pid=1028379 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@11 -- # local idx=1 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ idle != \b\u\s\y ]] 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ idle != \i\d\l\e ]] 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@18 -- # hash top 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j = 10 )) 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j != 0 )) 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@24 -- # top -bHn 1 -p 1028379 -w 256 00:29:39.477 11:40:22 reactor_set_interrupt -- interrupt/common.sh@24 -- # grep reactor_1 00:29:39.477 11:40:23 reactor_set_interrupt -- interrupt/common.sh@24 -- # top_reactor='1028406 root 20 0 128.2g 36864 23616 S 0.0 0.0 0:00.00 reactor_1' 00:29:39.477 11:40:23 reactor_set_interrupt -- interrupt/common.sh@25 -- # echo 1028406 root 20 0 128.2g 36864 23616 S 0.0 0.0 0:00.00 reactor_1 00:29:39.477 11:40:23 reactor_set_interrupt -- interrupt/common.sh@25 -- # sed -e 's/^\s*//g' 00:29:39.477 11:40:23 reactor_set_interrupt -- interrupt/common.sh@25 -- # awk '{print $9}' 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@25 -- # cpu_rate=0.0 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@26 -- # cpu_rate=0 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@28 -- # [[ idle = \b\u\s\y ]] 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ idle = \i\d\l\e ]] 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ 0 -gt 30 ]] 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@33 -- # return 0 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@29 -- # for i in {0..2} 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@30 -- # reactor_is_idle 1028379 2 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@51 -- # reactor_is_busy_or_idle 1028379 2 idle 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@10 -- # local pid=1028379 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@11 -- # local idx=2 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ idle != \b\u\s\y ]] 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ idle != \i\d\l\e ]] 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@18 -- # hash top 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j = 10 )) 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j != 0 )) 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@24 -- # top -bHn 1 -p 1028379 -w 256 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@24 -- # grep reactor_2 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@24 -- # top_reactor='1028407 root 20 0 128.2g 36864 23616 S 0.0 0.0 0:00.00 reactor_2' 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@25 -- # echo 1028407 root 20 0 128.2g 36864 23616 S 0.0 0.0 0:00.00 reactor_2 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@25 -- # sed -e 's/^\s*//g' 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@25 -- # awk '{print $9}' 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@25 -- # cpu_rate=0.0 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@26 -- # cpu_rate=0 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@28 -- # [[ idle = \b\u\s\y ]] 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ idle = \i\d\l\e ]] 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ 0 -gt 30 ]] 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/common.sh@33 -- # return 0 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@33 -- # '[' x '!=' x ']' 00:29:39.736 11:40:23 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@43 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py --plugin interrupt_plugin reactor_set_interrupt_mode 0 -d 00:29:39.994 [2024-07-15 11:40:23.483539] interrupt_tgt.c: 99:rpc_reactor_set_interrupt_mode: *NOTICE*: RPC Start to disable interrupt mode on reactor 0. 00:29:39.994 [2024-07-15 11:40:23.483747] thread.c:2099:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to poll mode from intr mode. 00:29:39.994 [2024-07-15 11:40:23.483827] interrupt_tgt.c: 36:rpc_reactor_set_interrupt_mode_cb: *NOTICE*: complete reactor switch 00:29:39.994 11:40:23 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@44 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py --plugin interrupt_plugin reactor_set_interrupt_mode 2 -d 00:29:40.252 [2024-07-15 11:40:23.736100] interrupt_tgt.c: 99:rpc_reactor_set_interrupt_mode: *NOTICE*: RPC Start to disable interrupt mode on reactor 2. 00:29:40.252 [2024-07-15 11:40:23.736285] interrupt_tgt.c: 36:rpc_reactor_set_interrupt_mode_cb: *NOTICE*: complete reactor switch 00:29:40.252 11:40:23 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@46 -- # for i in 0 2 00:29:40.252 11:40:23 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@47 -- # reactor_is_busy 1028379 0 00:29:40.252 11:40:23 reactor_set_interrupt -- interrupt/common.sh@47 -- # reactor_is_busy_or_idle 1028379 0 busy 00:29:40.252 11:40:23 reactor_set_interrupt -- interrupt/common.sh@10 -- # local pid=1028379 00:29:40.252 11:40:23 reactor_set_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:29:40.252 11:40:23 reactor_set_interrupt -- interrupt/common.sh@12 -- # local state=busy 00:29:40.252 11:40:23 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ busy != \b\u\s\y ]] 00:29:40.252 11:40:23 reactor_set_interrupt -- interrupt/common.sh@18 -- # hash top 00:29:40.252 11:40:23 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j = 10 )) 00:29:40.252 11:40:23 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j != 0 )) 00:29:40.253 11:40:23 reactor_set_interrupt -- interrupt/common.sh@24 -- # top -bHn 1 -p 1028379 -w 256 00:29:40.253 11:40:23 reactor_set_interrupt -- interrupt/common.sh@24 -- # grep reactor_0 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/common.sh@24 -- # top_reactor='1028379 root 20 0 128.2g 36864 23616 R 99.9 0.0 0:00.84 reactor_0' 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/common.sh@25 -- # sed -e 's/^\s*//g' 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/common.sh@25 -- # echo 1028379 root 20 0 128.2g 36864 23616 R 99.9 0.0 0:00.84 reactor_0 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/common.sh@25 -- # awk '{print $9}' 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/common.sh@25 -- # cpu_rate=99.9 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/common.sh@26 -- # cpu_rate=99 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/common.sh@28 -- # [[ busy = \b\u\s\y ]] 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/common.sh@28 -- # [[ 99 -lt 70 ]] 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ busy = \i\d\l\e ]] 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/common.sh@33 -- # return 0 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@46 -- # for i in 0 2 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@47 -- # reactor_is_busy 1028379 2 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/common.sh@47 -- # reactor_is_busy_or_idle 1028379 2 busy 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/common.sh@10 -- # local pid=1028379 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/common.sh@11 -- # local idx=2 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/common.sh@12 -- # local state=busy 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ busy != \b\u\s\y ]] 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/common.sh@18 -- # hash top 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j = 10 )) 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j != 0 )) 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/common.sh@24 -- # top -bHn 1 -p 1028379 -w 256 00:29:40.512 11:40:23 reactor_set_interrupt -- interrupt/common.sh@24 -- # grep reactor_2 00:29:40.512 11:40:24 reactor_set_interrupt -- interrupt/common.sh@24 -- # top_reactor='1028407 root 20 0 128.2g 36864 23616 R 99.9 0.0 0:00.36 reactor_2' 00:29:40.771 11:40:24 reactor_set_interrupt -- interrupt/common.sh@25 -- # echo 1028407 root 20 0 128.2g 36864 23616 R 99.9 0.0 0:00.36 reactor_2 00:29:40.771 11:40:24 reactor_set_interrupt -- interrupt/common.sh@25 -- # sed -e 's/^\s*//g' 00:29:40.771 11:40:24 reactor_set_interrupt -- interrupt/common.sh@25 -- # awk '{print $9}' 00:29:40.771 11:40:24 reactor_set_interrupt -- interrupt/common.sh@25 -- # cpu_rate=99.9 00:29:40.771 11:40:24 reactor_set_interrupt -- interrupt/common.sh@26 -- # cpu_rate=99 00:29:40.771 11:40:24 reactor_set_interrupt -- interrupt/common.sh@28 -- # [[ busy = \b\u\s\y ]] 00:29:40.771 11:40:24 reactor_set_interrupt -- interrupt/common.sh@28 -- # [[ 99 -lt 70 ]] 00:29:40.771 11:40:24 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ busy = \i\d\l\e ]] 00:29:40.771 11:40:24 reactor_set_interrupt -- interrupt/common.sh@33 -- # return 0 00:29:40.771 11:40:24 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@51 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py --plugin interrupt_plugin reactor_set_interrupt_mode 2 00:29:40.771 [2024-07-15 11:40:24.345851] interrupt_tgt.c: 99:rpc_reactor_set_interrupt_mode: *NOTICE*: RPC Start to enable interrupt mode on reactor 2. 00:29:40.771 [2024-07-15 11:40:24.345968] interrupt_tgt.c: 36:rpc_reactor_set_interrupt_mode_cb: *NOTICE*: complete reactor switch 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@52 -- # '[' x '!=' x ']' 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@59 -- # reactor_is_idle 1028379 2 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@51 -- # reactor_is_busy_or_idle 1028379 2 idle 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@10 -- # local pid=1028379 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@11 -- # local idx=2 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ idle != \b\u\s\y ]] 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ idle != \i\d\l\e ]] 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@18 -- # hash top 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j = 10 )) 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j != 0 )) 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@24 -- # top -bHn 1 -p 1028379 -w 256 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@24 -- # grep reactor_2 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@24 -- # top_reactor='1028407 root 20 0 128.2g 36864 23616 S 0.0 0.0 0:00.60 reactor_2' 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@25 -- # echo 1028407 root 20 0 128.2g 36864 23616 S 0.0 0.0 0:00.60 reactor_2 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@25 -- # sed -e 's/^\s*//g' 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@25 -- # awk '{print $9}' 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@25 -- # cpu_rate=0.0 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@26 -- # cpu_rate=0 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@28 -- # [[ idle = \b\u\s\y ]] 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ idle = \i\d\l\e ]] 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ 0 -gt 30 ]] 00:29:41.029 11:40:24 reactor_set_interrupt -- interrupt/common.sh@33 -- # return 0 00:29:41.030 11:40:24 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@62 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py --plugin interrupt_plugin reactor_set_interrupt_mode 0 00:29:41.288 [2024-07-15 11:40:24.774959] interrupt_tgt.c: 99:rpc_reactor_set_interrupt_mode: *NOTICE*: RPC Start to enable interrupt mode on reactor 0. 00:29:41.288 [2024-07-15 11:40:24.775170] thread.c:2099:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from poll mode. 00:29:41.288 [2024-07-15 11:40:24.775195] interrupt_tgt.c: 36:rpc_reactor_set_interrupt_mode_cb: *NOTICE*: complete reactor switch 00:29:41.288 11:40:24 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@63 -- # '[' x '!=' x ']' 00:29:41.288 11:40:24 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@70 -- # reactor_is_idle 1028379 0 00:29:41.288 11:40:24 reactor_set_interrupt -- interrupt/common.sh@51 -- # reactor_is_busy_or_idle 1028379 0 idle 00:29:41.288 11:40:24 reactor_set_interrupt -- interrupt/common.sh@10 -- # local pid=1028379 00:29:41.288 11:40:24 reactor_set_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:29:41.288 11:40:24 reactor_set_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:29:41.288 11:40:24 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ idle != \b\u\s\y ]] 00:29:41.288 11:40:24 reactor_set_interrupt -- interrupt/common.sh@14 -- # [[ idle != \i\d\l\e ]] 00:29:41.288 11:40:24 reactor_set_interrupt -- interrupt/common.sh@18 -- # hash top 00:29:41.288 11:40:24 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j = 10 )) 00:29:41.288 11:40:24 reactor_set_interrupt -- interrupt/common.sh@23 -- # (( j != 0 )) 00:29:41.288 11:40:24 reactor_set_interrupt -- interrupt/common.sh@24 -- # top -bHn 1 -p 1028379 -w 256 00:29:41.288 11:40:24 reactor_set_interrupt -- interrupt/common.sh@24 -- # grep reactor_0 00:29:41.548 11:40:24 reactor_set_interrupt -- interrupt/common.sh@24 -- # top_reactor='1028379 root 20 0 128.2g 36864 23616 S 0.0 0.0 0:01.69 reactor_0' 00:29:41.548 11:40:24 reactor_set_interrupt -- interrupt/common.sh@25 -- # echo 1028379 root 20 0 128.2g 36864 23616 S 0.0 0.0 0:01.69 reactor_0 00:29:41.548 11:40:24 reactor_set_interrupt -- interrupt/common.sh@25 -- # sed -e 's/^\s*//g' 00:29:41.548 11:40:24 reactor_set_interrupt -- interrupt/common.sh@25 -- # awk '{print $9}' 00:29:41.548 11:40:24 reactor_set_interrupt -- interrupt/common.sh@25 -- # cpu_rate=0.0 00:29:41.548 11:40:24 reactor_set_interrupt -- interrupt/common.sh@26 -- # cpu_rate=0 00:29:41.548 11:40:24 reactor_set_interrupt -- interrupt/common.sh@28 -- # [[ idle = \b\u\s\y ]] 00:29:41.548 11:40:24 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ idle = \i\d\l\e ]] 00:29:41.548 11:40:24 reactor_set_interrupt -- interrupt/common.sh@30 -- # [[ 0 -gt 30 ]] 00:29:41.548 11:40:24 reactor_set_interrupt -- interrupt/common.sh@33 -- # return 0 00:29:41.548 11:40:24 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@72 -- # return 0 00:29:41.549 11:40:24 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@82 -- # return 0 00:29:41.549 11:40:24 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@103 -- # trap - SIGINT SIGTERM EXIT 00:29:41.549 11:40:24 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@104 -- # killprocess 1028379 00:29:41.549 11:40:24 reactor_set_interrupt -- common/autotest_common.sh@948 -- # '[' -z 1028379 ']' 00:29:41.549 11:40:24 reactor_set_interrupt -- common/autotest_common.sh@952 -- # kill -0 1028379 00:29:41.549 11:40:24 reactor_set_interrupt -- common/autotest_common.sh@953 -- # uname 00:29:41.549 11:40:24 reactor_set_interrupt -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:29:41.549 11:40:24 reactor_set_interrupt -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1028379 00:29:41.549 11:40:25 reactor_set_interrupt -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:29:41.549 11:40:25 reactor_set_interrupt -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:29:41.549 11:40:25 reactor_set_interrupt -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1028379' 00:29:41.549 killing process with pid 1028379 00:29:41.549 11:40:25 reactor_set_interrupt -- common/autotest_common.sh@967 -- # kill 1028379 00:29:41.549 11:40:25 reactor_set_interrupt -- common/autotest_common.sh@972 -- # wait 1028379 00:29:41.808 11:40:25 reactor_set_interrupt -- interrupt/reactor_set_interrupt.sh@105 -- # cleanup 00:29:41.808 11:40:25 reactor_set_interrupt -- interrupt/common.sh@6 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/aiofile 00:29:41.808 00:29:41.808 real 0m10.095s 00:29:41.808 user 0m9.407s 00:29:41.808 sys 0m2.159s 00:29:41.808 11:40:25 reactor_set_interrupt -- common/autotest_common.sh@1124 -- # xtrace_disable 00:29:41.808 11:40:25 reactor_set_interrupt -- common/autotest_common.sh@10 -- # set +x 00:29:41.808 ************************************ 00:29:41.808 END TEST reactor_set_interrupt 00:29:41.808 ************************************ 00:29:41.808 11:40:25 -- common/autotest_common.sh@1142 -- # return 0 00:29:41.808 11:40:25 -- spdk/autotest.sh@194 -- # run_test reap_unregistered_poller /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/reap_unregistered_poller.sh 00:29:41.808 11:40:25 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:29:41.808 11:40:25 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:29:41.808 11:40:25 -- common/autotest_common.sh@10 -- # set +x 00:29:41.808 ************************************ 00:29:41.808 START TEST reap_unregistered_poller 00:29:41.808 ************************************ 00:29:41.808 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/reap_unregistered_poller.sh 00:29:42.068 * Looking for test storage... 00:29:42.068 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt 00:29:42.068 11:40:25 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@9 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/interrupt_common.sh 00:29:42.068 11:40:25 reap_unregistered_poller -- interrupt/interrupt_common.sh@5 -- # dirname /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/reap_unregistered_poller.sh 00:29:42.068 11:40:25 reap_unregistered_poller -- interrupt/interrupt_common.sh@5 -- # readlink -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt 00:29:42.068 11:40:25 reap_unregistered_poller -- interrupt/interrupt_common.sh@5 -- # testdir=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt 00:29:42.068 11:40:25 reap_unregistered_poller -- interrupt/interrupt_common.sh@6 -- # readlink -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/../.. 00:29:42.068 11:40:25 reap_unregistered_poller -- interrupt/interrupt_common.sh@6 -- # rootdir=/var/jenkins/workspace/crypto-phy-autotest/spdk 00:29:42.068 11:40:25 reap_unregistered_poller -- interrupt/interrupt_common.sh@7 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/common/autotest_common.sh 00:29:42.068 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@7 -- # rpc_py=rpc_cmd 00:29:42.068 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@34 -- # set -e 00:29:42.068 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@35 -- # shopt -s nullglob 00:29:42.068 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@36 -- # shopt -s extglob 00:29:42.068 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@37 -- # shopt -s inherit_errexit 00:29:42.068 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@39 -- # '[' -z /var/jenkins/workspace/crypto-phy-autotest/spdk/../output ']' 00:29:42.068 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@44 -- # [[ -e /var/jenkins/workspace/crypto-phy-autotest/spdk/test/common/build_config.sh ]] 00:29:42.068 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@45 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/common/build_config.sh 00:29:42.068 11:40:25 reap_unregistered_poller -- common/build_config.sh@1 -- # CONFIG_WPDK_DIR= 00:29:42.068 11:40:25 reap_unregistered_poller -- common/build_config.sh@2 -- # CONFIG_ASAN=n 00:29:42.068 11:40:25 reap_unregistered_poller -- common/build_config.sh@3 -- # CONFIG_VBDEV_COMPRESS=y 00:29:42.068 11:40:25 reap_unregistered_poller -- common/build_config.sh@4 -- # CONFIG_HAVE_EXECINFO_H=y 00:29:42.068 11:40:25 reap_unregistered_poller -- common/build_config.sh@5 -- # CONFIG_USDT=n 00:29:42.068 11:40:25 reap_unregistered_poller -- common/build_config.sh@6 -- # CONFIG_CUSTOMOCF=n 00:29:42.068 11:40:25 reap_unregistered_poller -- common/build_config.sh@7 -- # CONFIG_PREFIX=/usr/local 00:29:42.068 11:40:25 reap_unregistered_poller -- common/build_config.sh@8 -- # CONFIG_RBD=n 00:29:42.068 11:40:25 reap_unregistered_poller -- common/build_config.sh@9 -- # CONFIG_LIBDIR= 00:29:42.068 11:40:25 reap_unregistered_poller -- common/build_config.sh@10 -- # CONFIG_IDXD=y 00:29:42.068 11:40:25 reap_unregistered_poller -- common/build_config.sh@11 -- # CONFIG_NVME_CUSE=y 00:29:42.068 11:40:25 reap_unregistered_poller -- common/build_config.sh@12 -- # CONFIG_SMA=n 00:29:42.068 11:40:25 reap_unregistered_poller -- common/build_config.sh@13 -- # CONFIG_VTUNE=n 00:29:42.068 11:40:25 reap_unregistered_poller -- common/build_config.sh@14 -- # CONFIG_TSAN=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@15 -- # CONFIG_RDMA_SEND_WITH_INVAL=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@16 -- # CONFIG_VFIO_USER_DIR= 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@17 -- # CONFIG_PGO_CAPTURE=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@18 -- # CONFIG_HAVE_UUID_GENERATE_SHA1=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@19 -- # CONFIG_ENV=/var/jenkins/workspace/crypto-phy-autotest/spdk/lib/env_dpdk 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@20 -- # CONFIG_LTO=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@21 -- # CONFIG_ISCSI_INITIATOR=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@22 -- # CONFIG_CET=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@23 -- # CONFIG_VBDEV_COMPRESS_MLX5=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@24 -- # CONFIG_OCF_PATH= 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@25 -- # CONFIG_RDMA_SET_TOS=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@26 -- # CONFIG_HAVE_ARC4RANDOM=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@27 -- # CONFIG_HAVE_LIBARCHIVE=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@28 -- # CONFIG_UBLK=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@29 -- # CONFIG_ISAL_CRYPTO=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@30 -- # CONFIG_OPENSSL_PATH= 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@31 -- # CONFIG_OCF=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@32 -- # CONFIG_FUSE=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@33 -- # CONFIG_VTUNE_DIR= 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@34 -- # CONFIG_FUZZER_LIB= 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@35 -- # CONFIG_FUZZER=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@36 -- # CONFIG_DPDK_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@37 -- # CONFIG_CRYPTO=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@38 -- # CONFIG_PGO_USE=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@39 -- # CONFIG_VHOST=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@40 -- # CONFIG_DAOS=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@41 -- # CONFIG_DPDK_INC_DIR= 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@42 -- # CONFIG_DAOS_DIR= 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@43 -- # CONFIG_UNIT_TESTS=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@44 -- # CONFIG_RDMA_SET_ACK_TIMEOUT=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@45 -- # CONFIG_VIRTIO=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@46 -- # CONFIG_DPDK_UADK=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@47 -- # CONFIG_COVERAGE=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@48 -- # CONFIG_RDMA=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@49 -- # CONFIG_FIO_SOURCE_DIR=/usr/src/fio 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@50 -- # CONFIG_URING_PATH= 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@51 -- # CONFIG_XNVME=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@52 -- # CONFIG_VFIO_USER=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@53 -- # CONFIG_ARCH=native 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@54 -- # CONFIG_HAVE_EVP_MAC=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@55 -- # CONFIG_URING_ZNS=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@56 -- # CONFIG_WERROR=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@57 -- # CONFIG_HAVE_LIBBSD=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@58 -- # CONFIG_UBSAN=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@59 -- # CONFIG_IPSEC_MB_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/intel-ipsec-mb/lib 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@60 -- # CONFIG_GOLANG=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@61 -- # CONFIG_ISAL=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@62 -- # CONFIG_IDXD_KERNEL=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@63 -- # CONFIG_DPDK_LIB_DIR= 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@64 -- # CONFIG_RDMA_PROV=verbs 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@65 -- # CONFIG_APPS=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@66 -- # CONFIG_SHARED=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@67 -- # CONFIG_HAVE_KEYUTILS=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@68 -- # CONFIG_FC_PATH= 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@69 -- # CONFIG_DPDK_PKG_CONFIG=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@70 -- # CONFIG_FC=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@71 -- # CONFIG_AVAHI=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@72 -- # CONFIG_FIO_PLUGIN=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@73 -- # CONFIG_RAID5F=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@74 -- # CONFIG_EXAMPLES=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@75 -- # CONFIG_TESTS=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@76 -- # CONFIG_CRYPTO_MLX5=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@77 -- # CONFIG_MAX_LCORES=128 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@78 -- # CONFIG_IPSEC_MB=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@79 -- # CONFIG_PGO_DIR= 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@80 -- # CONFIG_DEBUG=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@81 -- # CONFIG_DPDK_COMPRESSDEV=y 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@82 -- # CONFIG_CROSS_PREFIX= 00:29:42.069 11:40:25 reap_unregistered_poller -- common/build_config.sh@83 -- # CONFIG_URING=n 00:29:42.069 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@54 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/common/applications.sh 00:29:42.069 11:40:25 reap_unregistered_poller -- common/applications.sh@8 -- # dirname /var/jenkins/workspace/crypto-phy-autotest/spdk/test/common/applications.sh 00:29:42.069 11:40:25 reap_unregistered_poller -- common/applications.sh@8 -- # readlink -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/common 00:29:42.069 11:40:25 reap_unregistered_poller -- common/applications.sh@8 -- # _root=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/common 00:29:42.069 11:40:25 reap_unregistered_poller -- common/applications.sh@9 -- # _root=/var/jenkins/workspace/crypto-phy-autotest/spdk 00:29:42.069 11:40:25 reap_unregistered_poller -- common/applications.sh@10 -- # _app_dir=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin 00:29:42.069 11:40:25 reap_unregistered_poller -- common/applications.sh@11 -- # _test_app_dir=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/app 00:29:42.069 11:40:25 reap_unregistered_poller -- common/applications.sh@12 -- # _examples_dir=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples 00:29:42.069 11:40:25 reap_unregistered_poller -- common/applications.sh@14 -- # VHOST_FUZZ_APP=("$_test_app_dir/fuzz/vhost_fuzz/vhost_fuzz") 00:29:42.069 11:40:25 reap_unregistered_poller -- common/applications.sh@15 -- # ISCSI_APP=("$_app_dir/iscsi_tgt") 00:29:42.069 11:40:25 reap_unregistered_poller -- common/applications.sh@16 -- # NVMF_APP=("$_app_dir/nvmf_tgt") 00:29:42.069 11:40:25 reap_unregistered_poller -- common/applications.sh@17 -- # VHOST_APP=("$_app_dir/vhost") 00:29:42.069 11:40:25 reap_unregistered_poller -- common/applications.sh@18 -- # DD_APP=("$_app_dir/spdk_dd") 00:29:42.069 11:40:25 reap_unregistered_poller -- common/applications.sh@19 -- # SPDK_APP=("$_app_dir/spdk_tgt") 00:29:42.069 11:40:25 reap_unregistered_poller -- common/applications.sh@22 -- # [[ -e /var/jenkins/workspace/crypto-phy-autotest/spdk/include/spdk/config.h ]] 00:29:42.069 11:40:25 reap_unregistered_poller -- common/applications.sh@23 -- # [[ #ifndef SPDK_CONFIG_H 00:29:42.069 #define SPDK_CONFIG_H 00:29:42.069 #define SPDK_CONFIG_APPS 1 00:29:42.069 #define SPDK_CONFIG_ARCH native 00:29:42.069 #undef SPDK_CONFIG_ASAN 00:29:42.069 #undef SPDK_CONFIG_AVAHI 00:29:42.069 #undef SPDK_CONFIG_CET 00:29:42.069 #define SPDK_CONFIG_COVERAGE 1 00:29:42.069 #define SPDK_CONFIG_CROSS_PREFIX 00:29:42.069 #define SPDK_CONFIG_CRYPTO 1 00:29:42.069 #define SPDK_CONFIG_CRYPTO_MLX5 1 00:29:42.069 #undef SPDK_CONFIG_CUSTOMOCF 00:29:42.069 #undef SPDK_CONFIG_DAOS 00:29:42.069 #define SPDK_CONFIG_DAOS_DIR 00:29:42.069 #define SPDK_CONFIG_DEBUG 1 00:29:42.069 #define SPDK_CONFIG_DPDK_COMPRESSDEV 1 00:29:42.069 #define SPDK_CONFIG_DPDK_DIR /var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build 00:29:42.069 #define SPDK_CONFIG_DPDK_INC_DIR 00:29:42.069 #define SPDK_CONFIG_DPDK_LIB_DIR 00:29:42.069 #undef SPDK_CONFIG_DPDK_PKG_CONFIG 00:29:42.069 #undef SPDK_CONFIG_DPDK_UADK 00:29:42.069 #define SPDK_CONFIG_ENV /var/jenkins/workspace/crypto-phy-autotest/spdk/lib/env_dpdk 00:29:42.069 #define SPDK_CONFIG_EXAMPLES 1 00:29:42.069 #undef SPDK_CONFIG_FC 00:29:42.069 #define SPDK_CONFIG_FC_PATH 00:29:42.069 #define SPDK_CONFIG_FIO_PLUGIN 1 00:29:42.069 #define SPDK_CONFIG_FIO_SOURCE_DIR /usr/src/fio 00:29:42.069 #undef SPDK_CONFIG_FUSE 00:29:42.069 #undef SPDK_CONFIG_FUZZER 00:29:42.069 #define SPDK_CONFIG_FUZZER_LIB 00:29:42.069 #undef SPDK_CONFIG_GOLANG 00:29:42.069 #define SPDK_CONFIG_HAVE_ARC4RANDOM 1 00:29:42.069 #define SPDK_CONFIG_HAVE_EVP_MAC 1 00:29:42.069 #define SPDK_CONFIG_HAVE_EXECINFO_H 1 00:29:42.069 #define SPDK_CONFIG_HAVE_KEYUTILS 1 00:29:42.069 #undef SPDK_CONFIG_HAVE_LIBARCHIVE 00:29:42.069 #undef SPDK_CONFIG_HAVE_LIBBSD 00:29:42.069 #define SPDK_CONFIG_HAVE_UUID_GENERATE_SHA1 1 00:29:42.069 #define SPDK_CONFIG_IDXD 1 00:29:42.069 #define SPDK_CONFIG_IDXD_KERNEL 1 00:29:42.069 #define SPDK_CONFIG_IPSEC_MB 1 00:29:42.069 #define SPDK_CONFIG_IPSEC_MB_DIR /var/jenkins/workspace/crypto-phy-autotest/spdk/intel-ipsec-mb/lib 00:29:42.069 #define SPDK_CONFIG_ISAL 1 00:29:42.069 #define SPDK_CONFIG_ISAL_CRYPTO 1 00:29:42.069 #define SPDK_CONFIG_ISCSI_INITIATOR 1 00:29:42.069 #define SPDK_CONFIG_LIBDIR 00:29:42.069 #undef SPDK_CONFIG_LTO 00:29:42.069 #define SPDK_CONFIG_MAX_LCORES 128 00:29:42.069 #define SPDK_CONFIG_NVME_CUSE 1 00:29:42.069 #undef SPDK_CONFIG_OCF 00:29:42.069 #define SPDK_CONFIG_OCF_PATH 00:29:42.069 #define SPDK_CONFIG_OPENSSL_PATH 00:29:42.069 #undef SPDK_CONFIG_PGO_CAPTURE 00:29:42.069 #define SPDK_CONFIG_PGO_DIR 00:29:42.069 #undef SPDK_CONFIG_PGO_USE 00:29:42.069 #define SPDK_CONFIG_PREFIX /usr/local 00:29:42.069 #undef SPDK_CONFIG_RAID5F 00:29:42.069 #undef SPDK_CONFIG_RBD 00:29:42.069 #define SPDK_CONFIG_RDMA 1 00:29:42.069 #define SPDK_CONFIG_RDMA_PROV verbs 00:29:42.069 #define SPDK_CONFIG_RDMA_SEND_WITH_INVAL 1 00:29:42.069 #define SPDK_CONFIG_RDMA_SET_ACK_TIMEOUT 1 00:29:42.069 #define SPDK_CONFIG_RDMA_SET_TOS 1 00:29:42.069 #define SPDK_CONFIG_SHARED 1 00:29:42.069 #undef SPDK_CONFIG_SMA 00:29:42.069 #define SPDK_CONFIG_TESTS 1 00:29:42.069 #undef SPDK_CONFIG_TSAN 00:29:42.069 #define SPDK_CONFIG_UBLK 1 00:29:42.069 #define SPDK_CONFIG_UBSAN 1 00:29:42.069 #undef SPDK_CONFIG_UNIT_TESTS 00:29:42.069 #undef SPDK_CONFIG_URING 00:29:42.069 #define SPDK_CONFIG_URING_PATH 00:29:42.069 #undef SPDK_CONFIG_URING_ZNS 00:29:42.069 #undef SPDK_CONFIG_USDT 00:29:42.069 #define SPDK_CONFIG_VBDEV_COMPRESS 1 00:29:42.070 #define SPDK_CONFIG_VBDEV_COMPRESS_MLX5 1 00:29:42.070 #undef SPDK_CONFIG_VFIO_USER 00:29:42.070 #define SPDK_CONFIG_VFIO_USER_DIR 00:29:42.070 #define SPDK_CONFIG_VHOST 1 00:29:42.070 #define SPDK_CONFIG_VIRTIO 1 00:29:42.070 #undef SPDK_CONFIG_VTUNE 00:29:42.070 #define SPDK_CONFIG_VTUNE_DIR 00:29:42.070 #define SPDK_CONFIG_WERROR 1 00:29:42.070 #define SPDK_CONFIG_WPDK_DIR 00:29:42.070 #undef SPDK_CONFIG_XNVME 00:29:42.070 #endif /* SPDK_CONFIG_H */ == *\#\d\e\f\i\n\e\ \S\P\D\K\_\C\O\N\F\I\G\_\D\E\B\U\G* ]] 00:29:42.070 11:40:25 reap_unregistered_poller -- common/applications.sh@24 -- # (( SPDK_AUTOTEST_DEBUG_APPS )) 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@55 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/common.sh 00:29:42.070 11:40:25 reap_unregistered_poller -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:29:42.070 11:40:25 reap_unregistered_poller -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:29:42.070 11:40:25 reap_unregistered_poller -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:29:42.070 11:40:25 reap_unregistered_poller -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:42.070 11:40:25 reap_unregistered_poller -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:42.070 11:40:25 reap_unregistered_poller -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:42.070 11:40:25 reap_unregistered_poller -- paths/export.sh@5 -- # export PATH 00:29:42.070 11:40:25 reap_unregistered_poller -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@56 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm/common 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@6 -- # dirname /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm/common 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@6 -- # readlink -f /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@6 -- # _pmdir=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@7 -- # readlink -f /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm/../../../ 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@7 -- # _pmrootdir=/var/jenkins/workspace/crypto-phy-autotest/spdk 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@64 -- # TEST_TAG=N/A 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@65 -- # TEST_TAG_FILE=/var/jenkins/workspace/crypto-phy-autotest/spdk/.run_test_name 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@67 -- # PM_OUTPUTDIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@68 -- # uname -s 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@68 -- # PM_OS=Linux 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@70 -- # MONITOR_RESOURCES_SUDO=() 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@70 -- # declare -A MONITOR_RESOURCES_SUDO 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@71 -- # MONITOR_RESOURCES_SUDO["collect-bmc-pm"]=1 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@72 -- # MONITOR_RESOURCES_SUDO["collect-cpu-load"]=0 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@73 -- # MONITOR_RESOURCES_SUDO["collect-cpu-temp"]=0 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@74 -- # MONITOR_RESOURCES_SUDO["collect-vmstat"]=0 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@76 -- # SUDO[0]= 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@76 -- # SUDO[1]='sudo -E' 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@78 -- # MONITOR_RESOURCES=(collect-cpu-load collect-vmstat) 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@79 -- # [[ Linux == FreeBSD ]] 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@81 -- # [[ Linux == Linux ]] 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@81 -- # [[ ............................... != QEMU ]] 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@81 -- # [[ ! -e /.dockerenv ]] 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@84 -- # MONITOR_RESOURCES+=(collect-cpu-temp) 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@85 -- # MONITOR_RESOURCES+=(collect-bmc-pm) 00:29:42.070 11:40:25 reap_unregistered_poller -- pm/common@88 -- # [[ ! -d /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power ]] 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@58 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@59 -- # export RUN_NIGHTLY 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@62 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@63 -- # export SPDK_AUTOTEST_DEBUG_APPS 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@64 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@65 -- # export SPDK_RUN_VALGRIND 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@66 -- # : 1 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@67 -- # export SPDK_RUN_FUNCTIONAL_TEST 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@68 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@69 -- # export SPDK_TEST_UNITTEST 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@70 -- # : 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@71 -- # export SPDK_TEST_AUTOBUILD 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@72 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@73 -- # export SPDK_TEST_RELEASE_BUILD 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@74 -- # : 1 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@75 -- # export SPDK_TEST_ISAL 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@76 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@77 -- # export SPDK_TEST_ISCSI 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@78 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@79 -- # export SPDK_TEST_ISCSI_INITIATOR 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@80 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@81 -- # export SPDK_TEST_NVME 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@82 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@83 -- # export SPDK_TEST_NVME_PMR 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@84 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@85 -- # export SPDK_TEST_NVME_BP 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@86 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@87 -- # export SPDK_TEST_NVME_CLI 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@88 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@89 -- # export SPDK_TEST_NVME_CUSE 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@90 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@91 -- # export SPDK_TEST_NVME_FDP 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@92 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@93 -- # export SPDK_TEST_NVMF 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@94 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@95 -- # export SPDK_TEST_VFIOUSER 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@96 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@97 -- # export SPDK_TEST_VFIOUSER_QEMU 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@98 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@99 -- # export SPDK_TEST_FUZZER 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@100 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@101 -- # export SPDK_TEST_FUZZER_SHORT 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@102 -- # : rdma 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@103 -- # export SPDK_TEST_NVMF_TRANSPORT 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@104 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@105 -- # export SPDK_TEST_RBD 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@106 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@107 -- # export SPDK_TEST_VHOST 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@108 -- # : 1 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@109 -- # export SPDK_TEST_BLOCKDEV 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@110 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@111 -- # export SPDK_TEST_IOAT 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@112 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@113 -- # export SPDK_TEST_BLOBFS 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@114 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@115 -- # export SPDK_TEST_VHOST_INIT 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@116 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@117 -- # export SPDK_TEST_LVOL 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@118 -- # : 1 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@119 -- # export SPDK_TEST_VBDEV_COMPRESS 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@120 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@121 -- # export SPDK_RUN_ASAN 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@122 -- # : 1 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@123 -- # export SPDK_RUN_UBSAN 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@124 -- # : 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@125 -- # export SPDK_RUN_EXTERNAL_DPDK 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@126 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@127 -- # export SPDK_RUN_NON_ROOT 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@128 -- # : 1 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@129 -- # export SPDK_TEST_CRYPTO 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@130 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@131 -- # export SPDK_TEST_FTL 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@132 -- # : 0 00:29:42.070 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@133 -- # export SPDK_TEST_OCF 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@134 -- # : 0 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@135 -- # export SPDK_TEST_VMD 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@136 -- # : 0 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@137 -- # export SPDK_TEST_OPAL 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@138 -- # : 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@139 -- # export SPDK_TEST_NATIVE_DPDK 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@140 -- # : true 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@141 -- # export SPDK_AUTOTEST_X 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@142 -- # : 0 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@143 -- # export SPDK_TEST_RAID5 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@144 -- # : 0 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@145 -- # export SPDK_TEST_URING 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@146 -- # : 0 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@147 -- # export SPDK_TEST_USDT 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@148 -- # : 0 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@149 -- # export SPDK_TEST_USE_IGB_UIO 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@150 -- # : 0 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@151 -- # export SPDK_TEST_SCHEDULER 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@152 -- # : 0 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@153 -- # export SPDK_TEST_SCANBUILD 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@154 -- # : 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@155 -- # export SPDK_TEST_NVMF_NICS 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@156 -- # : 0 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@157 -- # export SPDK_TEST_SMA 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@158 -- # : 0 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@159 -- # export SPDK_TEST_DAOS 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@160 -- # : 0 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@161 -- # export SPDK_TEST_XNVME 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@162 -- # : 0 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@163 -- # export SPDK_TEST_ACCEL_DSA 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@164 -- # : 0 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@165 -- # export SPDK_TEST_ACCEL_IAA 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@167 -- # : 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@168 -- # export SPDK_TEST_FUZZER_TARGET 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@169 -- # : 0 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@170 -- # export SPDK_TEST_NVMF_MDNS 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@171 -- # : 0 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@172 -- # export SPDK_JSONRPC_GO_CLIENT 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@175 -- # export SPDK_LIB_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/lib 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@175 -- # SPDK_LIB_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/lib 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@176 -- # export DPDK_LIB_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build/lib 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@176 -- # DPDK_LIB_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build/lib 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@177 -- # export VFIO_LIB_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@177 -- # VFIO_LIB_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@178 -- # export LD_LIBRARY_PATH=:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@178 -- # LD_LIBRARY_PATH=:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/crypto-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@181 -- # export PCI_BLOCK_SYNC_ON_RESET=yes 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@181 -- # PCI_BLOCK_SYNC_ON_RESET=yes 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@185 -- # export PYTHONPATH=:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/python 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@185 -- # PYTHONPATH=:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/python 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@189 -- # export PYTHONDONTWRITEBYTECODE=1 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@189 -- # PYTHONDONTWRITEBYTECODE=1 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@193 -- # export ASAN_OPTIONS=new_delete_type_mismatch=0:disable_coredump=0:abort_on_error=1:use_sigaltstack=0 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@193 -- # ASAN_OPTIONS=new_delete_type_mismatch=0:disable_coredump=0:abort_on_error=1:use_sigaltstack=0 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@194 -- # export UBSAN_OPTIONS=halt_on_error=1:print_stacktrace=1:abort_on_error=1:disable_coredump=0:exitcode=134 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@194 -- # UBSAN_OPTIONS=halt_on_error=1:print_stacktrace=1:abort_on_error=1:disable_coredump=0:exitcode=134 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@198 -- # asan_suppression_file=/var/tmp/asan_suppression_file 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@199 -- # rm -rf /var/tmp/asan_suppression_file 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@200 -- # cat 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@236 -- # echo leak:libfuse3.so 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@238 -- # export LSAN_OPTIONS=suppressions=/var/tmp/asan_suppression_file 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@238 -- # LSAN_OPTIONS=suppressions=/var/tmp/asan_suppression_file 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@240 -- # export DEFAULT_RPC_ADDR=/var/tmp/spdk.sock 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@240 -- # DEFAULT_RPC_ADDR=/var/tmp/spdk.sock 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@242 -- # '[' -z /var/spdk/dependencies ']' 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@245 -- # export DEPENDENCY_DIR 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@249 -- # export SPDK_BIN_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@249 -- # SPDK_BIN_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@250 -- # export SPDK_EXAMPLE_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@250 -- # SPDK_EXAMPLE_DIR=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@253 -- # export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@253 -- # QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@254 -- # export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@254 -- # VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@256 -- # export AR_TOOL=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/ar-xnvme-fixer 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@256 -- # AR_TOOL=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/ar-xnvme-fixer 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@259 -- # export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@259 -- # UNBIND_ENTIRE_IOMMU_GROUP=yes 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@262 -- # '[' 0 -eq 0 ']' 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@263 -- # export valgrind= 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@263 -- # valgrind= 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@269 -- # uname -s 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@269 -- # '[' Linux = Linux ']' 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@270 -- # HUGEMEM=4096 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@271 -- # export CLEAR_HUGE=yes 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@271 -- # CLEAR_HUGE=yes 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@272 -- # [[ 1 -eq 1 ]] 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@276 -- # export HUGE_EVEN_ALLOC=yes 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@276 -- # HUGE_EVEN_ALLOC=yes 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@279 -- # MAKE=make 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@280 -- # MAKEFLAGS=-j72 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@296 -- # export HUGEMEM=4096 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@296 -- # HUGEMEM=4096 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@298 -- # NO_HUGE=() 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@299 -- # TEST_MODE= 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@318 -- # [[ -z 1029183 ]] 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@318 -- # kill -0 1029183 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@1680 -- # set_test_storage 2147483648 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@328 -- # [[ -v testdir ]] 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@330 -- # local requested_size=2147483648 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@331 -- # local mount target_dir 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@333 -- # local -A mounts fss sizes avails uses 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@334 -- # local source fs size avail mount use 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@336 -- # local storage_fallback storage_candidates 00:29:42.071 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@338 -- # mktemp -udt spdk.XXXXXX 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@338 -- # storage_fallback=/tmp/spdk.J9NQA3 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@343 -- # storage_candidates=("$testdir" "$storage_fallback/tests/${testdir##*/}" "$storage_fallback") 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@345 -- # [[ -n '' ]] 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@350 -- # [[ -n '' ]] 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@355 -- # mkdir -p /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt /tmp/spdk.J9NQA3/tests/interrupt /tmp/spdk.J9NQA3 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@358 -- # requested_size=2214592512 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@327 -- # df -T 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@327 -- # grep -v Filesystem 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@361 -- # mounts["$mount"]=spdk_devtmpfs 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@361 -- # fss["$mount"]=devtmpfs 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@362 -- # avails["$mount"]=67108864 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@362 -- # sizes["$mount"]=67108864 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@363 -- # uses["$mount"]=0 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@361 -- # mounts["$mount"]=/dev/pmem0 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@361 -- # fss["$mount"]=ext2 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@362 -- # avails["$mount"]=946290688 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@362 -- # sizes["$mount"]=5284429824 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@363 -- # uses["$mount"]=4338139136 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@361 -- # mounts["$mount"]=spdk_root 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@361 -- # fss["$mount"]=overlay 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@362 -- # avails["$mount"]=88761401344 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@362 -- # sizes["$mount"]=94508515328 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@363 -- # uses["$mount"]=5747113984 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@361 -- # mounts["$mount"]=tmpfs 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@361 -- # fss["$mount"]=tmpfs 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@362 -- # avails["$mount"]=47249547264 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@362 -- # sizes["$mount"]=47254257664 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@363 -- # uses["$mount"]=4710400 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@361 -- # mounts["$mount"]=tmpfs 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@361 -- # fss["$mount"]=tmpfs 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@362 -- # avails["$mount"]=18892267520 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@362 -- # sizes["$mount"]=18901704704 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@363 -- # uses["$mount"]=9437184 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@361 -- # mounts["$mount"]=tmpfs 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@361 -- # fss["$mount"]=tmpfs 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@362 -- # avails["$mount"]=47253434368 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@362 -- # sizes["$mount"]=47254257664 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@363 -- # uses["$mount"]=823296 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@361 -- # mounts["$mount"]=tmpfs 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@361 -- # fss["$mount"]=tmpfs 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@362 -- # avails["$mount"]=9450844160 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@362 -- # sizes["$mount"]=9450848256 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@363 -- # uses["$mount"]=4096 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@366 -- # printf '* Looking for test storage...\n' 00:29:42.072 * Looking for test storage... 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@368 -- # local target_space new_size 00:29:42.072 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@369 -- # for target_dir in "${storage_candidates[@]}" 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@372 -- # df /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@372 -- # awk '$1 !~ /Filesystem/{print $6}' 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@372 -- # mount=/ 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@374 -- # target_space=88761401344 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@375 -- # (( target_space == 0 || target_space < requested_size )) 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@378 -- # (( target_space >= requested_size )) 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@380 -- # [[ overlay == tmpfs ]] 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@380 -- # [[ overlay == ramfs ]] 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@380 -- # [[ / == / ]] 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@381 -- # new_size=7961706496 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@382 -- # (( new_size * 100 / sizes[/] > 95 )) 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@387 -- # export SPDK_TEST_STORAGE=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@387 -- # SPDK_TEST_STORAGE=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@388 -- # printf '* Found test storage at %s\n' /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt 00:29:42.331 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@389 -- # return 0 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@1682 -- # set -o errtrace 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@1683 -- # shopt -s extdebug 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@1684 -- # trap 'trap - ERR; print_backtrace >&2' ERR 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@1686 -- # PS4=' \t ${test_domain:-} -- ${BASH_SOURCE#${BASH_SOURCE%/*/*}/}@${LINENO} -- \$ ' 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@1687 -- # true 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@1689 -- # xtrace_fd 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@25 -- # [[ -n 13 ]] 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@25 -- # [[ -e /proc/self/fd/13 ]] 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@27 -- # exec 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@29 -- # exec 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@31 -- # xtrace_restore 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@16 -- # unset -v 'X_STACK[0 - 1 < 0 ? 0 : 0 - 1]' 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@17 -- # (( 0 == 0 )) 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@18 -- # set -x 00:29:42.331 11:40:25 reap_unregistered_poller -- interrupt/interrupt_common.sh@8 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/common.sh 00:29:42.331 11:40:25 reap_unregistered_poller -- interrupt/interrupt_common.sh@10 -- # rpc_py=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:29:42.331 11:40:25 reap_unregistered_poller -- interrupt/interrupt_common.sh@12 -- # r0_mask=0x1 00:29:42.331 11:40:25 reap_unregistered_poller -- interrupt/interrupt_common.sh@13 -- # r1_mask=0x2 00:29:42.331 11:40:25 reap_unregistered_poller -- interrupt/interrupt_common.sh@14 -- # r2_mask=0x4 00:29:42.331 11:40:25 reap_unregistered_poller -- interrupt/interrupt_common.sh@16 -- # cpu_server_mask=0x07 00:29:42.331 11:40:25 reap_unregistered_poller -- interrupt/interrupt_common.sh@17 -- # rpc_server_addr=/var/tmp/spdk.sock 00:29:42.331 11:40:25 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@14 -- # export PYTHONPATH=:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/examples/interrupt_tgt 00:29:42.331 11:40:25 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@14 -- # PYTHONPATH=:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/crypto-phy-autotest/spdk/python:/var/jenkins/workspace/crypto-phy-autotest/spdk/examples/interrupt_tgt 00:29:42.331 11:40:25 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@17 -- # start_intr_tgt 00:29:42.331 11:40:25 reap_unregistered_poller -- interrupt/interrupt_common.sh@20 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:42.331 11:40:25 reap_unregistered_poller -- interrupt/interrupt_common.sh@21 -- # local cpu_mask=0x07 00:29:42.331 11:40:25 reap_unregistered_poller -- interrupt/interrupt_common.sh@24 -- # intr_tgt_pid=1029225 00:29:42.331 11:40:25 reap_unregistered_poller -- interrupt/interrupt_common.sh@25 -- # trap 'killprocess "$intr_tgt_pid"; cleanup; exit 1' SIGINT SIGTERM EXIT 00:29:42.331 11:40:25 reap_unregistered_poller -- interrupt/interrupt_common.sh@23 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/interrupt_tgt -m 0x07 -r /var/tmp/spdk.sock -E -g 00:29:42.331 11:40:25 reap_unregistered_poller -- interrupt/interrupt_common.sh@26 -- # waitforlisten 1029225 /var/tmp/spdk.sock 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@829 -- # '[' -z 1029225 ']' 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@834 -- # local max_retries=100 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:42.331 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@838 -- # xtrace_disable 00:29:42.331 11:40:25 reap_unregistered_poller -- common/autotest_common.sh@10 -- # set +x 00:29:42.331 [2024-07-15 11:40:25.719424] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:29:42.331 [2024-07-15 11:40:25.719498] [ DPDK EAL parameters: interrupt_tgt --no-shconf -c 0x07 --single-file-segments --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1029225 ] 00:29:42.331 [2024-07-15 11:40:25.852731] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:29:42.591 [2024-07-15 11:40:25.952028] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:29:42.591 [2024-07-15 11:40:25.952112] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:29:42.591 [2024-07-15 11:40:25.952117] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:29:42.591 [2024-07-15 11:40:26.026258] thread.c:2099:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:29:43.158 11:40:26 reap_unregistered_poller -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:29:43.158 11:40:26 reap_unregistered_poller -- common/autotest_common.sh@862 -- # return 0 00:29:43.158 11:40:26 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@20 -- # rpc_cmd thread_get_pollers 00:29:43.158 11:40:26 reap_unregistered_poller -- common/autotest_common.sh@559 -- # xtrace_disable 00:29:43.158 11:40:26 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@20 -- # jq -r '.threads[0]' 00:29:43.158 11:40:26 reap_unregistered_poller -- common/autotest_common.sh@10 -- # set +x 00:29:43.158 11:40:26 reap_unregistered_poller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:29:43.158 11:40:26 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@20 -- # app_thread='{ 00:29:43.158 "name": "app_thread", 00:29:43.158 "id": 1, 00:29:43.158 "active_pollers": [], 00:29:43.158 "timed_pollers": [ 00:29:43.158 { 00:29:43.158 "name": "rpc_subsystem_poll_servers", 00:29:43.158 "id": 1, 00:29:43.158 "state": "waiting", 00:29:43.158 "run_count": 0, 00:29:43.158 "busy_count": 0, 00:29:43.158 "period_ticks": 9200000 00:29:43.158 } 00:29:43.158 ], 00:29:43.158 "paused_pollers": [] 00:29:43.158 }' 00:29:43.158 11:40:26 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@21 -- # jq -r '.active_pollers[].name' 00:29:43.158 11:40:26 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@21 -- # native_pollers= 00:29:43.158 11:40:26 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@22 -- # native_pollers+=' ' 00:29:43.158 11:40:26 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@23 -- # jq -r '.timed_pollers[].name' 00:29:43.158 11:40:26 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@23 -- # native_pollers+=rpc_subsystem_poll_servers 00:29:43.158 11:40:26 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@28 -- # setup_bdev_aio 00:29:43.158 11:40:26 reap_unregistered_poller -- interrupt/common.sh@75 -- # uname -s 00:29:43.158 11:40:26 reap_unregistered_poller -- interrupt/common.sh@75 -- # [[ Linux != \F\r\e\e\B\S\D ]] 00:29:43.158 11:40:26 reap_unregistered_poller -- interrupt/common.sh@76 -- # dd if=/dev/zero of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/aiofile bs=2048 count=5000 00:29:43.416 5000+0 records in 00:29:43.416 5000+0 records out 00:29:43.416 10240000 bytes (10 MB, 9.8 MiB) copied, 0.0277009 s, 370 MB/s 00:29:43.416 11:40:26 reap_unregistered_poller -- interrupt/common.sh@77 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/aiofile AIO0 2048 00:29:43.675 AIO0 00:29:43.675 11:40:27 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@33 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:29:43.934 11:40:27 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@34 -- # sleep 0.1 00:29:43.934 11:40:27 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@37 -- # rpc_cmd thread_get_pollers 00:29:43.934 11:40:27 reap_unregistered_poller -- common/autotest_common.sh@559 -- # xtrace_disable 00:29:43.934 11:40:27 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@37 -- # jq -r '.threads[0]' 00:29:43.934 11:40:27 reap_unregistered_poller -- common/autotest_common.sh@10 -- # set +x 00:29:43.934 11:40:27 reap_unregistered_poller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:29:43.934 11:40:27 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@37 -- # app_thread='{ 00:29:43.934 "name": "app_thread", 00:29:43.934 "id": 1, 00:29:43.934 "active_pollers": [], 00:29:43.934 "timed_pollers": [ 00:29:43.934 { 00:29:43.934 "name": "rpc_subsystem_poll_servers", 00:29:43.934 "id": 1, 00:29:43.934 "state": "waiting", 00:29:43.934 "run_count": 0, 00:29:43.934 "busy_count": 0, 00:29:43.934 "period_ticks": 9200000 00:29:43.934 } 00:29:43.934 ], 00:29:43.934 "paused_pollers": [] 00:29:43.934 }' 00:29:43.934 11:40:27 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@38 -- # jq -r '.active_pollers[].name' 00:29:43.934 11:40:27 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@38 -- # remaining_pollers= 00:29:43.934 11:40:27 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@39 -- # remaining_pollers+=' ' 00:29:43.934 11:40:27 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@40 -- # jq -r '.timed_pollers[].name' 00:29:43.934 11:40:27 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@40 -- # remaining_pollers+=rpc_subsystem_poll_servers 00:29:43.934 11:40:27 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@44 -- # [[ rpc_subsystem_poll_servers == \ \r\p\c\_\s\u\b\s\y\s\t\e\m\_\p\o\l\l\_\s\e\r\v\e\r\s ]] 00:29:43.934 11:40:27 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@46 -- # trap - SIGINT SIGTERM EXIT 00:29:43.934 11:40:27 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@47 -- # killprocess 1029225 00:29:43.934 11:40:27 reap_unregistered_poller -- common/autotest_common.sh@948 -- # '[' -z 1029225 ']' 00:29:43.934 11:40:27 reap_unregistered_poller -- common/autotest_common.sh@952 -- # kill -0 1029225 00:29:43.934 11:40:27 reap_unregistered_poller -- common/autotest_common.sh@953 -- # uname 00:29:43.934 11:40:27 reap_unregistered_poller -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:29:43.934 11:40:27 reap_unregistered_poller -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1029225 00:29:44.192 11:40:27 reap_unregistered_poller -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:29:44.192 11:40:27 reap_unregistered_poller -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:29:44.192 11:40:27 reap_unregistered_poller -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1029225' 00:29:44.192 killing process with pid 1029225 00:29:44.192 11:40:27 reap_unregistered_poller -- common/autotest_common.sh@967 -- # kill 1029225 00:29:44.192 11:40:27 reap_unregistered_poller -- common/autotest_common.sh@972 -- # wait 1029225 00:29:44.192 11:40:27 reap_unregistered_poller -- interrupt/reap_unregistered_poller.sh@48 -- # cleanup 00:29:44.192 11:40:27 reap_unregistered_poller -- interrupt/common.sh@6 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/interrupt/aiofile 00:29:44.192 00:29:44.192 real 0m2.390s 00:29:44.192 user 0m1.480s 00:29:44.192 sys 0m0.664s 00:29:44.192 11:40:27 reap_unregistered_poller -- common/autotest_common.sh@1124 -- # xtrace_disable 00:29:44.192 11:40:27 reap_unregistered_poller -- common/autotest_common.sh@10 -- # set +x 00:29:44.192 ************************************ 00:29:44.192 END TEST reap_unregistered_poller 00:29:44.192 ************************************ 00:29:44.451 11:40:27 -- common/autotest_common.sh@1142 -- # return 0 00:29:44.451 11:40:27 -- spdk/autotest.sh@198 -- # uname -s 00:29:44.451 11:40:27 -- spdk/autotest.sh@198 -- # [[ Linux == Linux ]] 00:29:44.451 11:40:27 -- spdk/autotest.sh@199 -- # [[ 1 -eq 1 ]] 00:29:44.451 11:40:27 -- spdk/autotest.sh@205 -- # [[ 1 -eq 0 ]] 00:29:44.451 11:40:27 -- spdk/autotest.sh@211 -- # '[' 0 -eq 1 ']' 00:29:44.451 11:40:27 -- spdk/autotest.sh@256 -- # '[' 0 -eq 1 ']' 00:29:44.451 11:40:27 -- spdk/autotest.sh@260 -- # timing_exit lib 00:29:44.451 11:40:27 -- common/autotest_common.sh@728 -- # xtrace_disable 00:29:44.451 11:40:27 -- common/autotest_common.sh@10 -- # set +x 00:29:44.451 11:40:27 -- spdk/autotest.sh@262 -- # '[' 0 -eq 1 ']' 00:29:44.451 11:40:27 -- spdk/autotest.sh@270 -- # '[' 0 -eq 1 ']' 00:29:44.451 11:40:27 -- spdk/autotest.sh@279 -- # '[' 0 -eq 1 ']' 00:29:44.451 11:40:27 -- spdk/autotest.sh@308 -- # '[' 0 -eq 1 ']' 00:29:44.451 11:40:27 -- spdk/autotest.sh@312 -- # '[' 0 -eq 1 ']' 00:29:44.451 11:40:27 -- spdk/autotest.sh@316 -- # '[' 0 -eq 1 ']' 00:29:44.451 11:40:27 -- spdk/autotest.sh@321 -- # '[' 0 -eq 1 ']' 00:29:44.451 11:40:27 -- spdk/autotest.sh@330 -- # '[' 0 -eq 1 ']' 00:29:44.451 11:40:27 -- spdk/autotest.sh@335 -- # '[' 0 -eq 1 ']' 00:29:44.451 11:40:27 -- spdk/autotest.sh@339 -- # '[' 0 -eq 1 ']' 00:29:44.451 11:40:27 -- spdk/autotest.sh@343 -- # '[' 0 -eq 1 ']' 00:29:44.451 11:40:27 -- spdk/autotest.sh@347 -- # '[' 1 -eq 1 ']' 00:29:44.451 11:40:27 -- spdk/autotest.sh@348 -- # run_test compress_compdev /var/jenkins/workspace/crypto-phy-autotest/spdk/test/compress/compress.sh compdev 00:29:44.451 11:40:27 -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:29:44.451 11:40:27 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:29:44.451 11:40:27 -- common/autotest_common.sh@10 -- # set +x 00:29:44.451 ************************************ 00:29:44.451 START TEST compress_compdev 00:29:44.451 ************************************ 00:29:44.451 11:40:27 compress_compdev -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/compress/compress.sh compdev 00:29:44.451 * Looking for test storage... 00:29:44.451 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/compress 00:29:44.451 11:40:28 compress_compdev -- compress/compress.sh@13 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/nvmf/common.sh 00:29:44.451 11:40:28 compress_compdev -- nvmf/common.sh@7 -- # uname -s 00:29:44.451 11:40:28 compress_compdev -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:29:44.451 11:40:28 compress_compdev -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:29:44.451 11:40:28 compress_compdev -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:29:44.451 11:40:28 compress_compdev -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:29:44.451 11:40:28 compress_compdev -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:29:44.451 11:40:28 compress_compdev -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:29:44.451 11:40:28 compress_compdev -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:29:44.452 11:40:28 compress_compdev -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:29:44.452 11:40:28 compress_compdev -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:29:44.452 11:40:28 compress_compdev -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:29:44.452 11:40:28 compress_compdev -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:809e2efd-7f71-e711-906e-0017a4403562 00:29:44.452 11:40:28 compress_compdev -- nvmf/common.sh@18 -- # NVME_HOSTID=809e2efd-7f71-e711-906e-0017a4403562 00:29:44.452 11:40:28 compress_compdev -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:29:44.452 11:40:28 compress_compdev -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:29:44.452 11:40:28 compress_compdev -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:29:44.452 11:40:28 compress_compdev -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:29:44.452 11:40:28 compress_compdev -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/common.sh 00:29:44.452 11:40:28 compress_compdev -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:29:44.452 11:40:28 compress_compdev -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:29:44.452 11:40:28 compress_compdev -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:29:44.452 11:40:28 compress_compdev -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:44.452 11:40:28 compress_compdev -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:44.452 11:40:28 compress_compdev -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:44.452 11:40:28 compress_compdev -- paths/export.sh@5 -- # export PATH 00:29:44.452 11:40:28 compress_compdev -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:44.452 11:40:28 compress_compdev -- nvmf/common.sh@47 -- # : 0 00:29:44.452 11:40:28 compress_compdev -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:29:44.452 11:40:28 compress_compdev -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:29:44.452 11:40:28 compress_compdev -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:29:44.452 11:40:28 compress_compdev -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:29:44.452 11:40:28 compress_compdev -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:29:44.710 11:40:28 compress_compdev -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:29:44.710 11:40:28 compress_compdev -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:29:44.710 11:40:28 compress_compdev -- nvmf/common.sh@51 -- # have_pci_nics=0 00:29:44.710 11:40:28 compress_compdev -- compress/compress.sh@17 -- # rpc_py=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:29:44.710 11:40:28 compress_compdev -- compress/compress.sh@81 -- # mkdir -p /tmp/pmem 00:29:44.710 11:40:28 compress_compdev -- compress/compress.sh@82 -- # test_type=compdev 00:29:44.710 11:40:28 compress_compdev -- compress/compress.sh@86 -- # run_bdevperf 32 4096 3 00:29:44.710 11:40:28 compress_compdev -- compress/compress.sh@66 -- # [[ compdev == \c\o\m\p\d\e\v ]] 00:29:44.710 11:40:28 compress_compdev -- compress/compress.sh@71 -- # bdevperf_pid=1029669 00:29:44.710 11:40:28 compress_compdev -- compress/compress.sh@72 -- # trap 'killprocess $bdevperf_pid; error_cleanup; exit 1' SIGINT SIGTERM EXIT 00:29:44.710 11:40:28 compress_compdev -- compress/compress.sh@73 -- # waitforlisten 1029669 00:29:44.710 11:40:28 compress_compdev -- common/autotest_common.sh@829 -- # '[' -z 1029669 ']' 00:29:44.710 11:40:28 compress_compdev -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:44.710 11:40:28 compress_compdev -- common/autotest_common.sh@834 -- # local max_retries=100 00:29:44.710 11:40:28 compress_compdev -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:44.710 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:44.710 11:40:28 compress_compdev -- common/autotest_common.sh@838 -- # xtrace_disable 00:29:44.710 11:40:28 compress_compdev -- compress/compress.sh@67 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -z -q 32 -o 4096 -w verify -t 3 -C -m 0x6 -c /var/jenkins/workspace/crypto-phy-autotest/spdk/test/compress/dpdk.json 00:29:44.710 11:40:28 compress_compdev -- common/autotest_common.sh@10 -- # set +x 00:29:44.710 [2024-07-15 11:40:28.113924] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:29:44.710 [2024-07-15 11:40:28.114003] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x6 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1029669 ] 00:29:44.710 [2024-07-15 11:40:28.233892] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:29:44.967 [2024-07-15 11:40:28.338675] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:29:44.967 [2024-07-15 11:40:28.338681] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:29:45.532 [2024-07-15 11:40:29.087836] accel_dpdk_compressdev.c: 296:accel_init_compress_drivers: *NOTICE*: initialized QAT PMD 00:29:45.790 11:40:29 compress_compdev -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:29:45.790 11:40:29 compress_compdev -- common/autotest_common.sh@862 -- # return 0 00:29:45.790 11:40:29 compress_compdev -- compress/compress.sh@74 -- # create_vols 00:29:45.790 11:40:29 compress_compdev -- compress/compress.sh@34 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/gen_nvme.sh 00:29:45.790 11:40:29 compress_compdev -- compress/compress.sh@34 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py load_subsystem_config 00:29:46.365 [2024-07-15 11:40:29.726921] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x144c3c0 PMD being used: compress_qat 00:29:46.365 11:40:29 compress_compdev -- compress/compress.sh@35 -- # waitforbdev Nvme0n1 00:29:46.365 11:40:29 compress_compdev -- common/autotest_common.sh@897 -- # local bdev_name=Nvme0n1 00:29:46.365 11:40:29 compress_compdev -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:29:46.365 11:40:29 compress_compdev -- common/autotest_common.sh@899 -- # local i 00:29:46.365 11:40:29 compress_compdev -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:29:46.365 11:40:29 compress_compdev -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:29:46.365 11:40:29 compress_compdev -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:29:46.623 11:40:30 compress_compdev -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b Nvme0n1 -t 2000 00:29:46.881 [ 00:29:46.881 { 00:29:46.881 "name": "Nvme0n1", 00:29:46.881 "aliases": [ 00:29:46.881 "01000000-0000-0000-5cd2-e43197705251" 00:29:46.881 ], 00:29:46.881 "product_name": "NVMe disk", 00:29:46.881 "block_size": 512, 00:29:46.881 "num_blocks": 15002931888, 00:29:46.881 "uuid": "01000000-0000-0000-5cd2-e43197705251", 00:29:46.881 "assigned_rate_limits": { 00:29:46.881 "rw_ios_per_sec": 0, 00:29:46.881 "rw_mbytes_per_sec": 0, 00:29:46.881 "r_mbytes_per_sec": 0, 00:29:46.881 "w_mbytes_per_sec": 0 00:29:46.881 }, 00:29:46.881 "claimed": false, 00:29:46.881 "zoned": false, 00:29:46.881 "supported_io_types": { 00:29:46.881 "read": true, 00:29:46.881 "write": true, 00:29:46.881 "unmap": true, 00:29:46.881 "flush": true, 00:29:46.881 "reset": true, 00:29:46.881 "nvme_admin": true, 00:29:46.881 "nvme_io": true, 00:29:46.881 "nvme_io_md": false, 00:29:46.881 "write_zeroes": true, 00:29:46.881 "zcopy": false, 00:29:46.881 "get_zone_info": false, 00:29:46.881 "zone_management": false, 00:29:46.881 "zone_append": false, 00:29:46.881 "compare": false, 00:29:46.881 "compare_and_write": false, 00:29:46.881 "abort": true, 00:29:46.881 "seek_hole": false, 00:29:46.881 "seek_data": false, 00:29:46.881 "copy": false, 00:29:46.881 "nvme_iov_md": false 00:29:46.881 }, 00:29:46.881 "driver_specific": { 00:29:46.881 "nvme": [ 00:29:46.881 { 00:29:46.881 "pci_address": "0000:5e:00.0", 00:29:46.881 "trid": { 00:29:46.881 "trtype": "PCIe", 00:29:46.881 "traddr": "0000:5e:00.0" 00:29:46.881 }, 00:29:46.881 "ctrlr_data": { 00:29:46.881 "cntlid": 0, 00:29:46.881 "vendor_id": "0x8086", 00:29:46.881 "model_number": "INTEL SSDPF2KX076TZO", 00:29:46.881 "serial_number": "PHAC0301002G7P6CGN", 00:29:46.881 "firmware_revision": "JCV10200", 00:29:46.881 "subnqn": "nqn.2020-07.com.intel:PHAC0301002G7P6CGN ", 00:29:46.881 "oacs": { 00:29:46.881 "security": 1, 00:29:46.881 "format": 1, 00:29:46.881 "firmware": 1, 00:29:46.881 "ns_manage": 1 00:29:46.881 }, 00:29:46.881 "multi_ctrlr": false, 00:29:46.881 "ana_reporting": false 00:29:46.881 }, 00:29:46.881 "vs": { 00:29:46.881 "nvme_version": "1.3" 00:29:46.881 }, 00:29:46.881 "ns_data": { 00:29:46.881 "id": 1, 00:29:46.881 "can_share": false 00:29:46.881 }, 00:29:46.881 "security": { 00:29:46.881 "opal": true 00:29:46.881 } 00:29:46.881 } 00:29:46.881 ], 00:29:46.881 "mp_policy": "active_passive" 00:29:46.881 } 00:29:46.881 } 00:29:46.881 ] 00:29:46.881 11:40:30 compress_compdev -- common/autotest_common.sh@905 -- # return 0 00:29:46.881 11:40:30 compress_compdev -- compress/compress.sh@37 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --clear-method none Nvme0n1 lvs0 00:29:47.138 [2024-07-15 11:40:30.488761] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x12b10d0 PMD being used: compress_qat 00:29:49.662 5472b5dd-de10-45ea-8ecc-f5e648656e9d 00:29:49.662 11:40:32 compress_compdev -- compress/compress.sh@38 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -t -l lvs0 lv0 100 00:29:49.663 78aa9267-00d5-4e32-a430-17e509091577 00:29:49.663 11:40:32 compress_compdev -- compress/compress.sh@39 -- # waitforbdev lvs0/lv0 00:29:49.663 11:40:32 compress_compdev -- common/autotest_common.sh@897 -- # local bdev_name=lvs0/lv0 00:29:49.663 11:40:32 compress_compdev -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:29:49.663 11:40:32 compress_compdev -- common/autotest_common.sh@899 -- # local i 00:29:49.663 11:40:32 compress_compdev -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:29:49.663 11:40:32 compress_compdev -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:29:49.663 11:40:32 compress_compdev -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:29:49.663 11:40:33 compress_compdev -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b lvs0/lv0 -t 2000 00:29:49.919 [ 00:29:49.919 { 00:29:49.919 "name": "78aa9267-00d5-4e32-a430-17e509091577", 00:29:49.919 "aliases": [ 00:29:49.919 "lvs0/lv0" 00:29:49.919 ], 00:29:49.919 "product_name": "Logical Volume", 00:29:49.919 "block_size": 512, 00:29:49.919 "num_blocks": 204800, 00:29:49.919 "uuid": "78aa9267-00d5-4e32-a430-17e509091577", 00:29:49.919 "assigned_rate_limits": { 00:29:49.919 "rw_ios_per_sec": 0, 00:29:49.919 "rw_mbytes_per_sec": 0, 00:29:49.919 "r_mbytes_per_sec": 0, 00:29:49.919 "w_mbytes_per_sec": 0 00:29:49.919 }, 00:29:49.919 "claimed": false, 00:29:49.919 "zoned": false, 00:29:49.919 "supported_io_types": { 00:29:49.919 "read": true, 00:29:49.919 "write": true, 00:29:49.919 "unmap": true, 00:29:49.919 "flush": false, 00:29:49.919 "reset": true, 00:29:49.919 "nvme_admin": false, 00:29:49.919 "nvme_io": false, 00:29:49.919 "nvme_io_md": false, 00:29:49.919 "write_zeroes": true, 00:29:49.919 "zcopy": false, 00:29:49.919 "get_zone_info": false, 00:29:49.919 "zone_management": false, 00:29:49.919 "zone_append": false, 00:29:49.919 "compare": false, 00:29:49.919 "compare_and_write": false, 00:29:49.919 "abort": false, 00:29:49.919 "seek_hole": true, 00:29:49.919 "seek_data": true, 00:29:49.919 "copy": false, 00:29:49.919 "nvme_iov_md": false 00:29:49.920 }, 00:29:49.920 "driver_specific": { 00:29:49.920 "lvol": { 00:29:49.920 "lvol_store_uuid": "5472b5dd-de10-45ea-8ecc-f5e648656e9d", 00:29:49.920 "base_bdev": "Nvme0n1", 00:29:49.920 "thin_provision": true, 00:29:49.920 "num_allocated_clusters": 0, 00:29:49.920 "snapshot": false, 00:29:49.920 "clone": false, 00:29:49.920 "esnap_clone": false 00:29:49.920 } 00:29:49.920 } 00:29:49.920 } 00:29:49.920 ] 00:29:49.920 11:40:33 compress_compdev -- common/autotest_common.sh@905 -- # return 0 00:29:49.920 11:40:33 compress_compdev -- compress/compress.sh@41 -- # '[' -z '' ']' 00:29:49.920 11:40:33 compress_compdev -- compress/compress.sh@42 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_compress_create -b lvs0/lv0 -p /tmp/pmem 00:29:50.178 [2024-07-15 11:40:33.579254] vbdev_compress.c:1016:vbdev_compress_claim: *NOTICE*: registered io_device and virtual bdev for: COMP_lvs0/lv0 00:29:50.178 COMP_lvs0/lv0 00:29:50.178 11:40:33 compress_compdev -- compress/compress.sh@46 -- # waitforbdev COMP_lvs0/lv0 00:29:50.178 11:40:33 compress_compdev -- common/autotest_common.sh@897 -- # local bdev_name=COMP_lvs0/lv0 00:29:50.178 11:40:33 compress_compdev -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:29:50.178 11:40:33 compress_compdev -- common/autotest_common.sh@899 -- # local i 00:29:50.178 11:40:33 compress_compdev -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:29:50.178 11:40:33 compress_compdev -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:29:50.178 11:40:33 compress_compdev -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:29:50.436 11:40:33 compress_compdev -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b COMP_lvs0/lv0 -t 2000 00:29:50.694 [ 00:29:50.694 { 00:29:50.694 "name": "COMP_lvs0/lv0", 00:29:50.694 "aliases": [ 00:29:50.694 "b1754044-0747-59ef-8c45-ecd8c7f14068" 00:29:50.694 ], 00:29:50.694 "product_name": "compress", 00:29:50.694 "block_size": 512, 00:29:50.694 "num_blocks": 200704, 00:29:50.694 "uuid": "b1754044-0747-59ef-8c45-ecd8c7f14068", 00:29:50.694 "assigned_rate_limits": { 00:29:50.694 "rw_ios_per_sec": 0, 00:29:50.694 "rw_mbytes_per_sec": 0, 00:29:50.694 "r_mbytes_per_sec": 0, 00:29:50.694 "w_mbytes_per_sec": 0 00:29:50.694 }, 00:29:50.694 "claimed": false, 00:29:50.694 "zoned": false, 00:29:50.694 "supported_io_types": { 00:29:50.694 "read": true, 00:29:50.694 "write": true, 00:29:50.694 "unmap": false, 00:29:50.694 "flush": false, 00:29:50.694 "reset": false, 00:29:50.694 "nvme_admin": false, 00:29:50.694 "nvme_io": false, 00:29:50.694 "nvme_io_md": false, 00:29:50.694 "write_zeroes": true, 00:29:50.694 "zcopy": false, 00:29:50.694 "get_zone_info": false, 00:29:50.694 "zone_management": false, 00:29:50.694 "zone_append": false, 00:29:50.694 "compare": false, 00:29:50.694 "compare_and_write": false, 00:29:50.694 "abort": false, 00:29:50.694 "seek_hole": false, 00:29:50.694 "seek_data": false, 00:29:50.694 "copy": false, 00:29:50.694 "nvme_iov_md": false 00:29:50.694 }, 00:29:50.694 "driver_specific": { 00:29:50.694 "compress": { 00:29:50.694 "name": "COMP_lvs0/lv0", 00:29:50.694 "base_bdev_name": "78aa9267-00d5-4e32-a430-17e509091577" 00:29:50.694 } 00:29:50.694 } 00:29:50.694 } 00:29:50.694 ] 00:29:50.694 11:40:34 compress_compdev -- common/autotest_common.sh@905 -- # return 0 00:29:50.694 11:40:34 compress_compdev -- compress/compress.sh@75 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:29:50.694 [2024-07-15 11:40:34.193636] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x7f3b141b15c0 PMD being used: compress_qat 00:29:50.694 [2024-07-15 11:40:34.195879] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x1449670 PMD being used: compress_qat 00:29:50.694 Running I/O for 3 seconds... 00:29:53.987 00:29:53.987 Latency(us) 00:29:53.987 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:53.987 Job: COMP_lvs0/lv0 (Core Mask 0x2, workload: verify, depth: 32, IO size: 4096) 00:29:53.987 Verification LBA range: start 0x0 length 0x3100 00:29:53.987 COMP_lvs0/lv0 : 3.00 5118.60 19.99 0.00 0.00 6198.53 555.63 6097.70 00:29:53.987 Job: COMP_lvs0/lv0 (Core Mask 0x4, workload: verify, depth: 32, IO size: 4096) 00:29:53.987 Verification LBA range: start 0x3100 length 0x3100 00:29:53.987 COMP_lvs0/lv0 : 3.00 5401.58 21.10 0.00 0.00 5888.03 318.78 5698.78 00:29:53.987 =================================================================================================================== 00:29:53.987 Total : 10520.18 41.09 0.00 0.00 6039.12 318.78 6097.70 00:29:53.987 0 00:29:53.987 11:40:37 compress_compdev -- compress/compress.sh@76 -- # destroy_vols 00:29:53.987 11:40:37 compress_compdev -- compress/compress.sh@29 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_compress_delete COMP_lvs0/lv0 00:29:53.987 11:40:37 compress_compdev -- compress/compress.sh@30 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -l lvs0 00:29:54.246 11:40:37 compress_compdev -- compress/compress.sh@77 -- # trap - SIGINT SIGTERM EXIT 00:29:54.246 11:40:37 compress_compdev -- compress/compress.sh@78 -- # killprocess 1029669 00:29:54.246 11:40:37 compress_compdev -- common/autotest_common.sh@948 -- # '[' -z 1029669 ']' 00:29:54.246 11:40:37 compress_compdev -- common/autotest_common.sh@952 -- # kill -0 1029669 00:29:54.246 11:40:37 compress_compdev -- common/autotest_common.sh@953 -- # uname 00:29:54.246 11:40:37 compress_compdev -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:29:54.246 11:40:37 compress_compdev -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1029669 00:29:54.246 11:40:37 compress_compdev -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:29:54.246 11:40:37 compress_compdev -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:29:54.246 11:40:37 compress_compdev -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1029669' 00:29:54.246 killing process with pid 1029669 00:29:54.246 11:40:37 compress_compdev -- common/autotest_common.sh@967 -- # kill 1029669 00:29:54.246 Received shutdown signal, test time was about 3.000000 seconds 00:29:54.246 00:29:54.246 Latency(us) 00:29:54.246 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:54.246 =================================================================================================================== 00:29:54.246 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:29:54.246 11:40:37 compress_compdev -- common/autotest_common.sh@972 -- # wait 1029669 00:29:57.559 11:40:40 compress_compdev -- compress/compress.sh@87 -- # run_bdevperf 32 4096 3 512 00:29:57.559 11:40:40 compress_compdev -- compress/compress.sh@66 -- # [[ compdev == \c\o\m\p\d\e\v ]] 00:29:57.559 11:40:40 compress_compdev -- compress/compress.sh@71 -- # bdevperf_pid=1031272 00:29:57.559 11:40:40 compress_compdev -- compress/compress.sh@72 -- # trap 'killprocess $bdevperf_pid; error_cleanup; exit 1' SIGINT SIGTERM EXIT 00:29:57.559 11:40:40 compress_compdev -- compress/compress.sh@67 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -z -q 32 -o 4096 -w verify -t 3 -C -m 0x6 -c /var/jenkins/workspace/crypto-phy-autotest/spdk/test/compress/dpdk.json 00:29:57.559 11:40:40 compress_compdev -- compress/compress.sh@73 -- # waitforlisten 1031272 00:29:57.559 11:40:40 compress_compdev -- common/autotest_common.sh@829 -- # '[' -z 1031272 ']' 00:29:57.559 11:40:40 compress_compdev -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:57.559 11:40:40 compress_compdev -- common/autotest_common.sh@834 -- # local max_retries=100 00:29:57.559 11:40:40 compress_compdev -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:57.559 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:57.559 11:40:40 compress_compdev -- common/autotest_common.sh@838 -- # xtrace_disable 00:29:57.559 11:40:40 compress_compdev -- common/autotest_common.sh@10 -- # set +x 00:29:57.559 [2024-07-15 11:40:40.842369] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:29:57.559 [2024-07-15 11:40:40.842439] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x6 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1031272 ] 00:29:57.559 [2024-07-15 11:40:40.961149] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:29:57.559 [2024-07-15 11:40:41.067118] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:29:57.559 [2024-07-15 11:40:41.067125] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:29:58.494 [2024-07-15 11:40:41.813310] accel_dpdk_compressdev.c: 296:accel_init_compress_drivers: *NOTICE*: initialized QAT PMD 00:29:58.494 11:40:41 compress_compdev -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:29:58.494 11:40:41 compress_compdev -- common/autotest_common.sh@862 -- # return 0 00:29:58.494 11:40:41 compress_compdev -- compress/compress.sh@74 -- # create_vols 512 00:29:58.494 11:40:41 compress_compdev -- compress/compress.sh@34 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/gen_nvme.sh 00:29:58.494 11:40:41 compress_compdev -- compress/compress.sh@34 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py load_subsystem_config 00:29:59.060 [2024-07-15 11:40:42.455149] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x15cc3c0 PMD being used: compress_qat 00:29:59.060 11:40:42 compress_compdev -- compress/compress.sh@35 -- # waitforbdev Nvme0n1 00:29:59.060 11:40:42 compress_compdev -- common/autotest_common.sh@897 -- # local bdev_name=Nvme0n1 00:29:59.060 11:40:42 compress_compdev -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:29:59.060 11:40:42 compress_compdev -- common/autotest_common.sh@899 -- # local i 00:29:59.060 11:40:42 compress_compdev -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:29:59.060 11:40:42 compress_compdev -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:29:59.060 11:40:42 compress_compdev -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:29:59.318 11:40:42 compress_compdev -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b Nvme0n1 -t 2000 00:29:59.577 [ 00:29:59.577 { 00:29:59.577 "name": "Nvme0n1", 00:29:59.577 "aliases": [ 00:29:59.577 "01000000-0000-0000-5cd2-e43197705251" 00:29:59.577 ], 00:29:59.577 "product_name": "NVMe disk", 00:29:59.577 "block_size": 512, 00:29:59.577 "num_blocks": 15002931888, 00:29:59.577 "uuid": "01000000-0000-0000-5cd2-e43197705251", 00:29:59.577 "assigned_rate_limits": { 00:29:59.577 "rw_ios_per_sec": 0, 00:29:59.577 "rw_mbytes_per_sec": 0, 00:29:59.577 "r_mbytes_per_sec": 0, 00:29:59.577 "w_mbytes_per_sec": 0 00:29:59.577 }, 00:29:59.577 "claimed": false, 00:29:59.577 "zoned": false, 00:29:59.577 "supported_io_types": { 00:29:59.577 "read": true, 00:29:59.577 "write": true, 00:29:59.577 "unmap": true, 00:29:59.577 "flush": true, 00:29:59.577 "reset": true, 00:29:59.577 "nvme_admin": true, 00:29:59.577 "nvme_io": true, 00:29:59.577 "nvme_io_md": false, 00:29:59.577 "write_zeroes": true, 00:29:59.577 "zcopy": false, 00:29:59.577 "get_zone_info": false, 00:29:59.577 "zone_management": false, 00:29:59.577 "zone_append": false, 00:29:59.577 "compare": false, 00:29:59.577 "compare_and_write": false, 00:29:59.577 "abort": true, 00:29:59.577 "seek_hole": false, 00:29:59.577 "seek_data": false, 00:29:59.577 "copy": false, 00:29:59.577 "nvme_iov_md": false 00:29:59.577 }, 00:29:59.577 "driver_specific": { 00:29:59.577 "nvme": [ 00:29:59.577 { 00:29:59.577 "pci_address": "0000:5e:00.0", 00:29:59.577 "trid": { 00:29:59.577 "trtype": "PCIe", 00:29:59.577 "traddr": "0000:5e:00.0" 00:29:59.577 }, 00:29:59.577 "ctrlr_data": { 00:29:59.577 "cntlid": 0, 00:29:59.577 "vendor_id": "0x8086", 00:29:59.577 "model_number": "INTEL SSDPF2KX076TZO", 00:29:59.577 "serial_number": "PHAC0301002G7P6CGN", 00:29:59.577 "firmware_revision": "JCV10200", 00:29:59.577 "subnqn": "nqn.2020-07.com.intel:PHAC0301002G7P6CGN ", 00:29:59.577 "oacs": { 00:29:59.577 "security": 1, 00:29:59.577 "format": 1, 00:29:59.577 "firmware": 1, 00:29:59.577 "ns_manage": 1 00:29:59.577 }, 00:29:59.577 "multi_ctrlr": false, 00:29:59.577 "ana_reporting": false 00:29:59.577 }, 00:29:59.577 "vs": { 00:29:59.577 "nvme_version": "1.3" 00:29:59.577 }, 00:29:59.577 "ns_data": { 00:29:59.577 "id": 1, 00:29:59.577 "can_share": false 00:29:59.577 }, 00:29:59.577 "security": { 00:29:59.577 "opal": true 00:29:59.577 } 00:29:59.577 } 00:29:59.577 ], 00:29:59.577 "mp_policy": "active_passive" 00:29:59.577 } 00:29:59.577 } 00:29:59.577 ] 00:29:59.577 11:40:42 compress_compdev -- common/autotest_common.sh@905 -- # return 0 00:29:59.577 11:40:42 compress_compdev -- compress/compress.sh@37 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --clear-method none Nvme0n1 lvs0 00:29:59.844 [2024-07-15 11:40:43.216814] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x1431660 PMD being used: compress_qat 00:30:02.374 4eadd9d2-4637-4618-8673-de951c4e2579 00:30:02.374 11:40:45 compress_compdev -- compress/compress.sh@38 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -t -l lvs0 lv0 100 00:30:02.374 b176c997-69be-4763-9197-3888298ca74d 00:30:02.374 11:40:45 compress_compdev -- compress/compress.sh@39 -- # waitforbdev lvs0/lv0 00:30:02.374 11:40:45 compress_compdev -- common/autotest_common.sh@897 -- # local bdev_name=lvs0/lv0 00:30:02.374 11:40:45 compress_compdev -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:30:02.374 11:40:45 compress_compdev -- common/autotest_common.sh@899 -- # local i 00:30:02.374 11:40:45 compress_compdev -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:30:02.374 11:40:45 compress_compdev -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:30:02.374 11:40:45 compress_compdev -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:30:02.374 11:40:45 compress_compdev -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b lvs0/lv0 -t 2000 00:30:02.633 [ 00:30:02.633 { 00:30:02.633 "name": "b176c997-69be-4763-9197-3888298ca74d", 00:30:02.633 "aliases": [ 00:30:02.633 "lvs0/lv0" 00:30:02.633 ], 00:30:02.633 "product_name": "Logical Volume", 00:30:02.633 "block_size": 512, 00:30:02.633 "num_blocks": 204800, 00:30:02.633 "uuid": "b176c997-69be-4763-9197-3888298ca74d", 00:30:02.633 "assigned_rate_limits": { 00:30:02.633 "rw_ios_per_sec": 0, 00:30:02.633 "rw_mbytes_per_sec": 0, 00:30:02.633 "r_mbytes_per_sec": 0, 00:30:02.633 "w_mbytes_per_sec": 0 00:30:02.633 }, 00:30:02.633 "claimed": false, 00:30:02.633 "zoned": false, 00:30:02.633 "supported_io_types": { 00:30:02.633 "read": true, 00:30:02.633 "write": true, 00:30:02.633 "unmap": true, 00:30:02.633 "flush": false, 00:30:02.633 "reset": true, 00:30:02.633 "nvme_admin": false, 00:30:02.633 "nvme_io": false, 00:30:02.633 "nvme_io_md": false, 00:30:02.633 "write_zeroes": true, 00:30:02.633 "zcopy": false, 00:30:02.633 "get_zone_info": false, 00:30:02.633 "zone_management": false, 00:30:02.633 "zone_append": false, 00:30:02.633 "compare": false, 00:30:02.633 "compare_and_write": false, 00:30:02.633 "abort": false, 00:30:02.633 "seek_hole": true, 00:30:02.633 "seek_data": true, 00:30:02.633 "copy": false, 00:30:02.633 "nvme_iov_md": false 00:30:02.633 }, 00:30:02.633 "driver_specific": { 00:30:02.633 "lvol": { 00:30:02.633 "lvol_store_uuid": "4eadd9d2-4637-4618-8673-de951c4e2579", 00:30:02.633 "base_bdev": "Nvme0n1", 00:30:02.633 "thin_provision": true, 00:30:02.633 "num_allocated_clusters": 0, 00:30:02.633 "snapshot": false, 00:30:02.633 "clone": false, 00:30:02.633 "esnap_clone": false 00:30:02.633 } 00:30:02.633 } 00:30:02.633 } 00:30:02.633 ] 00:30:02.633 11:40:46 compress_compdev -- common/autotest_common.sh@905 -- # return 0 00:30:02.633 11:40:46 compress_compdev -- compress/compress.sh@41 -- # '[' -z 512 ']' 00:30:02.633 11:40:46 compress_compdev -- compress/compress.sh@44 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_compress_create -b lvs0/lv0 -p /tmp/pmem -l 512 00:30:02.891 [2024-07-15 11:40:46.423281] vbdev_compress.c:1016:vbdev_compress_claim: *NOTICE*: registered io_device and virtual bdev for: COMP_lvs0/lv0 00:30:02.891 COMP_lvs0/lv0 00:30:02.891 11:40:46 compress_compdev -- compress/compress.sh@46 -- # waitforbdev COMP_lvs0/lv0 00:30:02.891 11:40:46 compress_compdev -- common/autotest_common.sh@897 -- # local bdev_name=COMP_lvs0/lv0 00:30:02.891 11:40:46 compress_compdev -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:30:02.891 11:40:46 compress_compdev -- common/autotest_common.sh@899 -- # local i 00:30:02.891 11:40:46 compress_compdev -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:30:02.891 11:40:46 compress_compdev -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:30:02.891 11:40:46 compress_compdev -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:30:03.149 11:40:46 compress_compdev -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b COMP_lvs0/lv0 -t 2000 00:30:03.408 [ 00:30:03.408 { 00:30:03.408 "name": "COMP_lvs0/lv0", 00:30:03.408 "aliases": [ 00:30:03.408 "8e8544a6-d131-5ddb-b7d0-6e4da48e7c52" 00:30:03.408 ], 00:30:03.408 "product_name": "compress", 00:30:03.408 "block_size": 512, 00:30:03.408 "num_blocks": 200704, 00:30:03.408 "uuid": "8e8544a6-d131-5ddb-b7d0-6e4da48e7c52", 00:30:03.408 "assigned_rate_limits": { 00:30:03.408 "rw_ios_per_sec": 0, 00:30:03.408 "rw_mbytes_per_sec": 0, 00:30:03.408 "r_mbytes_per_sec": 0, 00:30:03.408 "w_mbytes_per_sec": 0 00:30:03.408 }, 00:30:03.408 "claimed": false, 00:30:03.408 "zoned": false, 00:30:03.408 "supported_io_types": { 00:30:03.408 "read": true, 00:30:03.408 "write": true, 00:30:03.408 "unmap": false, 00:30:03.408 "flush": false, 00:30:03.408 "reset": false, 00:30:03.408 "nvme_admin": false, 00:30:03.408 "nvme_io": false, 00:30:03.408 "nvme_io_md": false, 00:30:03.408 "write_zeroes": true, 00:30:03.408 "zcopy": false, 00:30:03.408 "get_zone_info": false, 00:30:03.408 "zone_management": false, 00:30:03.408 "zone_append": false, 00:30:03.408 "compare": false, 00:30:03.408 "compare_and_write": false, 00:30:03.408 "abort": false, 00:30:03.408 "seek_hole": false, 00:30:03.408 "seek_data": false, 00:30:03.408 "copy": false, 00:30:03.408 "nvme_iov_md": false 00:30:03.408 }, 00:30:03.408 "driver_specific": { 00:30:03.408 "compress": { 00:30:03.408 "name": "COMP_lvs0/lv0", 00:30:03.408 "base_bdev_name": "b176c997-69be-4763-9197-3888298ca74d" 00:30:03.408 } 00:30:03.408 } 00:30:03.408 } 00:30:03.408 ] 00:30:03.408 11:40:46 compress_compdev -- common/autotest_common.sh@905 -- # return 0 00:30:03.408 11:40:46 compress_compdev -- compress/compress.sh@75 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:30:03.666 [2024-07-15 11:40:47.013603] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x7f3b041b15c0 PMD being used: compress_qat 00:30:03.666 [2024-07-15 11:40:47.015814] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x15c9770 PMD being used: compress_qat 00:30:03.666 Running I/O for 3 seconds... 00:30:06.950 00:30:06.950 Latency(us) 00:30:06.950 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:30:06.950 Job: COMP_lvs0/lv0 (Core Mask 0x2, workload: verify, depth: 32, IO size: 4096) 00:30:06.950 Verification LBA range: start 0x0 length 0x3100 00:30:06.950 COMP_lvs0/lv0 : 3.00 5132.33 20.05 0.00 0.00 6184.18 516.45 6069.20 00:30:06.950 Job: COMP_lvs0/lv0 (Core Mask 0x4, workload: verify, depth: 32, IO size: 4096) 00:30:06.950 Verification LBA range: start 0x3100 length 0x3100 00:30:06.950 COMP_lvs0/lv0 : 3.00 5402.89 21.11 0.00 0.00 5885.99 354.39 5869.75 00:30:06.950 =================================================================================================================== 00:30:06.950 Total : 10535.22 41.15 0.00 0.00 6031.25 354.39 6069.20 00:30:06.950 0 00:30:06.950 11:40:50 compress_compdev -- compress/compress.sh@76 -- # destroy_vols 00:30:06.950 11:40:50 compress_compdev -- compress/compress.sh@29 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_compress_delete COMP_lvs0/lv0 00:30:06.950 11:40:50 compress_compdev -- compress/compress.sh@30 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -l lvs0 00:30:07.209 11:40:50 compress_compdev -- compress/compress.sh@77 -- # trap - SIGINT SIGTERM EXIT 00:30:07.209 11:40:50 compress_compdev -- compress/compress.sh@78 -- # killprocess 1031272 00:30:07.209 11:40:50 compress_compdev -- common/autotest_common.sh@948 -- # '[' -z 1031272 ']' 00:30:07.209 11:40:50 compress_compdev -- common/autotest_common.sh@952 -- # kill -0 1031272 00:30:07.209 11:40:50 compress_compdev -- common/autotest_common.sh@953 -- # uname 00:30:07.209 11:40:50 compress_compdev -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:30:07.209 11:40:50 compress_compdev -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1031272 00:30:07.209 11:40:50 compress_compdev -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:30:07.209 11:40:50 compress_compdev -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:30:07.209 11:40:50 compress_compdev -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1031272' 00:30:07.209 killing process with pid 1031272 00:30:07.209 11:40:50 compress_compdev -- common/autotest_common.sh@967 -- # kill 1031272 00:30:07.209 Received shutdown signal, test time was about 3.000000 seconds 00:30:07.209 00:30:07.209 Latency(us) 00:30:07.209 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:30:07.209 =================================================================================================================== 00:30:07.209 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:30:07.209 11:40:50 compress_compdev -- common/autotest_common.sh@972 -- # wait 1031272 00:30:10.510 11:40:53 compress_compdev -- compress/compress.sh@88 -- # run_bdevperf 32 4096 3 4096 00:30:10.510 11:40:53 compress_compdev -- compress/compress.sh@66 -- # [[ compdev == \c\o\m\p\d\e\v ]] 00:30:10.510 11:40:53 compress_compdev -- compress/compress.sh@71 -- # bdevperf_pid=1032877 00:30:10.510 11:40:53 compress_compdev -- compress/compress.sh@72 -- # trap 'killprocess $bdevperf_pid; error_cleanup; exit 1' SIGINT SIGTERM EXIT 00:30:10.510 11:40:53 compress_compdev -- compress/compress.sh@67 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -z -q 32 -o 4096 -w verify -t 3 -C -m 0x6 -c /var/jenkins/workspace/crypto-phy-autotest/spdk/test/compress/dpdk.json 00:30:10.510 11:40:53 compress_compdev -- compress/compress.sh@73 -- # waitforlisten 1032877 00:30:10.510 11:40:53 compress_compdev -- common/autotest_common.sh@829 -- # '[' -z 1032877 ']' 00:30:10.510 11:40:53 compress_compdev -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:10.510 11:40:53 compress_compdev -- common/autotest_common.sh@834 -- # local max_retries=100 00:30:10.510 11:40:53 compress_compdev -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:10.510 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:10.510 11:40:53 compress_compdev -- common/autotest_common.sh@838 -- # xtrace_disable 00:30:10.510 11:40:53 compress_compdev -- common/autotest_common.sh@10 -- # set +x 00:30:10.510 [2024-07-15 11:40:53.660426] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:30:10.510 [2024-07-15 11:40:53.660493] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x6 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1032877 ] 00:30:10.510 [2024-07-15 11:40:53.777935] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:30:10.510 [2024-07-15 11:40:53.875604] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:30:10.510 [2024-07-15 11:40:53.875610] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:30:11.078 [2024-07-15 11:40:54.618548] accel_dpdk_compressdev.c: 296:accel_init_compress_drivers: *NOTICE*: initialized QAT PMD 00:30:11.336 11:40:54 compress_compdev -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:30:11.336 11:40:54 compress_compdev -- common/autotest_common.sh@862 -- # return 0 00:30:11.336 11:40:54 compress_compdev -- compress/compress.sh@74 -- # create_vols 4096 00:30:11.336 11:40:54 compress_compdev -- compress/compress.sh@34 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/gen_nvme.sh 00:30:11.336 11:40:54 compress_compdev -- compress/compress.sh@34 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py load_subsystem_config 00:30:11.903 [2024-07-15 11:40:55.254936] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0xeff3c0 PMD being used: compress_qat 00:30:11.903 11:40:55 compress_compdev -- compress/compress.sh@35 -- # waitforbdev Nvme0n1 00:30:11.903 11:40:55 compress_compdev -- common/autotest_common.sh@897 -- # local bdev_name=Nvme0n1 00:30:11.903 11:40:55 compress_compdev -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:30:11.903 11:40:55 compress_compdev -- common/autotest_common.sh@899 -- # local i 00:30:11.903 11:40:55 compress_compdev -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:30:11.903 11:40:55 compress_compdev -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:30:11.903 11:40:55 compress_compdev -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:30:12.161 11:40:55 compress_compdev -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b Nvme0n1 -t 2000 00:30:12.161 [ 00:30:12.161 { 00:30:12.161 "name": "Nvme0n1", 00:30:12.161 "aliases": [ 00:30:12.161 "01000000-0000-0000-5cd2-e43197705251" 00:30:12.161 ], 00:30:12.161 "product_name": "NVMe disk", 00:30:12.161 "block_size": 512, 00:30:12.161 "num_blocks": 15002931888, 00:30:12.161 "uuid": "01000000-0000-0000-5cd2-e43197705251", 00:30:12.161 "assigned_rate_limits": { 00:30:12.161 "rw_ios_per_sec": 0, 00:30:12.161 "rw_mbytes_per_sec": 0, 00:30:12.161 "r_mbytes_per_sec": 0, 00:30:12.161 "w_mbytes_per_sec": 0 00:30:12.161 }, 00:30:12.161 "claimed": false, 00:30:12.161 "zoned": false, 00:30:12.161 "supported_io_types": { 00:30:12.161 "read": true, 00:30:12.161 "write": true, 00:30:12.161 "unmap": true, 00:30:12.161 "flush": true, 00:30:12.161 "reset": true, 00:30:12.161 "nvme_admin": true, 00:30:12.161 "nvme_io": true, 00:30:12.161 "nvme_io_md": false, 00:30:12.161 "write_zeroes": true, 00:30:12.162 "zcopy": false, 00:30:12.162 "get_zone_info": false, 00:30:12.162 "zone_management": false, 00:30:12.162 "zone_append": false, 00:30:12.162 "compare": false, 00:30:12.162 "compare_and_write": false, 00:30:12.162 "abort": true, 00:30:12.162 "seek_hole": false, 00:30:12.162 "seek_data": false, 00:30:12.162 "copy": false, 00:30:12.162 "nvme_iov_md": false 00:30:12.162 }, 00:30:12.162 "driver_specific": { 00:30:12.162 "nvme": [ 00:30:12.162 { 00:30:12.162 "pci_address": "0000:5e:00.0", 00:30:12.162 "trid": { 00:30:12.162 "trtype": "PCIe", 00:30:12.162 "traddr": "0000:5e:00.0" 00:30:12.162 }, 00:30:12.162 "ctrlr_data": { 00:30:12.162 "cntlid": 0, 00:30:12.162 "vendor_id": "0x8086", 00:30:12.162 "model_number": "INTEL SSDPF2KX076TZO", 00:30:12.162 "serial_number": "PHAC0301002G7P6CGN", 00:30:12.162 "firmware_revision": "JCV10200", 00:30:12.162 "subnqn": "nqn.2020-07.com.intel:PHAC0301002G7P6CGN ", 00:30:12.162 "oacs": { 00:30:12.162 "security": 1, 00:30:12.162 "format": 1, 00:30:12.162 "firmware": 1, 00:30:12.162 "ns_manage": 1 00:30:12.162 }, 00:30:12.162 "multi_ctrlr": false, 00:30:12.162 "ana_reporting": false 00:30:12.162 }, 00:30:12.162 "vs": { 00:30:12.162 "nvme_version": "1.3" 00:30:12.162 }, 00:30:12.162 "ns_data": { 00:30:12.162 "id": 1, 00:30:12.162 "can_share": false 00:30:12.162 }, 00:30:12.162 "security": { 00:30:12.162 "opal": true 00:30:12.162 } 00:30:12.162 } 00:30:12.162 ], 00:30:12.162 "mp_policy": "active_passive" 00:30:12.162 } 00:30:12.162 } 00:30:12.162 ] 00:30:12.420 11:40:55 compress_compdev -- common/autotest_common.sh@905 -- # return 0 00:30:12.420 11:40:55 compress_compdev -- compress/compress.sh@37 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --clear-method none Nvme0n1 lvs0 00:30:12.420 [2024-07-15 11:40:55.996525] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0xd64660 PMD being used: compress_qat 00:30:14.949 4aa1d2c5-5550-4c68-88ea-5f413f6ddc05 00:30:14.949 11:40:58 compress_compdev -- compress/compress.sh@38 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -t -l lvs0 lv0 100 00:30:14.949 323ff736-a538-461e-8cfb-b20abc572c3e 00:30:14.949 11:40:58 compress_compdev -- compress/compress.sh@39 -- # waitforbdev lvs0/lv0 00:30:14.949 11:40:58 compress_compdev -- common/autotest_common.sh@897 -- # local bdev_name=lvs0/lv0 00:30:14.949 11:40:58 compress_compdev -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:30:14.949 11:40:58 compress_compdev -- common/autotest_common.sh@899 -- # local i 00:30:14.949 11:40:58 compress_compdev -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:30:14.949 11:40:58 compress_compdev -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:30:14.949 11:40:58 compress_compdev -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:30:15.207 11:40:58 compress_compdev -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b lvs0/lv0 -t 2000 00:30:15.465 [ 00:30:15.465 { 00:30:15.465 "name": "323ff736-a538-461e-8cfb-b20abc572c3e", 00:30:15.465 "aliases": [ 00:30:15.465 "lvs0/lv0" 00:30:15.465 ], 00:30:15.465 "product_name": "Logical Volume", 00:30:15.465 "block_size": 512, 00:30:15.465 "num_blocks": 204800, 00:30:15.465 "uuid": "323ff736-a538-461e-8cfb-b20abc572c3e", 00:30:15.465 "assigned_rate_limits": { 00:30:15.465 "rw_ios_per_sec": 0, 00:30:15.465 "rw_mbytes_per_sec": 0, 00:30:15.465 "r_mbytes_per_sec": 0, 00:30:15.465 "w_mbytes_per_sec": 0 00:30:15.465 }, 00:30:15.465 "claimed": false, 00:30:15.465 "zoned": false, 00:30:15.465 "supported_io_types": { 00:30:15.465 "read": true, 00:30:15.465 "write": true, 00:30:15.465 "unmap": true, 00:30:15.465 "flush": false, 00:30:15.465 "reset": true, 00:30:15.465 "nvme_admin": false, 00:30:15.465 "nvme_io": false, 00:30:15.465 "nvme_io_md": false, 00:30:15.465 "write_zeroes": true, 00:30:15.465 "zcopy": false, 00:30:15.465 "get_zone_info": false, 00:30:15.465 "zone_management": false, 00:30:15.465 "zone_append": false, 00:30:15.465 "compare": false, 00:30:15.465 "compare_and_write": false, 00:30:15.465 "abort": false, 00:30:15.465 "seek_hole": true, 00:30:15.465 "seek_data": true, 00:30:15.465 "copy": false, 00:30:15.465 "nvme_iov_md": false 00:30:15.465 }, 00:30:15.465 "driver_specific": { 00:30:15.465 "lvol": { 00:30:15.465 "lvol_store_uuid": "4aa1d2c5-5550-4c68-88ea-5f413f6ddc05", 00:30:15.465 "base_bdev": "Nvme0n1", 00:30:15.465 "thin_provision": true, 00:30:15.465 "num_allocated_clusters": 0, 00:30:15.465 "snapshot": false, 00:30:15.465 "clone": false, 00:30:15.465 "esnap_clone": false 00:30:15.465 } 00:30:15.465 } 00:30:15.465 } 00:30:15.465 ] 00:30:15.465 11:40:58 compress_compdev -- common/autotest_common.sh@905 -- # return 0 00:30:15.465 11:40:58 compress_compdev -- compress/compress.sh@41 -- # '[' -z 4096 ']' 00:30:15.465 11:40:58 compress_compdev -- compress/compress.sh@44 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_compress_create -b lvs0/lv0 -p /tmp/pmem -l 4096 00:30:15.724 [2024-07-15 11:40:59.191180] vbdev_compress.c:1016:vbdev_compress_claim: *NOTICE*: registered io_device and virtual bdev for: COMP_lvs0/lv0 00:30:15.724 COMP_lvs0/lv0 00:30:15.724 11:40:59 compress_compdev -- compress/compress.sh@46 -- # waitforbdev COMP_lvs0/lv0 00:30:15.724 11:40:59 compress_compdev -- common/autotest_common.sh@897 -- # local bdev_name=COMP_lvs0/lv0 00:30:15.724 11:40:59 compress_compdev -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:30:15.724 11:40:59 compress_compdev -- common/autotest_common.sh@899 -- # local i 00:30:15.724 11:40:59 compress_compdev -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:30:15.724 11:40:59 compress_compdev -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:30:15.724 11:40:59 compress_compdev -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:30:16.070 11:40:59 compress_compdev -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b COMP_lvs0/lv0 -t 2000 00:30:16.329 [ 00:30:16.329 { 00:30:16.329 "name": "COMP_lvs0/lv0", 00:30:16.329 "aliases": [ 00:30:16.329 "88ad5300-3014-56b4-9cfe-ae5ffdfab0e2" 00:30:16.329 ], 00:30:16.329 "product_name": "compress", 00:30:16.329 "block_size": 4096, 00:30:16.329 "num_blocks": 25088, 00:30:16.329 "uuid": "88ad5300-3014-56b4-9cfe-ae5ffdfab0e2", 00:30:16.329 "assigned_rate_limits": { 00:30:16.329 "rw_ios_per_sec": 0, 00:30:16.329 "rw_mbytes_per_sec": 0, 00:30:16.329 "r_mbytes_per_sec": 0, 00:30:16.329 "w_mbytes_per_sec": 0 00:30:16.329 }, 00:30:16.329 "claimed": false, 00:30:16.329 "zoned": false, 00:30:16.329 "supported_io_types": { 00:30:16.329 "read": true, 00:30:16.329 "write": true, 00:30:16.329 "unmap": false, 00:30:16.329 "flush": false, 00:30:16.329 "reset": false, 00:30:16.329 "nvme_admin": false, 00:30:16.329 "nvme_io": false, 00:30:16.329 "nvme_io_md": false, 00:30:16.329 "write_zeroes": true, 00:30:16.329 "zcopy": false, 00:30:16.329 "get_zone_info": false, 00:30:16.329 "zone_management": false, 00:30:16.329 "zone_append": false, 00:30:16.329 "compare": false, 00:30:16.329 "compare_and_write": false, 00:30:16.329 "abort": false, 00:30:16.329 "seek_hole": false, 00:30:16.329 "seek_data": false, 00:30:16.329 "copy": false, 00:30:16.329 "nvme_iov_md": false 00:30:16.329 }, 00:30:16.329 "driver_specific": { 00:30:16.329 "compress": { 00:30:16.329 "name": "COMP_lvs0/lv0", 00:30:16.329 "base_bdev_name": "323ff736-a538-461e-8cfb-b20abc572c3e" 00:30:16.329 } 00:30:16.329 } 00:30:16.329 } 00:30:16.329 ] 00:30:16.329 11:40:59 compress_compdev -- common/autotest_common.sh@905 -- # return 0 00:30:16.329 11:40:59 compress_compdev -- compress/compress.sh@75 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:30:16.329 [2024-07-15 11:40:59.761328] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x7f7ec81b15c0 PMD being used: compress_qat 00:30:16.329 [2024-07-15 11:40:59.763561] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0xefc770 PMD being used: compress_qat 00:30:16.329 Running I/O for 3 seconds... 00:30:19.615 00:30:19.615 Latency(us) 00:30:19.615 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:30:19.615 Job: COMP_lvs0/lv0 (Core Mask 0x2, workload: verify, depth: 32, IO size: 4096) 00:30:19.615 Verification LBA range: start 0x0 length 0x3100 00:30:19.615 COMP_lvs0/lv0 : 3.00 5122.52 20.01 0.00 0.00 6195.64 329.46 5841.25 00:30:19.615 Job: COMP_lvs0/lv0 (Core Mask 0x4, workload: verify, depth: 32, IO size: 4096) 00:30:19.615 Verification LBA range: start 0x3100 length 0x3100 00:30:19.615 COMP_lvs0/lv0 : 3.00 5367.90 20.97 0.00 0.00 5924.04 336.58 5641.79 00:30:19.615 =================================================================================================================== 00:30:19.615 Total : 10490.41 40.98 0.00 0.00 6056.66 329.46 5841.25 00:30:19.615 0 00:30:19.615 11:41:02 compress_compdev -- compress/compress.sh@76 -- # destroy_vols 00:30:19.615 11:41:02 compress_compdev -- compress/compress.sh@29 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_compress_delete COMP_lvs0/lv0 00:30:19.615 11:41:03 compress_compdev -- compress/compress.sh@30 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -l lvs0 00:30:19.873 11:41:03 compress_compdev -- compress/compress.sh@77 -- # trap - SIGINT SIGTERM EXIT 00:30:19.873 11:41:03 compress_compdev -- compress/compress.sh@78 -- # killprocess 1032877 00:30:19.873 11:41:03 compress_compdev -- common/autotest_common.sh@948 -- # '[' -z 1032877 ']' 00:30:19.873 11:41:03 compress_compdev -- common/autotest_common.sh@952 -- # kill -0 1032877 00:30:19.873 11:41:03 compress_compdev -- common/autotest_common.sh@953 -- # uname 00:30:19.873 11:41:03 compress_compdev -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:30:19.873 11:41:03 compress_compdev -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1032877 00:30:19.873 11:41:03 compress_compdev -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:30:19.873 11:41:03 compress_compdev -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:30:19.873 11:41:03 compress_compdev -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1032877' 00:30:19.873 killing process with pid 1032877 00:30:19.873 11:41:03 compress_compdev -- common/autotest_common.sh@967 -- # kill 1032877 00:30:19.873 Received shutdown signal, test time was about 3.000000 seconds 00:30:19.873 00:30:19.873 Latency(us) 00:30:19.873 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:30:19.873 =================================================================================================================== 00:30:19.873 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:30:19.873 11:41:03 compress_compdev -- common/autotest_common.sh@972 -- # wait 1032877 00:30:23.159 11:41:06 compress_compdev -- compress/compress.sh@89 -- # run_bdevio 00:30:23.159 11:41:06 compress_compdev -- compress/compress.sh@50 -- # [[ compdev == \c\o\m\p\d\e\v ]] 00:30:23.159 11:41:06 compress_compdev -- compress/compress.sh@55 -- # bdevio_pid=1034545 00:30:23.159 11:41:06 compress_compdev -- compress/compress.sh@56 -- # trap 'killprocess $bdevio_pid; error_cleanup; exit 1' SIGINT SIGTERM EXIT 00:30:23.159 11:41:06 compress_compdev -- compress/compress.sh@51 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevio/bdevio -c /var/jenkins/workspace/crypto-phy-autotest/spdk/test/compress/dpdk.json -w 00:30:23.159 11:41:06 compress_compdev -- compress/compress.sh@57 -- # waitforlisten 1034545 00:30:23.159 11:41:06 compress_compdev -- common/autotest_common.sh@829 -- # '[' -z 1034545 ']' 00:30:23.159 11:41:06 compress_compdev -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:23.159 11:41:06 compress_compdev -- common/autotest_common.sh@834 -- # local max_retries=100 00:30:23.159 11:41:06 compress_compdev -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:23.159 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:23.159 11:41:06 compress_compdev -- common/autotest_common.sh@838 -- # xtrace_disable 00:30:23.159 11:41:06 compress_compdev -- common/autotest_common.sh@10 -- # set +x 00:30:23.159 [2024-07-15 11:41:06.393200] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:30:23.159 [2024-07-15 11:41:06.393272] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1034545 ] 00:30:23.159 [2024-07-15 11:41:06.522906] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:30:23.159 [2024-07-15 11:41:06.631038] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:30:23.159 [2024-07-15 11:41:06.631122] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:30:23.159 [2024-07-15 11:41:06.631128] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:30:24.095 [2024-07-15 11:41:07.384340] accel_dpdk_compressdev.c: 296:accel_init_compress_drivers: *NOTICE*: initialized QAT PMD 00:30:24.095 11:41:07 compress_compdev -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:30:24.095 11:41:07 compress_compdev -- common/autotest_common.sh@862 -- # return 0 00:30:24.095 11:41:07 compress_compdev -- compress/compress.sh@58 -- # create_vols 00:30:24.095 11:41:07 compress_compdev -- compress/compress.sh@34 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/gen_nvme.sh 00:30:24.095 11:41:07 compress_compdev -- compress/compress.sh@34 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py load_subsystem_config 00:30:24.660 [2024-07-15 11:41:08.031324] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x24b0f20 PMD being used: compress_qat 00:30:24.660 11:41:08 compress_compdev -- compress/compress.sh@35 -- # waitforbdev Nvme0n1 00:30:24.660 11:41:08 compress_compdev -- common/autotest_common.sh@897 -- # local bdev_name=Nvme0n1 00:30:24.660 11:41:08 compress_compdev -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:30:24.660 11:41:08 compress_compdev -- common/autotest_common.sh@899 -- # local i 00:30:24.660 11:41:08 compress_compdev -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:30:24.660 11:41:08 compress_compdev -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:30:24.660 11:41:08 compress_compdev -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:30:24.917 11:41:08 compress_compdev -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b Nvme0n1 -t 2000 00:30:24.917 [ 00:30:24.917 { 00:30:24.917 "name": "Nvme0n1", 00:30:24.917 "aliases": [ 00:30:24.917 "01000000-0000-0000-5cd2-e43197705251" 00:30:24.917 ], 00:30:24.917 "product_name": "NVMe disk", 00:30:24.917 "block_size": 512, 00:30:24.917 "num_blocks": 15002931888, 00:30:24.917 "uuid": "01000000-0000-0000-5cd2-e43197705251", 00:30:24.917 "assigned_rate_limits": { 00:30:24.917 "rw_ios_per_sec": 0, 00:30:24.917 "rw_mbytes_per_sec": 0, 00:30:24.917 "r_mbytes_per_sec": 0, 00:30:24.917 "w_mbytes_per_sec": 0 00:30:24.917 }, 00:30:24.917 "claimed": false, 00:30:24.917 "zoned": false, 00:30:24.917 "supported_io_types": { 00:30:24.917 "read": true, 00:30:24.917 "write": true, 00:30:24.917 "unmap": true, 00:30:24.917 "flush": true, 00:30:24.917 "reset": true, 00:30:24.917 "nvme_admin": true, 00:30:24.917 "nvme_io": true, 00:30:24.917 "nvme_io_md": false, 00:30:24.917 "write_zeroes": true, 00:30:24.917 "zcopy": false, 00:30:24.917 "get_zone_info": false, 00:30:24.917 "zone_management": false, 00:30:24.917 "zone_append": false, 00:30:24.917 "compare": false, 00:30:24.917 "compare_and_write": false, 00:30:24.917 "abort": true, 00:30:24.917 "seek_hole": false, 00:30:24.917 "seek_data": false, 00:30:24.917 "copy": false, 00:30:24.917 "nvme_iov_md": false 00:30:24.917 }, 00:30:24.917 "driver_specific": { 00:30:24.917 "nvme": [ 00:30:24.917 { 00:30:24.917 "pci_address": "0000:5e:00.0", 00:30:24.917 "trid": { 00:30:24.917 "trtype": "PCIe", 00:30:24.917 "traddr": "0000:5e:00.0" 00:30:24.917 }, 00:30:24.917 "ctrlr_data": { 00:30:24.917 "cntlid": 0, 00:30:24.917 "vendor_id": "0x8086", 00:30:24.917 "model_number": "INTEL SSDPF2KX076TZO", 00:30:24.917 "serial_number": "PHAC0301002G7P6CGN", 00:30:24.917 "firmware_revision": "JCV10200", 00:30:24.917 "subnqn": "nqn.2020-07.com.intel:PHAC0301002G7P6CGN ", 00:30:24.917 "oacs": { 00:30:24.917 "security": 1, 00:30:24.917 "format": 1, 00:30:24.917 "firmware": 1, 00:30:24.917 "ns_manage": 1 00:30:24.917 }, 00:30:24.917 "multi_ctrlr": false, 00:30:24.917 "ana_reporting": false 00:30:24.917 }, 00:30:24.917 "vs": { 00:30:24.917 "nvme_version": "1.3" 00:30:24.917 }, 00:30:24.917 "ns_data": { 00:30:24.917 "id": 1, 00:30:24.917 "can_share": false 00:30:24.917 }, 00:30:24.917 "security": { 00:30:24.917 "opal": true 00:30:24.917 } 00:30:24.917 } 00:30:24.917 ], 00:30:24.917 "mp_policy": "active_passive" 00:30:24.917 } 00:30:24.917 } 00:30:24.917 ] 00:30:24.917 11:41:08 compress_compdev -- common/autotest_common.sh@905 -- # return 0 00:30:24.917 11:41:08 compress_compdev -- compress/compress.sh@37 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --clear-method none Nvme0n1 lvs0 00:30:25.175 [2024-07-15 11:41:08.736533] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x22ff440 PMD being used: compress_qat 00:30:27.705 f7c0ea6a-b640-4d00-9511-0f35b534791c 00:30:27.705 11:41:10 compress_compdev -- compress/compress.sh@38 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -t -l lvs0 lv0 100 00:30:27.705 53fec2fb-be0c-4358-bba3-02ce3f8a02ee 00:30:27.705 11:41:11 compress_compdev -- compress/compress.sh@39 -- # waitforbdev lvs0/lv0 00:30:27.705 11:41:11 compress_compdev -- common/autotest_common.sh@897 -- # local bdev_name=lvs0/lv0 00:30:27.705 11:41:11 compress_compdev -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:30:27.705 11:41:11 compress_compdev -- common/autotest_common.sh@899 -- # local i 00:30:27.705 11:41:11 compress_compdev -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:30:27.705 11:41:11 compress_compdev -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:30:27.705 11:41:11 compress_compdev -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:30:27.962 11:41:11 compress_compdev -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b lvs0/lv0 -t 2000 00:30:28.219 [ 00:30:28.219 { 00:30:28.219 "name": "53fec2fb-be0c-4358-bba3-02ce3f8a02ee", 00:30:28.219 "aliases": [ 00:30:28.219 "lvs0/lv0" 00:30:28.219 ], 00:30:28.219 "product_name": "Logical Volume", 00:30:28.219 "block_size": 512, 00:30:28.219 "num_blocks": 204800, 00:30:28.219 "uuid": "53fec2fb-be0c-4358-bba3-02ce3f8a02ee", 00:30:28.219 "assigned_rate_limits": { 00:30:28.219 "rw_ios_per_sec": 0, 00:30:28.219 "rw_mbytes_per_sec": 0, 00:30:28.219 "r_mbytes_per_sec": 0, 00:30:28.219 "w_mbytes_per_sec": 0 00:30:28.219 }, 00:30:28.219 "claimed": false, 00:30:28.219 "zoned": false, 00:30:28.219 "supported_io_types": { 00:30:28.219 "read": true, 00:30:28.219 "write": true, 00:30:28.219 "unmap": true, 00:30:28.219 "flush": false, 00:30:28.219 "reset": true, 00:30:28.219 "nvme_admin": false, 00:30:28.219 "nvme_io": false, 00:30:28.219 "nvme_io_md": false, 00:30:28.219 "write_zeroes": true, 00:30:28.219 "zcopy": false, 00:30:28.219 "get_zone_info": false, 00:30:28.219 "zone_management": false, 00:30:28.219 "zone_append": false, 00:30:28.219 "compare": false, 00:30:28.219 "compare_and_write": false, 00:30:28.219 "abort": false, 00:30:28.219 "seek_hole": true, 00:30:28.219 "seek_data": true, 00:30:28.219 "copy": false, 00:30:28.219 "nvme_iov_md": false 00:30:28.219 }, 00:30:28.219 "driver_specific": { 00:30:28.219 "lvol": { 00:30:28.219 "lvol_store_uuid": "f7c0ea6a-b640-4d00-9511-0f35b534791c", 00:30:28.219 "base_bdev": "Nvme0n1", 00:30:28.219 "thin_provision": true, 00:30:28.219 "num_allocated_clusters": 0, 00:30:28.219 "snapshot": false, 00:30:28.219 "clone": false, 00:30:28.219 "esnap_clone": false 00:30:28.219 } 00:30:28.219 } 00:30:28.219 } 00:30:28.219 ] 00:30:28.219 11:41:11 compress_compdev -- common/autotest_common.sh@905 -- # return 0 00:30:28.219 11:41:11 compress_compdev -- compress/compress.sh@41 -- # '[' -z '' ']' 00:30:28.219 11:41:11 compress_compdev -- compress/compress.sh@42 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_compress_create -b lvs0/lv0 -p /tmp/pmem 00:30:28.477 [2024-07-15 11:41:11.968445] vbdev_compress.c:1016:vbdev_compress_claim: *NOTICE*: registered io_device and virtual bdev for: COMP_lvs0/lv0 00:30:28.477 COMP_lvs0/lv0 00:30:28.477 11:41:11 compress_compdev -- compress/compress.sh@46 -- # waitforbdev COMP_lvs0/lv0 00:30:28.477 11:41:11 compress_compdev -- common/autotest_common.sh@897 -- # local bdev_name=COMP_lvs0/lv0 00:30:28.477 11:41:11 compress_compdev -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:30:28.477 11:41:11 compress_compdev -- common/autotest_common.sh@899 -- # local i 00:30:28.477 11:41:11 compress_compdev -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:30:28.477 11:41:11 compress_compdev -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:30:28.477 11:41:11 compress_compdev -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:30:28.734 11:41:12 compress_compdev -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b COMP_lvs0/lv0 -t 2000 00:30:28.992 [ 00:30:28.992 { 00:30:28.992 "name": "COMP_lvs0/lv0", 00:30:28.992 "aliases": [ 00:30:28.992 "0d7f475d-c2cc-54b5-9d28-6891b7969b79" 00:30:28.992 ], 00:30:28.992 "product_name": "compress", 00:30:28.992 "block_size": 512, 00:30:28.992 "num_blocks": 200704, 00:30:28.992 "uuid": "0d7f475d-c2cc-54b5-9d28-6891b7969b79", 00:30:28.992 "assigned_rate_limits": { 00:30:28.992 "rw_ios_per_sec": 0, 00:30:28.992 "rw_mbytes_per_sec": 0, 00:30:28.992 "r_mbytes_per_sec": 0, 00:30:28.992 "w_mbytes_per_sec": 0 00:30:28.992 }, 00:30:28.992 "claimed": false, 00:30:28.992 "zoned": false, 00:30:28.992 "supported_io_types": { 00:30:28.992 "read": true, 00:30:28.992 "write": true, 00:30:28.992 "unmap": false, 00:30:28.992 "flush": false, 00:30:28.992 "reset": false, 00:30:28.992 "nvme_admin": false, 00:30:28.992 "nvme_io": false, 00:30:28.992 "nvme_io_md": false, 00:30:28.992 "write_zeroes": true, 00:30:28.992 "zcopy": false, 00:30:28.992 "get_zone_info": false, 00:30:28.992 "zone_management": false, 00:30:28.992 "zone_append": false, 00:30:28.992 "compare": false, 00:30:28.992 "compare_and_write": false, 00:30:28.992 "abort": false, 00:30:28.992 "seek_hole": false, 00:30:28.992 "seek_data": false, 00:30:28.992 "copy": false, 00:30:28.992 "nvme_iov_md": false 00:30:28.992 }, 00:30:28.992 "driver_specific": { 00:30:28.992 "compress": { 00:30:28.992 "name": "COMP_lvs0/lv0", 00:30:28.992 "base_bdev_name": "53fec2fb-be0c-4358-bba3-02ce3f8a02ee" 00:30:28.992 } 00:30:28.992 } 00:30:28.992 } 00:30:28.992 ] 00:30:28.992 11:41:12 compress_compdev -- common/autotest_common.sh@905 -- # return 0 00:30:28.992 11:41:12 compress_compdev -- compress/compress.sh@59 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevio/tests.py perform_tests 00:30:29.250 [2024-07-15 11:41:12.593395] accel_dpdk_compressdev.c: 690:_set_pmd: *NOTICE*: Channel 0x7f46a81b1350 PMD being used: compress_qat 00:30:29.250 I/O targets: 00:30:29.250 COMP_lvs0/lv0: 200704 blocks of 512 bytes (98 MiB) 00:30:29.250 00:30:29.250 00:30:29.250 CUnit - A unit testing framework for C - Version 2.1-3 00:30:29.250 http://cunit.sourceforge.net/ 00:30:29.250 00:30:29.250 00:30:29.250 Suite: bdevio tests on: COMP_lvs0/lv0 00:30:29.250 Test: blockdev write read block ...passed 00:30:29.250 Test: blockdev write zeroes read block ...passed 00:30:29.250 Test: blockdev write zeroes read no split ...passed 00:30:29.250 Test: blockdev write zeroes read split ...passed 00:30:29.250 Test: blockdev write zeroes read split partial ...passed 00:30:29.250 Test: blockdev reset ...[2024-07-15 11:41:12.630743] vbdev_compress.c: 252:vbdev_compress_submit_request: *ERROR*: Unknown I/O type 5 00:30:29.250 passed 00:30:29.250 Test: blockdev write read 8 blocks ...passed 00:30:29.250 Test: blockdev write read size > 128k ...passed 00:30:29.250 Test: blockdev write read invalid size ...passed 00:30:29.250 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:30:29.250 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:30:29.250 Test: blockdev write read max offset ...passed 00:30:29.250 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:30:29.250 Test: blockdev writev readv 8 blocks ...passed 00:30:29.250 Test: blockdev writev readv 30 x 1block ...passed 00:30:29.250 Test: blockdev writev readv block ...passed 00:30:29.250 Test: blockdev writev readv size > 128k ...passed 00:30:29.250 Test: blockdev writev readv size > 128k in two iovs ...passed 00:30:29.250 Test: blockdev comparev and writev ...passed 00:30:29.250 Test: blockdev nvme passthru rw ...passed 00:30:29.250 Test: blockdev nvme passthru vendor specific ...passed 00:30:29.250 Test: blockdev nvme admin passthru ...passed 00:30:29.250 Test: blockdev copy ...passed 00:30:29.250 00:30:29.250 Run Summary: Type Total Ran Passed Failed Inactive 00:30:29.250 suites 1 1 n/a 0 0 00:30:29.250 tests 23 23 23 0 0 00:30:29.250 asserts 130 130 130 0 n/a 00:30:29.250 00:30:29.250 Elapsed time = 0.092 seconds 00:30:29.250 0 00:30:29.250 11:41:12 compress_compdev -- compress/compress.sh@60 -- # destroy_vols 00:30:29.250 11:41:12 compress_compdev -- compress/compress.sh@29 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_compress_delete COMP_lvs0/lv0 00:30:29.507 11:41:12 compress_compdev -- compress/compress.sh@30 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -l lvs0 00:30:29.768 11:41:13 compress_compdev -- compress/compress.sh@61 -- # trap - SIGINT SIGTERM EXIT 00:30:29.768 11:41:13 compress_compdev -- compress/compress.sh@62 -- # killprocess 1034545 00:30:29.768 11:41:13 compress_compdev -- common/autotest_common.sh@948 -- # '[' -z 1034545 ']' 00:30:29.768 11:41:13 compress_compdev -- common/autotest_common.sh@952 -- # kill -0 1034545 00:30:29.768 11:41:13 compress_compdev -- common/autotest_common.sh@953 -- # uname 00:30:29.768 11:41:13 compress_compdev -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:30:29.768 11:41:13 compress_compdev -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1034545 00:30:29.768 11:41:13 compress_compdev -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:30:29.768 11:41:13 compress_compdev -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:30:29.768 11:41:13 compress_compdev -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1034545' 00:30:29.768 killing process with pid 1034545 00:30:29.768 11:41:13 compress_compdev -- common/autotest_common.sh@967 -- # kill 1034545 00:30:29.768 11:41:13 compress_compdev -- common/autotest_common.sh@972 -- # wait 1034545 00:30:33.049 11:41:16 compress_compdev -- compress/compress.sh@91 -- # '[' 0 -eq 1 ']' 00:30:33.049 11:41:16 compress_compdev -- compress/compress.sh@120 -- # rm -rf /tmp/pmem 00:30:33.049 00:30:33.049 real 0m48.279s 00:30:33.049 user 1m51.563s 00:30:33.049 sys 0m5.746s 00:30:33.049 11:41:16 compress_compdev -- common/autotest_common.sh@1124 -- # xtrace_disable 00:30:33.049 11:41:16 compress_compdev -- common/autotest_common.sh@10 -- # set +x 00:30:33.049 ************************************ 00:30:33.049 END TEST compress_compdev 00:30:33.049 ************************************ 00:30:33.049 11:41:16 -- common/autotest_common.sh@1142 -- # return 0 00:30:33.049 11:41:16 -- spdk/autotest.sh@349 -- # run_test compress_isal /var/jenkins/workspace/crypto-phy-autotest/spdk/test/compress/compress.sh isal 00:30:33.049 11:41:16 -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:30:33.049 11:41:16 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:30:33.049 11:41:16 -- common/autotest_common.sh@10 -- # set +x 00:30:33.049 ************************************ 00:30:33.049 START TEST compress_isal 00:30:33.049 ************************************ 00:30:33.049 11:41:16 compress_isal -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/compress/compress.sh isal 00:30:33.049 * Looking for test storage... 00:30:33.049 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/compress 00:30:33.049 11:41:16 compress_isal -- compress/compress.sh@13 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/nvmf/common.sh 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@7 -- # uname -s 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:809e2efd-7f71-e711-906e-0017a4403562 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@18 -- # NVME_HOSTID=809e2efd-7f71-e711-906e-0017a4403562 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/common.sh 00:30:33.049 11:41:16 compress_isal -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:30:33.049 11:41:16 compress_isal -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:30:33.049 11:41:16 compress_isal -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:30:33.049 11:41:16 compress_isal -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:33.049 11:41:16 compress_isal -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:33.049 11:41:16 compress_isal -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:33.049 11:41:16 compress_isal -- paths/export.sh@5 -- # export PATH 00:30:33.049 11:41:16 compress_isal -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@47 -- # : 0 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:30:33.049 11:41:16 compress_isal -- nvmf/common.sh@51 -- # have_pci_nics=0 00:30:33.049 11:41:16 compress_isal -- compress/compress.sh@17 -- # rpc_py=/var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py 00:30:33.049 11:41:16 compress_isal -- compress/compress.sh@81 -- # mkdir -p /tmp/pmem 00:30:33.049 11:41:16 compress_isal -- compress/compress.sh@82 -- # test_type=isal 00:30:33.049 11:41:16 compress_isal -- compress/compress.sh@86 -- # run_bdevperf 32 4096 3 00:30:33.049 11:41:16 compress_isal -- compress/compress.sh@66 -- # [[ isal == \c\o\m\p\d\e\v ]] 00:30:33.049 11:41:16 compress_isal -- compress/compress.sh@71 -- # bdevperf_pid=1035952 00:30:33.049 11:41:16 compress_isal -- compress/compress.sh@72 -- # trap 'killprocess $bdevperf_pid; error_cleanup; exit 1' SIGINT SIGTERM EXIT 00:30:33.049 11:41:16 compress_isal -- compress/compress.sh@73 -- # waitforlisten 1035952 00:30:33.049 11:41:16 compress_isal -- common/autotest_common.sh@829 -- # '[' -z 1035952 ']' 00:30:33.049 11:41:16 compress_isal -- compress/compress.sh@69 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -z -q 32 -o 4096 -w verify -t 3 -C -m 0x6 00:30:33.049 11:41:16 compress_isal -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:33.049 11:41:16 compress_isal -- common/autotest_common.sh@834 -- # local max_retries=100 00:30:33.049 11:41:16 compress_isal -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:33.049 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:33.049 11:41:16 compress_isal -- common/autotest_common.sh@838 -- # xtrace_disable 00:30:33.049 11:41:16 compress_isal -- common/autotest_common.sh@10 -- # set +x 00:30:33.049 [2024-07-15 11:41:16.465399] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:30:33.049 [2024-07-15 11:41:16.465468] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x6 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1035952 ] 00:30:33.049 [2024-07-15 11:41:16.583847] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:30:33.309 [2024-07-15 11:41:16.693228] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:30:33.309 [2024-07-15 11:41:16.693234] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:30:33.873 11:41:17 compress_isal -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:30:33.873 11:41:17 compress_isal -- common/autotest_common.sh@862 -- # return 0 00:30:33.873 11:41:17 compress_isal -- compress/compress.sh@74 -- # create_vols 00:30:33.873 11:41:17 compress_isal -- compress/compress.sh@34 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/gen_nvme.sh 00:30:33.873 11:41:17 compress_isal -- compress/compress.sh@34 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py load_subsystem_config 00:30:34.439 11:41:17 compress_isal -- compress/compress.sh@35 -- # waitforbdev Nvme0n1 00:30:34.439 11:41:17 compress_isal -- common/autotest_common.sh@897 -- # local bdev_name=Nvme0n1 00:30:34.439 11:41:17 compress_isal -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:30:34.439 11:41:17 compress_isal -- common/autotest_common.sh@899 -- # local i 00:30:34.439 11:41:17 compress_isal -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:30:34.439 11:41:17 compress_isal -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:30:34.439 11:41:17 compress_isal -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:30:34.698 11:41:18 compress_isal -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b Nvme0n1 -t 2000 00:30:34.956 [ 00:30:34.956 { 00:30:34.956 "name": "Nvme0n1", 00:30:34.956 "aliases": [ 00:30:34.956 "01000000-0000-0000-5cd2-e43197705251" 00:30:34.956 ], 00:30:34.956 "product_name": "NVMe disk", 00:30:34.956 "block_size": 512, 00:30:34.956 "num_blocks": 15002931888, 00:30:34.956 "uuid": "01000000-0000-0000-5cd2-e43197705251", 00:30:34.956 "assigned_rate_limits": { 00:30:34.956 "rw_ios_per_sec": 0, 00:30:34.956 "rw_mbytes_per_sec": 0, 00:30:34.956 "r_mbytes_per_sec": 0, 00:30:34.956 "w_mbytes_per_sec": 0 00:30:34.956 }, 00:30:34.956 "claimed": false, 00:30:34.956 "zoned": false, 00:30:34.956 "supported_io_types": { 00:30:34.956 "read": true, 00:30:34.956 "write": true, 00:30:34.956 "unmap": true, 00:30:34.956 "flush": true, 00:30:34.956 "reset": true, 00:30:34.956 "nvme_admin": true, 00:30:34.956 "nvme_io": true, 00:30:34.956 "nvme_io_md": false, 00:30:34.956 "write_zeroes": true, 00:30:34.956 "zcopy": false, 00:30:34.956 "get_zone_info": false, 00:30:34.956 "zone_management": false, 00:30:34.956 "zone_append": false, 00:30:34.956 "compare": false, 00:30:34.956 "compare_and_write": false, 00:30:34.956 "abort": true, 00:30:34.956 "seek_hole": false, 00:30:34.956 "seek_data": false, 00:30:34.956 "copy": false, 00:30:34.956 "nvme_iov_md": false 00:30:34.956 }, 00:30:34.956 "driver_specific": { 00:30:34.956 "nvme": [ 00:30:34.956 { 00:30:34.956 "pci_address": "0000:5e:00.0", 00:30:34.956 "trid": { 00:30:34.956 "trtype": "PCIe", 00:30:34.956 "traddr": "0000:5e:00.0" 00:30:34.956 }, 00:30:34.956 "ctrlr_data": { 00:30:34.956 "cntlid": 0, 00:30:34.956 "vendor_id": "0x8086", 00:30:34.956 "model_number": "INTEL SSDPF2KX076TZO", 00:30:34.956 "serial_number": "PHAC0301002G7P6CGN", 00:30:34.956 "firmware_revision": "JCV10200", 00:30:34.956 "subnqn": "nqn.2020-07.com.intel:PHAC0301002G7P6CGN ", 00:30:34.956 "oacs": { 00:30:34.956 "security": 1, 00:30:34.956 "format": 1, 00:30:34.956 "firmware": 1, 00:30:34.956 "ns_manage": 1 00:30:34.956 }, 00:30:34.956 "multi_ctrlr": false, 00:30:34.956 "ana_reporting": false 00:30:34.956 }, 00:30:34.956 "vs": { 00:30:34.956 "nvme_version": "1.3" 00:30:34.956 }, 00:30:34.956 "ns_data": { 00:30:34.956 "id": 1, 00:30:34.956 "can_share": false 00:30:34.956 }, 00:30:34.956 "security": { 00:30:34.956 "opal": true 00:30:34.956 } 00:30:34.956 } 00:30:34.956 ], 00:30:34.956 "mp_policy": "active_passive" 00:30:34.956 } 00:30:34.956 } 00:30:34.956 ] 00:30:34.956 11:41:18 compress_isal -- common/autotest_common.sh@905 -- # return 0 00:30:34.956 11:41:18 compress_isal -- compress/compress.sh@37 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --clear-method none Nvme0n1 lvs0 00:30:37.484 5a5a5b53-e506-4fc5-aefb-68800b133261 00:30:37.484 11:41:20 compress_isal -- compress/compress.sh@38 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -t -l lvs0 lv0 100 00:30:38.048 1116588c-205f-4d3e-9845-1f7bde4af53d 00:30:38.048 11:41:21 compress_isal -- compress/compress.sh@39 -- # waitforbdev lvs0/lv0 00:30:38.048 11:41:21 compress_isal -- common/autotest_common.sh@897 -- # local bdev_name=lvs0/lv0 00:30:38.048 11:41:21 compress_isal -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:30:38.048 11:41:21 compress_isal -- common/autotest_common.sh@899 -- # local i 00:30:38.048 11:41:21 compress_isal -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:30:38.048 11:41:21 compress_isal -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:30:38.048 11:41:21 compress_isal -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:30:38.306 11:41:21 compress_isal -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b lvs0/lv0 -t 2000 00:30:38.306 [ 00:30:38.306 { 00:30:38.306 "name": "1116588c-205f-4d3e-9845-1f7bde4af53d", 00:30:38.306 "aliases": [ 00:30:38.306 "lvs0/lv0" 00:30:38.306 ], 00:30:38.306 "product_name": "Logical Volume", 00:30:38.306 "block_size": 512, 00:30:38.306 "num_blocks": 204800, 00:30:38.306 "uuid": "1116588c-205f-4d3e-9845-1f7bde4af53d", 00:30:38.306 "assigned_rate_limits": { 00:30:38.306 "rw_ios_per_sec": 0, 00:30:38.306 "rw_mbytes_per_sec": 0, 00:30:38.306 "r_mbytes_per_sec": 0, 00:30:38.306 "w_mbytes_per_sec": 0 00:30:38.306 }, 00:30:38.306 "claimed": false, 00:30:38.306 "zoned": false, 00:30:38.306 "supported_io_types": { 00:30:38.306 "read": true, 00:30:38.306 "write": true, 00:30:38.306 "unmap": true, 00:30:38.306 "flush": false, 00:30:38.306 "reset": true, 00:30:38.306 "nvme_admin": false, 00:30:38.306 "nvme_io": false, 00:30:38.306 "nvme_io_md": false, 00:30:38.306 "write_zeroes": true, 00:30:38.306 "zcopy": false, 00:30:38.306 "get_zone_info": false, 00:30:38.306 "zone_management": false, 00:30:38.306 "zone_append": false, 00:30:38.306 "compare": false, 00:30:38.306 "compare_and_write": false, 00:30:38.306 "abort": false, 00:30:38.306 "seek_hole": true, 00:30:38.306 "seek_data": true, 00:30:38.306 "copy": false, 00:30:38.306 "nvme_iov_md": false 00:30:38.306 }, 00:30:38.306 "driver_specific": { 00:30:38.306 "lvol": { 00:30:38.306 "lvol_store_uuid": "5a5a5b53-e506-4fc5-aefb-68800b133261", 00:30:38.306 "base_bdev": "Nvme0n1", 00:30:38.306 "thin_provision": true, 00:30:38.306 "num_allocated_clusters": 0, 00:30:38.306 "snapshot": false, 00:30:38.306 "clone": false, 00:30:38.306 "esnap_clone": false 00:30:38.306 } 00:30:38.306 } 00:30:38.306 } 00:30:38.306 ] 00:30:38.565 11:41:21 compress_isal -- common/autotest_common.sh@905 -- # return 0 00:30:38.565 11:41:21 compress_isal -- compress/compress.sh@41 -- # '[' -z '' ']' 00:30:38.565 11:41:21 compress_isal -- compress/compress.sh@42 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_compress_create -b lvs0/lv0 -p /tmp/pmem 00:30:38.823 [2024-07-15 11:41:22.397615] vbdev_compress.c:1016:vbdev_compress_claim: *NOTICE*: registered io_device and virtual bdev for: COMP_lvs0/lv0 00:30:38.823 COMP_lvs0/lv0 00:30:38.823 11:41:22 compress_isal -- compress/compress.sh@46 -- # waitforbdev COMP_lvs0/lv0 00:30:38.823 11:41:22 compress_isal -- common/autotest_common.sh@897 -- # local bdev_name=COMP_lvs0/lv0 00:30:38.823 11:41:22 compress_isal -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:30:38.823 11:41:22 compress_isal -- common/autotest_common.sh@899 -- # local i 00:30:38.823 11:41:22 compress_isal -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:30:38.823 11:41:22 compress_isal -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:30:38.823 11:41:22 compress_isal -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:30:39.081 11:41:22 compress_isal -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b COMP_lvs0/lv0 -t 2000 00:30:39.649 [ 00:30:39.649 { 00:30:39.649 "name": "COMP_lvs0/lv0", 00:30:39.649 "aliases": [ 00:30:39.649 "548d122f-3017-50e3-8cec-e0fe4fdca1db" 00:30:39.649 ], 00:30:39.649 "product_name": "compress", 00:30:39.649 "block_size": 512, 00:30:39.649 "num_blocks": 200704, 00:30:39.649 "uuid": "548d122f-3017-50e3-8cec-e0fe4fdca1db", 00:30:39.649 "assigned_rate_limits": { 00:30:39.649 "rw_ios_per_sec": 0, 00:30:39.649 "rw_mbytes_per_sec": 0, 00:30:39.649 "r_mbytes_per_sec": 0, 00:30:39.649 "w_mbytes_per_sec": 0 00:30:39.649 }, 00:30:39.649 "claimed": false, 00:30:39.649 "zoned": false, 00:30:39.649 "supported_io_types": { 00:30:39.649 "read": true, 00:30:39.649 "write": true, 00:30:39.649 "unmap": false, 00:30:39.649 "flush": false, 00:30:39.649 "reset": false, 00:30:39.649 "nvme_admin": false, 00:30:39.649 "nvme_io": false, 00:30:39.649 "nvme_io_md": false, 00:30:39.649 "write_zeroes": true, 00:30:39.649 "zcopy": false, 00:30:39.649 "get_zone_info": false, 00:30:39.649 "zone_management": false, 00:30:39.649 "zone_append": false, 00:30:39.649 "compare": false, 00:30:39.649 "compare_and_write": false, 00:30:39.649 "abort": false, 00:30:39.649 "seek_hole": false, 00:30:39.649 "seek_data": false, 00:30:39.649 "copy": false, 00:30:39.649 "nvme_iov_md": false 00:30:39.649 }, 00:30:39.649 "driver_specific": { 00:30:39.649 "compress": { 00:30:39.649 "name": "COMP_lvs0/lv0", 00:30:39.649 "base_bdev_name": "1116588c-205f-4d3e-9845-1f7bde4af53d" 00:30:39.649 } 00:30:39.649 } 00:30:39.649 } 00:30:39.649 ] 00:30:39.649 11:41:23 compress_isal -- common/autotest_common.sh@905 -- # return 0 00:30:39.649 11:41:23 compress_isal -- compress/compress.sh@75 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:30:39.907 Running I/O for 3 seconds... 00:30:43.190 00:30:43.190 Latency(us) 00:30:43.190 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:30:43.190 Job: COMP_lvs0/lv0 (Core Mask 0x2, workload: verify, depth: 32, IO size: 4096) 00:30:43.190 Verification LBA range: start 0x0 length 0x3100 00:30:43.190 COMP_lvs0/lv0 : 3.01 2913.10 11.38 0.00 0.00 10939.90 655.36 9687.93 00:30:43.190 Job: COMP_lvs0/lv0 (Core Mask 0x4, workload: verify, depth: 32, IO size: 4096) 00:30:43.190 Verification LBA range: start 0x3100 length 0x3100 00:30:43.190 COMP_lvs0/lv0 : 3.01 2909.83 11.37 0.00 0.00 10959.98 1089.89 9744.92 00:30:43.190 =================================================================================================================== 00:30:43.190 Total : 5822.93 22.75 0.00 0.00 10949.94 655.36 9744.92 00:30:43.190 0 00:30:43.190 11:41:26 compress_isal -- compress/compress.sh@76 -- # destroy_vols 00:30:43.190 11:41:26 compress_isal -- compress/compress.sh@29 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_compress_delete COMP_lvs0/lv0 00:30:43.190 11:41:26 compress_isal -- compress/compress.sh@30 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -l lvs0 00:30:43.448 11:41:26 compress_isal -- compress/compress.sh@77 -- # trap - SIGINT SIGTERM EXIT 00:30:43.448 11:41:26 compress_isal -- compress/compress.sh@78 -- # killprocess 1035952 00:30:43.448 11:41:26 compress_isal -- common/autotest_common.sh@948 -- # '[' -z 1035952 ']' 00:30:43.448 11:41:26 compress_isal -- common/autotest_common.sh@952 -- # kill -0 1035952 00:30:43.448 11:41:26 compress_isal -- common/autotest_common.sh@953 -- # uname 00:30:43.448 11:41:26 compress_isal -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:30:43.448 11:41:26 compress_isal -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1035952 00:30:43.448 11:41:26 compress_isal -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:30:43.448 11:41:26 compress_isal -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:30:43.448 11:41:26 compress_isal -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1035952' 00:30:43.448 killing process with pid 1035952 00:30:43.448 11:41:26 compress_isal -- common/autotest_common.sh@967 -- # kill 1035952 00:30:43.448 Received shutdown signal, test time was about 3.000000 seconds 00:30:43.448 00:30:43.448 Latency(us) 00:30:43.448 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:30:43.448 =================================================================================================================== 00:30:43.448 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:30:43.448 11:41:26 compress_isal -- common/autotest_common.sh@972 -- # wait 1035952 00:30:46.724 11:41:29 compress_isal -- compress/compress.sh@87 -- # run_bdevperf 32 4096 3 512 00:30:46.724 11:41:29 compress_isal -- compress/compress.sh@66 -- # [[ isal == \c\o\m\p\d\e\v ]] 00:30:46.724 11:41:29 compress_isal -- compress/compress.sh@71 -- # bdevperf_pid=1037624 00:30:46.724 11:41:29 compress_isal -- compress/compress.sh@69 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -z -q 32 -o 4096 -w verify -t 3 -C -m 0x6 00:30:46.724 11:41:29 compress_isal -- compress/compress.sh@72 -- # trap 'killprocess $bdevperf_pid; error_cleanup; exit 1' SIGINT SIGTERM EXIT 00:30:46.724 11:41:29 compress_isal -- compress/compress.sh@73 -- # waitforlisten 1037624 00:30:46.724 11:41:29 compress_isal -- common/autotest_common.sh@829 -- # '[' -z 1037624 ']' 00:30:46.724 11:41:29 compress_isal -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:46.724 11:41:29 compress_isal -- common/autotest_common.sh@834 -- # local max_retries=100 00:30:46.724 11:41:29 compress_isal -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:46.724 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:46.724 11:41:29 compress_isal -- common/autotest_common.sh@838 -- # xtrace_disable 00:30:46.724 11:41:29 compress_isal -- common/autotest_common.sh@10 -- # set +x 00:30:46.724 [2024-07-15 11:41:29.902100] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:30:46.724 [2024-07-15 11:41:29.902177] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x6 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1037624 ] 00:30:46.724 [2024-07-15 11:41:30.023378] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:30:46.724 [2024-07-15 11:41:30.132979] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:30:46.724 [2024-07-15 11:41:30.132985] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:30:47.291 11:41:30 compress_isal -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:30:47.291 11:41:30 compress_isal -- common/autotest_common.sh@862 -- # return 0 00:30:47.291 11:41:30 compress_isal -- compress/compress.sh@74 -- # create_vols 512 00:30:47.291 11:41:30 compress_isal -- compress/compress.sh@34 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py load_subsystem_config 00:30:47.291 11:41:30 compress_isal -- compress/compress.sh@34 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/gen_nvme.sh 00:30:47.857 11:41:31 compress_isal -- compress/compress.sh@35 -- # waitforbdev Nvme0n1 00:30:47.857 11:41:31 compress_isal -- common/autotest_common.sh@897 -- # local bdev_name=Nvme0n1 00:30:47.857 11:41:31 compress_isal -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:30:47.857 11:41:31 compress_isal -- common/autotest_common.sh@899 -- # local i 00:30:47.857 11:41:31 compress_isal -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:30:47.857 11:41:31 compress_isal -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:30:47.857 11:41:31 compress_isal -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:30:48.115 11:41:31 compress_isal -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b Nvme0n1 -t 2000 00:30:48.374 [ 00:30:48.374 { 00:30:48.374 "name": "Nvme0n1", 00:30:48.374 "aliases": [ 00:30:48.374 "01000000-0000-0000-5cd2-e43197705251" 00:30:48.374 ], 00:30:48.374 "product_name": "NVMe disk", 00:30:48.374 "block_size": 512, 00:30:48.374 "num_blocks": 15002931888, 00:30:48.374 "uuid": "01000000-0000-0000-5cd2-e43197705251", 00:30:48.374 "assigned_rate_limits": { 00:30:48.374 "rw_ios_per_sec": 0, 00:30:48.374 "rw_mbytes_per_sec": 0, 00:30:48.374 "r_mbytes_per_sec": 0, 00:30:48.374 "w_mbytes_per_sec": 0 00:30:48.374 }, 00:30:48.374 "claimed": false, 00:30:48.374 "zoned": false, 00:30:48.374 "supported_io_types": { 00:30:48.374 "read": true, 00:30:48.374 "write": true, 00:30:48.374 "unmap": true, 00:30:48.374 "flush": true, 00:30:48.374 "reset": true, 00:30:48.374 "nvme_admin": true, 00:30:48.374 "nvme_io": true, 00:30:48.374 "nvme_io_md": false, 00:30:48.374 "write_zeroes": true, 00:30:48.374 "zcopy": false, 00:30:48.374 "get_zone_info": false, 00:30:48.374 "zone_management": false, 00:30:48.374 "zone_append": false, 00:30:48.374 "compare": false, 00:30:48.374 "compare_and_write": false, 00:30:48.374 "abort": true, 00:30:48.374 "seek_hole": false, 00:30:48.374 "seek_data": false, 00:30:48.374 "copy": false, 00:30:48.374 "nvme_iov_md": false 00:30:48.374 }, 00:30:48.374 "driver_specific": { 00:30:48.374 "nvme": [ 00:30:48.374 { 00:30:48.374 "pci_address": "0000:5e:00.0", 00:30:48.374 "trid": { 00:30:48.374 "trtype": "PCIe", 00:30:48.374 "traddr": "0000:5e:00.0" 00:30:48.374 }, 00:30:48.374 "ctrlr_data": { 00:30:48.374 "cntlid": 0, 00:30:48.374 "vendor_id": "0x8086", 00:30:48.374 "model_number": "INTEL SSDPF2KX076TZO", 00:30:48.374 "serial_number": "PHAC0301002G7P6CGN", 00:30:48.374 "firmware_revision": "JCV10200", 00:30:48.374 "subnqn": "nqn.2020-07.com.intel:PHAC0301002G7P6CGN ", 00:30:48.374 "oacs": { 00:30:48.374 "security": 1, 00:30:48.374 "format": 1, 00:30:48.374 "firmware": 1, 00:30:48.374 "ns_manage": 1 00:30:48.374 }, 00:30:48.374 "multi_ctrlr": false, 00:30:48.374 "ana_reporting": false 00:30:48.374 }, 00:30:48.374 "vs": { 00:30:48.374 "nvme_version": "1.3" 00:30:48.374 }, 00:30:48.374 "ns_data": { 00:30:48.374 "id": 1, 00:30:48.374 "can_share": false 00:30:48.374 }, 00:30:48.374 "security": { 00:30:48.374 "opal": true 00:30:48.374 } 00:30:48.374 } 00:30:48.374 ], 00:30:48.374 "mp_policy": "active_passive" 00:30:48.374 } 00:30:48.374 } 00:30:48.374 ] 00:30:48.374 11:41:31 compress_isal -- common/autotest_common.sh@905 -- # return 0 00:30:48.374 11:41:31 compress_isal -- compress/compress.sh@37 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --clear-method none Nvme0n1 lvs0 00:30:50.900 0a9a0895-a95e-4bfc-ab9b-1b10ea9c8611 00:30:50.900 11:41:34 compress_isal -- compress/compress.sh@38 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -t -l lvs0 lv0 100 00:30:51.158 f48509aa-980b-44ad-9860-2a0e35ad8560 00:30:51.158 11:41:34 compress_isal -- compress/compress.sh@39 -- # waitforbdev lvs0/lv0 00:30:51.158 11:41:34 compress_isal -- common/autotest_common.sh@897 -- # local bdev_name=lvs0/lv0 00:30:51.158 11:41:34 compress_isal -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:30:51.158 11:41:34 compress_isal -- common/autotest_common.sh@899 -- # local i 00:30:51.158 11:41:34 compress_isal -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:30:51.158 11:41:34 compress_isal -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:30:51.158 11:41:34 compress_isal -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:30:51.415 11:41:34 compress_isal -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b lvs0/lv0 -t 2000 00:30:51.673 [ 00:30:51.673 { 00:30:51.673 "name": "f48509aa-980b-44ad-9860-2a0e35ad8560", 00:30:51.673 "aliases": [ 00:30:51.673 "lvs0/lv0" 00:30:51.673 ], 00:30:51.673 "product_name": "Logical Volume", 00:30:51.673 "block_size": 512, 00:30:51.673 "num_blocks": 204800, 00:30:51.673 "uuid": "f48509aa-980b-44ad-9860-2a0e35ad8560", 00:30:51.673 "assigned_rate_limits": { 00:30:51.673 "rw_ios_per_sec": 0, 00:30:51.673 "rw_mbytes_per_sec": 0, 00:30:51.673 "r_mbytes_per_sec": 0, 00:30:51.673 "w_mbytes_per_sec": 0 00:30:51.673 }, 00:30:51.673 "claimed": false, 00:30:51.673 "zoned": false, 00:30:51.673 "supported_io_types": { 00:30:51.673 "read": true, 00:30:51.673 "write": true, 00:30:51.673 "unmap": true, 00:30:51.673 "flush": false, 00:30:51.673 "reset": true, 00:30:51.673 "nvme_admin": false, 00:30:51.673 "nvme_io": false, 00:30:51.673 "nvme_io_md": false, 00:30:51.673 "write_zeroes": true, 00:30:51.673 "zcopy": false, 00:30:51.673 "get_zone_info": false, 00:30:51.673 "zone_management": false, 00:30:51.673 "zone_append": false, 00:30:51.673 "compare": false, 00:30:51.673 "compare_and_write": false, 00:30:51.673 "abort": false, 00:30:51.673 "seek_hole": true, 00:30:51.673 "seek_data": true, 00:30:51.673 "copy": false, 00:30:51.673 "nvme_iov_md": false 00:30:51.673 }, 00:30:51.673 "driver_specific": { 00:30:51.673 "lvol": { 00:30:51.673 "lvol_store_uuid": "0a9a0895-a95e-4bfc-ab9b-1b10ea9c8611", 00:30:51.673 "base_bdev": "Nvme0n1", 00:30:51.673 "thin_provision": true, 00:30:51.673 "num_allocated_clusters": 0, 00:30:51.673 "snapshot": false, 00:30:51.673 "clone": false, 00:30:51.673 "esnap_clone": false 00:30:51.673 } 00:30:51.673 } 00:30:51.673 } 00:30:51.673 ] 00:30:51.673 11:41:35 compress_isal -- common/autotest_common.sh@905 -- # return 0 00:30:51.673 11:41:35 compress_isal -- compress/compress.sh@41 -- # '[' -z 512 ']' 00:30:51.673 11:41:35 compress_isal -- compress/compress.sh@44 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_compress_create -b lvs0/lv0 -p /tmp/pmem -l 512 00:30:51.960 [2024-07-15 11:41:35.279004] vbdev_compress.c:1016:vbdev_compress_claim: *NOTICE*: registered io_device and virtual bdev for: COMP_lvs0/lv0 00:30:51.960 COMP_lvs0/lv0 00:30:51.960 11:41:35 compress_isal -- compress/compress.sh@46 -- # waitforbdev COMP_lvs0/lv0 00:30:51.960 11:41:35 compress_isal -- common/autotest_common.sh@897 -- # local bdev_name=COMP_lvs0/lv0 00:30:51.960 11:41:35 compress_isal -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:30:51.960 11:41:35 compress_isal -- common/autotest_common.sh@899 -- # local i 00:30:51.960 11:41:35 compress_isal -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:30:51.960 11:41:35 compress_isal -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:30:51.960 11:41:35 compress_isal -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:30:51.960 11:41:35 compress_isal -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b COMP_lvs0/lv0 -t 2000 00:30:52.220 [ 00:30:52.220 { 00:30:52.220 "name": "COMP_lvs0/lv0", 00:30:52.220 "aliases": [ 00:30:52.220 "2b9c2956-1e33-531e-8168-70223297d829" 00:30:52.220 ], 00:30:52.220 "product_name": "compress", 00:30:52.220 "block_size": 512, 00:30:52.220 "num_blocks": 200704, 00:30:52.220 "uuid": "2b9c2956-1e33-531e-8168-70223297d829", 00:30:52.220 "assigned_rate_limits": { 00:30:52.220 "rw_ios_per_sec": 0, 00:30:52.220 "rw_mbytes_per_sec": 0, 00:30:52.220 "r_mbytes_per_sec": 0, 00:30:52.220 "w_mbytes_per_sec": 0 00:30:52.220 }, 00:30:52.220 "claimed": false, 00:30:52.220 "zoned": false, 00:30:52.220 "supported_io_types": { 00:30:52.220 "read": true, 00:30:52.220 "write": true, 00:30:52.220 "unmap": false, 00:30:52.220 "flush": false, 00:30:52.220 "reset": false, 00:30:52.220 "nvme_admin": false, 00:30:52.220 "nvme_io": false, 00:30:52.220 "nvme_io_md": false, 00:30:52.220 "write_zeroes": true, 00:30:52.220 "zcopy": false, 00:30:52.220 "get_zone_info": false, 00:30:52.220 "zone_management": false, 00:30:52.220 "zone_append": false, 00:30:52.220 "compare": false, 00:30:52.220 "compare_and_write": false, 00:30:52.220 "abort": false, 00:30:52.220 "seek_hole": false, 00:30:52.220 "seek_data": false, 00:30:52.220 "copy": false, 00:30:52.220 "nvme_iov_md": false 00:30:52.220 }, 00:30:52.220 "driver_specific": { 00:30:52.220 "compress": { 00:30:52.220 "name": "COMP_lvs0/lv0", 00:30:52.220 "base_bdev_name": "f48509aa-980b-44ad-9860-2a0e35ad8560" 00:30:52.220 } 00:30:52.220 } 00:30:52.220 } 00:30:52.220 ] 00:30:52.220 11:41:35 compress_isal -- common/autotest_common.sh@905 -- # return 0 00:30:52.220 11:41:35 compress_isal -- compress/compress.sh@75 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:30:52.479 Running I/O for 3 seconds... 00:30:55.799 00:30:55.799 Latency(us) 00:30:55.799 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:30:55.799 Job: COMP_lvs0/lv0 (Core Mask 0x2, workload: verify, depth: 32, IO size: 4096) 00:30:55.799 Verification LBA range: start 0x0 length 0x3100 00:30:55.799 COMP_lvs0/lv0 : 3.00 3959.06 15.47 0.00 0.00 8028.18 644.67 6981.01 00:30:55.799 Job: COMP_lvs0/lv0 (Core Mask 0x4, workload: verify, depth: 32, IO size: 4096) 00:30:55.799 Verification LBA range: start 0x3100 length 0x3100 00:30:55.799 COMP_lvs0/lv0 : 3.00 3965.62 15.49 0.00 0.00 8027.77 544.95 6781.55 00:30:55.799 =================================================================================================================== 00:30:55.799 Total : 7924.68 30.96 0.00 0.00 8027.98 544.95 6981.01 00:30:55.799 0 00:30:55.799 11:41:38 compress_isal -- compress/compress.sh@76 -- # destroy_vols 00:30:55.799 11:41:38 compress_isal -- compress/compress.sh@29 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_compress_delete COMP_lvs0/lv0 00:30:55.799 11:41:39 compress_isal -- compress/compress.sh@30 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -l lvs0 00:30:56.058 11:41:39 compress_isal -- compress/compress.sh@77 -- # trap - SIGINT SIGTERM EXIT 00:30:56.058 11:41:39 compress_isal -- compress/compress.sh@78 -- # killprocess 1037624 00:30:56.058 11:41:39 compress_isal -- common/autotest_common.sh@948 -- # '[' -z 1037624 ']' 00:30:56.058 11:41:39 compress_isal -- common/autotest_common.sh@952 -- # kill -0 1037624 00:30:56.058 11:41:39 compress_isal -- common/autotest_common.sh@953 -- # uname 00:30:56.058 11:41:39 compress_isal -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:30:56.058 11:41:39 compress_isal -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1037624 00:30:56.058 11:41:39 compress_isal -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:30:56.058 11:41:39 compress_isal -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:30:56.058 11:41:39 compress_isal -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1037624' 00:30:56.058 killing process with pid 1037624 00:30:56.058 11:41:39 compress_isal -- common/autotest_common.sh@967 -- # kill 1037624 00:30:56.058 Received shutdown signal, test time was about 3.000000 seconds 00:30:56.058 00:30:56.058 Latency(us) 00:30:56.058 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:30:56.058 =================================================================================================================== 00:30:56.058 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:30:56.058 11:41:39 compress_isal -- common/autotest_common.sh@972 -- # wait 1037624 00:30:59.341 11:41:42 compress_isal -- compress/compress.sh@88 -- # run_bdevperf 32 4096 3 4096 00:30:59.341 11:41:42 compress_isal -- compress/compress.sh@66 -- # [[ isal == \c\o\m\p\d\e\v ]] 00:30:59.341 11:41:42 compress_isal -- compress/compress.sh@71 -- # bdevperf_pid=1039331 00:30:59.341 11:41:42 compress_isal -- compress/compress.sh@72 -- # trap 'killprocess $bdevperf_pid; error_cleanup; exit 1' SIGINT SIGTERM EXIT 00:30:59.341 11:41:42 compress_isal -- compress/compress.sh@69 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -z -q 32 -o 4096 -w verify -t 3 -C -m 0x6 00:30:59.341 11:41:42 compress_isal -- compress/compress.sh@73 -- # waitforlisten 1039331 00:30:59.341 11:41:42 compress_isal -- common/autotest_common.sh@829 -- # '[' -z 1039331 ']' 00:30:59.341 11:41:42 compress_isal -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:59.341 11:41:42 compress_isal -- common/autotest_common.sh@834 -- # local max_retries=100 00:30:59.341 11:41:42 compress_isal -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:59.341 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:59.341 11:41:42 compress_isal -- common/autotest_common.sh@838 -- # xtrace_disable 00:30:59.341 11:41:42 compress_isal -- common/autotest_common.sh@10 -- # set +x 00:30:59.341 [2024-07-15 11:41:42.507772] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:30:59.341 [2024-07-15 11:41:42.507845] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x6 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1039331 ] 00:30:59.341 [2024-07-15 11:41:42.628294] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:30:59.341 [2024-07-15 11:41:42.731142] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:30:59.341 [2024-07-15 11:41:42.731148] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:30:59.906 11:41:43 compress_isal -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:30:59.906 11:41:43 compress_isal -- common/autotest_common.sh@862 -- # return 0 00:30:59.906 11:41:43 compress_isal -- compress/compress.sh@74 -- # create_vols 4096 00:30:59.906 11:41:43 compress_isal -- compress/compress.sh@34 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/gen_nvme.sh 00:30:59.907 11:41:43 compress_isal -- compress/compress.sh@34 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py load_subsystem_config 00:31:00.472 11:41:44 compress_isal -- compress/compress.sh@35 -- # waitforbdev Nvme0n1 00:31:00.472 11:41:44 compress_isal -- common/autotest_common.sh@897 -- # local bdev_name=Nvme0n1 00:31:00.472 11:41:44 compress_isal -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:31:00.472 11:41:44 compress_isal -- common/autotest_common.sh@899 -- # local i 00:31:00.472 11:41:44 compress_isal -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:31:00.472 11:41:44 compress_isal -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:31:00.472 11:41:44 compress_isal -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:31:00.730 11:41:44 compress_isal -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b Nvme0n1 -t 2000 00:31:00.987 [ 00:31:00.987 { 00:31:00.987 "name": "Nvme0n1", 00:31:00.987 "aliases": [ 00:31:00.987 "01000000-0000-0000-5cd2-e43197705251" 00:31:00.987 ], 00:31:00.987 "product_name": "NVMe disk", 00:31:00.987 "block_size": 512, 00:31:00.987 "num_blocks": 15002931888, 00:31:00.987 "uuid": "01000000-0000-0000-5cd2-e43197705251", 00:31:00.987 "assigned_rate_limits": { 00:31:00.987 "rw_ios_per_sec": 0, 00:31:00.987 "rw_mbytes_per_sec": 0, 00:31:00.987 "r_mbytes_per_sec": 0, 00:31:00.987 "w_mbytes_per_sec": 0 00:31:00.987 }, 00:31:00.987 "claimed": false, 00:31:00.987 "zoned": false, 00:31:00.987 "supported_io_types": { 00:31:00.987 "read": true, 00:31:00.987 "write": true, 00:31:00.987 "unmap": true, 00:31:00.987 "flush": true, 00:31:00.987 "reset": true, 00:31:00.987 "nvme_admin": true, 00:31:00.987 "nvme_io": true, 00:31:00.987 "nvme_io_md": false, 00:31:00.987 "write_zeroes": true, 00:31:00.987 "zcopy": false, 00:31:00.987 "get_zone_info": false, 00:31:00.987 "zone_management": false, 00:31:00.987 "zone_append": false, 00:31:00.987 "compare": false, 00:31:00.987 "compare_and_write": false, 00:31:00.987 "abort": true, 00:31:00.987 "seek_hole": false, 00:31:00.987 "seek_data": false, 00:31:00.987 "copy": false, 00:31:00.987 "nvme_iov_md": false 00:31:00.987 }, 00:31:00.987 "driver_specific": { 00:31:00.987 "nvme": [ 00:31:00.987 { 00:31:00.987 "pci_address": "0000:5e:00.0", 00:31:00.987 "trid": { 00:31:00.987 "trtype": "PCIe", 00:31:00.987 "traddr": "0000:5e:00.0" 00:31:00.987 }, 00:31:00.987 "ctrlr_data": { 00:31:00.987 "cntlid": 0, 00:31:00.987 "vendor_id": "0x8086", 00:31:00.987 "model_number": "INTEL SSDPF2KX076TZO", 00:31:00.987 "serial_number": "PHAC0301002G7P6CGN", 00:31:00.987 "firmware_revision": "JCV10200", 00:31:00.987 "subnqn": "nqn.2020-07.com.intel:PHAC0301002G7P6CGN ", 00:31:00.987 "oacs": { 00:31:00.987 "security": 1, 00:31:00.987 "format": 1, 00:31:00.987 "firmware": 1, 00:31:00.987 "ns_manage": 1 00:31:00.987 }, 00:31:00.987 "multi_ctrlr": false, 00:31:00.987 "ana_reporting": false 00:31:00.987 }, 00:31:00.987 "vs": { 00:31:00.987 "nvme_version": "1.3" 00:31:00.987 }, 00:31:00.987 "ns_data": { 00:31:00.987 "id": 1, 00:31:00.987 "can_share": false 00:31:00.987 }, 00:31:00.987 "security": { 00:31:00.987 "opal": true 00:31:00.987 } 00:31:00.987 } 00:31:00.987 ], 00:31:00.987 "mp_policy": "active_passive" 00:31:00.987 } 00:31:00.987 } 00:31:00.987 ] 00:31:00.987 11:41:44 compress_isal -- common/autotest_common.sh@905 -- # return 0 00:31:00.987 11:41:44 compress_isal -- compress/compress.sh@37 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --clear-method none Nvme0n1 lvs0 00:31:03.515 1fc4f935-4ef7-486c-999c-a6a85f0d3f83 00:31:03.515 11:41:46 compress_isal -- compress/compress.sh@38 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -t -l lvs0 lv0 100 00:31:03.773 9e001426-0710-4fe5-b155-423759f5baaf 00:31:03.773 11:41:47 compress_isal -- compress/compress.sh@39 -- # waitforbdev lvs0/lv0 00:31:03.773 11:41:47 compress_isal -- common/autotest_common.sh@897 -- # local bdev_name=lvs0/lv0 00:31:03.773 11:41:47 compress_isal -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:31:03.773 11:41:47 compress_isal -- common/autotest_common.sh@899 -- # local i 00:31:03.773 11:41:47 compress_isal -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:31:03.773 11:41:47 compress_isal -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:31:03.773 11:41:47 compress_isal -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:31:04.031 11:41:47 compress_isal -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b lvs0/lv0 -t 2000 00:31:04.290 [ 00:31:04.290 { 00:31:04.290 "name": "9e001426-0710-4fe5-b155-423759f5baaf", 00:31:04.290 "aliases": [ 00:31:04.290 "lvs0/lv0" 00:31:04.290 ], 00:31:04.290 "product_name": "Logical Volume", 00:31:04.290 "block_size": 512, 00:31:04.290 "num_blocks": 204800, 00:31:04.290 "uuid": "9e001426-0710-4fe5-b155-423759f5baaf", 00:31:04.290 "assigned_rate_limits": { 00:31:04.290 "rw_ios_per_sec": 0, 00:31:04.290 "rw_mbytes_per_sec": 0, 00:31:04.290 "r_mbytes_per_sec": 0, 00:31:04.290 "w_mbytes_per_sec": 0 00:31:04.290 }, 00:31:04.290 "claimed": false, 00:31:04.290 "zoned": false, 00:31:04.290 "supported_io_types": { 00:31:04.290 "read": true, 00:31:04.290 "write": true, 00:31:04.290 "unmap": true, 00:31:04.290 "flush": false, 00:31:04.291 "reset": true, 00:31:04.291 "nvme_admin": false, 00:31:04.291 "nvme_io": false, 00:31:04.291 "nvme_io_md": false, 00:31:04.291 "write_zeroes": true, 00:31:04.291 "zcopy": false, 00:31:04.291 "get_zone_info": false, 00:31:04.291 "zone_management": false, 00:31:04.291 "zone_append": false, 00:31:04.291 "compare": false, 00:31:04.291 "compare_and_write": false, 00:31:04.291 "abort": false, 00:31:04.291 "seek_hole": true, 00:31:04.291 "seek_data": true, 00:31:04.291 "copy": false, 00:31:04.291 "nvme_iov_md": false 00:31:04.291 }, 00:31:04.291 "driver_specific": { 00:31:04.291 "lvol": { 00:31:04.291 "lvol_store_uuid": "1fc4f935-4ef7-486c-999c-a6a85f0d3f83", 00:31:04.291 "base_bdev": "Nvme0n1", 00:31:04.291 "thin_provision": true, 00:31:04.291 "num_allocated_clusters": 0, 00:31:04.291 "snapshot": false, 00:31:04.291 "clone": false, 00:31:04.291 "esnap_clone": false 00:31:04.291 } 00:31:04.291 } 00:31:04.291 } 00:31:04.291 ] 00:31:04.291 11:41:47 compress_isal -- common/autotest_common.sh@905 -- # return 0 00:31:04.291 11:41:47 compress_isal -- compress/compress.sh@41 -- # '[' -z 4096 ']' 00:31:04.291 11:41:47 compress_isal -- compress/compress.sh@44 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_compress_create -b lvs0/lv0 -p /tmp/pmem -l 4096 00:31:04.291 [2024-07-15 11:41:47.879209] vbdev_compress.c:1016:vbdev_compress_claim: *NOTICE*: registered io_device and virtual bdev for: COMP_lvs0/lv0 00:31:04.291 COMP_lvs0/lv0 00:31:04.575 11:41:47 compress_isal -- compress/compress.sh@46 -- # waitforbdev COMP_lvs0/lv0 00:31:04.575 11:41:47 compress_isal -- common/autotest_common.sh@897 -- # local bdev_name=COMP_lvs0/lv0 00:31:04.575 11:41:47 compress_isal -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:31:04.575 11:41:47 compress_isal -- common/autotest_common.sh@899 -- # local i 00:31:04.575 11:41:47 compress_isal -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:31:04.575 11:41:47 compress_isal -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:31:04.575 11:41:47 compress_isal -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:31:04.575 11:41:48 compress_isal -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b COMP_lvs0/lv0 -t 2000 00:31:04.833 [ 00:31:04.833 { 00:31:04.833 "name": "COMP_lvs0/lv0", 00:31:04.833 "aliases": [ 00:31:04.833 "0775e99c-77c9-5015-a20c-5cfddb749b73" 00:31:04.833 ], 00:31:04.833 "product_name": "compress", 00:31:04.833 "block_size": 4096, 00:31:04.833 "num_blocks": 25088, 00:31:04.833 "uuid": "0775e99c-77c9-5015-a20c-5cfddb749b73", 00:31:04.833 "assigned_rate_limits": { 00:31:04.833 "rw_ios_per_sec": 0, 00:31:04.833 "rw_mbytes_per_sec": 0, 00:31:04.833 "r_mbytes_per_sec": 0, 00:31:04.833 "w_mbytes_per_sec": 0 00:31:04.833 }, 00:31:04.833 "claimed": false, 00:31:04.833 "zoned": false, 00:31:04.833 "supported_io_types": { 00:31:04.833 "read": true, 00:31:04.833 "write": true, 00:31:04.833 "unmap": false, 00:31:04.833 "flush": false, 00:31:04.833 "reset": false, 00:31:04.833 "nvme_admin": false, 00:31:04.833 "nvme_io": false, 00:31:04.833 "nvme_io_md": false, 00:31:04.833 "write_zeroes": true, 00:31:04.833 "zcopy": false, 00:31:04.833 "get_zone_info": false, 00:31:04.833 "zone_management": false, 00:31:04.833 "zone_append": false, 00:31:04.833 "compare": false, 00:31:04.833 "compare_and_write": false, 00:31:04.833 "abort": false, 00:31:04.833 "seek_hole": false, 00:31:04.833 "seek_data": false, 00:31:04.833 "copy": false, 00:31:04.833 "nvme_iov_md": false 00:31:04.833 }, 00:31:04.833 "driver_specific": { 00:31:04.833 "compress": { 00:31:04.833 "name": "COMP_lvs0/lv0", 00:31:04.833 "base_bdev_name": "9e001426-0710-4fe5-b155-423759f5baaf" 00:31:04.833 } 00:31:04.833 } 00:31:04.833 } 00:31:04.833 ] 00:31:04.833 11:41:48 compress_isal -- common/autotest_common.sh@905 -- # return 0 00:31:04.833 11:41:48 compress_isal -- compress/compress.sh@75 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:31:04.833 Running I/O for 3 seconds... 00:31:08.115 00:31:08.115 Latency(us) 00:31:08.115 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:31:08.115 Job: COMP_lvs0/lv0 (Core Mask 0x2, workload: verify, depth: 32, IO size: 4096) 00:31:08.115 Verification LBA range: start 0x0 length 0x3100 00:31:08.115 COMP_lvs0/lv0 : 3.00 3953.92 15.44 0.00 0.00 8039.27 712.35 8263.23 00:31:08.115 Job: COMP_lvs0/lv0 (Core Mask 0x4, workload: verify, depth: 32, IO size: 4096) 00:31:08.115 Verification LBA range: start 0x3100 length 0x3100 00:31:08.115 COMP_lvs0/lv0 : 3.00 3957.85 15.46 0.00 0.00 8044.02 505.77 8263.23 00:31:08.115 =================================================================================================================== 00:31:08.115 Total : 7911.76 30.91 0.00 0.00 8041.64 505.77 8263.23 00:31:08.115 0 00:31:08.115 11:41:51 compress_isal -- compress/compress.sh@76 -- # destroy_vols 00:31:08.115 11:41:51 compress_isal -- compress/compress.sh@29 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_compress_delete COMP_lvs0/lv0 00:31:08.115 11:41:51 compress_isal -- compress/compress.sh@30 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -l lvs0 00:31:08.374 11:41:51 compress_isal -- compress/compress.sh@77 -- # trap - SIGINT SIGTERM EXIT 00:31:08.374 11:41:51 compress_isal -- compress/compress.sh@78 -- # killprocess 1039331 00:31:08.374 11:41:51 compress_isal -- common/autotest_common.sh@948 -- # '[' -z 1039331 ']' 00:31:08.374 11:41:51 compress_isal -- common/autotest_common.sh@952 -- # kill -0 1039331 00:31:08.374 11:41:51 compress_isal -- common/autotest_common.sh@953 -- # uname 00:31:08.374 11:41:51 compress_isal -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:31:08.374 11:41:51 compress_isal -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1039331 00:31:08.633 11:41:52 compress_isal -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:31:08.633 11:41:52 compress_isal -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:31:08.633 11:41:52 compress_isal -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1039331' 00:31:08.633 killing process with pid 1039331 00:31:08.633 11:41:52 compress_isal -- common/autotest_common.sh@967 -- # kill 1039331 00:31:08.633 Received shutdown signal, test time was about 3.000000 seconds 00:31:08.633 00:31:08.633 Latency(us) 00:31:08.633 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:31:08.633 =================================================================================================================== 00:31:08.633 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:31:08.633 11:41:52 compress_isal -- common/autotest_common.sh@972 -- # wait 1039331 00:31:11.921 11:41:54 compress_isal -- compress/compress.sh@89 -- # run_bdevio 00:31:11.921 11:41:54 compress_isal -- compress/compress.sh@50 -- # [[ isal == \c\o\m\p\d\e\v ]] 00:31:11.921 11:41:54 compress_isal -- compress/compress.sh@55 -- # bdevio_pid=1040935 00:31:11.921 11:41:54 compress_isal -- compress/compress.sh@56 -- # trap 'killprocess $bdevio_pid; error_cleanup; exit 1' SIGINT SIGTERM EXIT 00:31:11.921 11:41:54 compress_isal -- compress/compress.sh@53 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevio/bdevio -w 00:31:11.921 11:41:54 compress_isal -- compress/compress.sh@57 -- # waitforlisten 1040935 00:31:11.921 11:41:54 compress_isal -- common/autotest_common.sh@829 -- # '[' -z 1040935 ']' 00:31:11.921 11:41:54 compress_isal -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:31:11.921 11:41:54 compress_isal -- common/autotest_common.sh@834 -- # local max_retries=100 00:31:11.921 11:41:54 compress_isal -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:31:11.921 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:31:11.921 11:41:54 compress_isal -- common/autotest_common.sh@838 -- # xtrace_disable 00:31:11.921 11:41:54 compress_isal -- common/autotest_common.sh@10 -- # set +x 00:31:11.921 [2024-07-15 11:41:55.037980] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:31:11.921 [2024-07-15 11:41:55.038050] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1040935 ] 00:31:11.921 [2024-07-15 11:41:55.168814] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:31:11.921 [2024-07-15 11:41:55.272488] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:31:11.921 [2024-07-15 11:41:55.272577] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:31:11.921 [2024-07-15 11:41:55.272582] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:31:12.501 11:41:55 compress_isal -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:31:12.501 11:41:55 compress_isal -- common/autotest_common.sh@862 -- # return 0 00:31:12.501 11:41:55 compress_isal -- compress/compress.sh@58 -- # create_vols 00:31:12.501 11:41:55 compress_isal -- compress/compress.sh@34 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/gen_nvme.sh 00:31:12.501 11:41:55 compress_isal -- compress/compress.sh@34 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py load_subsystem_config 00:31:13.069 11:41:56 compress_isal -- compress/compress.sh@35 -- # waitforbdev Nvme0n1 00:31:13.069 11:41:56 compress_isal -- common/autotest_common.sh@897 -- # local bdev_name=Nvme0n1 00:31:13.069 11:41:56 compress_isal -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:31:13.069 11:41:56 compress_isal -- common/autotest_common.sh@899 -- # local i 00:31:13.069 11:41:56 compress_isal -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:31:13.069 11:41:56 compress_isal -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:31:13.069 11:41:56 compress_isal -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:31:13.328 11:41:56 compress_isal -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b Nvme0n1 -t 2000 00:31:13.588 [ 00:31:13.588 { 00:31:13.588 "name": "Nvme0n1", 00:31:13.588 "aliases": [ 00:31:13.588 "01000000-0000-0000-5cd2-e43197705251" 00:31:13.588 ], 00:31:13.588 "product_name": "NVMe disk", 00:31:13.588 "block_size": 512, 00:31:13.588 "num_blocks": 15002931888, 00:31:13.588 "uuid": "01000000-0000-0000-5cd2-e43197705251", 00:31:13.588 "assigned_rate_limits": { 00:31:13.588 "rw_ios_per_sec": 0, 00:31:13.588 "rw_mbytes_per_sec": 0, 00:31:13.588 "r_mbytes_per_sec": 0, 00:31:13.588 "w_mbytes_per_sec": 0 00:31:13.588 }, 00:31:13.588 "claimed": false, 00:31:13.588 "zoned": false, 00:31:13.588 "supported_io_types": { 00:31:13.588 "read": true, 00:31:13.588 "write": true, 00:31:13.588 "unmap": true, 00:31:13.588 "flush": true, 00:31:13.588 "reset": true, 00:31:13.588 "nvme_admin": true, 00:31:13.588 "nvme_io": true, 00:31:13.588 "nvme_io_md": false, 00:31:13.588 "write_zeroes": true, 00:31:13.588 "zcopy": false, 00:31:13.588 "get_zone_info": false, 00:31:13.588 "zone_management": false, 00:31:13.588 "zone_append": false, 00:31:13.588 "compare": false, 00:31:13.588 "compare_and_write": false, 00:31:13.588 "abort": true, 00:31:13.588 "seek_hole": false, 00:31:13.588 "seek_data": false, 00:31:13.588 "copy": false, 00:31:13.588 "nvme_iov_md": false 00:31:13.588 }, 00:31:13.588 "driver_specific": { 00:31:13.588 "nvme": [ 00:31:13.588 { 00:31:13.588 "pci_address": "0000:5e:00.0", 00:31:13.588 "trid": { 00:31:13.588 "trtype": "PCIe", 00:31:13.588 "traddr": "0000:5e:00.0" 00:31:13.588 }, 00:31:13.588 "ctrlr_data": { 00:31:13.588 "cntlid": 0, 00:31:13.588 "vendor_id": "0x8086", 00:31:13.588 "model_number": "INTEL SSDPF2KX076TZO", 00:31:13.588 "serial_number": "PHAC0301002G7P6CGN", 00:31:13.588 "firmware_revision": "JCV10200", 00:31:13.588 "subnqn": "nqn.2020-07.com.intel:PHAC0301002G7P6CGN ", 00:31:13.588 "oacs": { 00:31:13.588 "security": 1, 00:31:13.588 "format": 1, 00:31:13.588 "firmware": 1, 00:31:13.588 "ns_manage": 1 00:31:13.588 }, 00:31:13.588 "multi_ctrlr": false, 00:31:13.588 "ana_reporting": false 00:31:13.588 }, 00:31:13.588 "vs": { 00:31:13.588 "nvme_version": "1.3" 00:31:13.588 }, 00:31:13.588 "ns_data": { 00:31:13.588 "id": 1, 00:31:13.588 "can_share": false 00:31:13.588 }, 00:31:13.588 "security": { 00:31:13.588 "opal": true 00:31:13.588 } 00:31:13.588 } 00:31:13.588 ], 00:31:13.588 "mp_policy": "active_passive" 00:31:13.588 } 00:31:13.588 } 00:31:13.588 ] 00:31:13.588 11:41:57 compress_isal -- common/autotest_common.sh@905 -- # return 0 00:31:13.588 11:41:57 compress_isal -- compress/compress.sh@37 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --clear-method none Nvme0n1 lvs0 00:31:16.127 ad267a27-d1c9-495b-9d14-8646ed067c01 00:31:16.127 11:41:59 compress_isal -- compress/compress.sh@38 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -t -l lvs0 lv0 100 00:31:16.420 90cefb7c-7033-410a-a4ec-a8c8ea881b06 00:31:16.420 11:41:59 compress_isal -- compress/compress.sh@39 -- # waitforbdev lvs0/lv0 00:31:16.420 11:41:59 compress_isal -- common/autotest_common.sh@897 -- # local bdev_name=lvs0/lv0 00:31:16.420 11:41:59 compress_isal -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:31:16.420 11:41:59 compress_isal -- common/autotest_common.sh@899 -- # local i 00:31:16.420 11:41:59 compress_isal -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:31:16.420 11:41:59 compress_isal -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:31:16.420 11:41:59 compress_isal -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:31:16.420 11:41:59 compress_isal -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b lvs0/lv0 -t 2000 00:31:16.679 [ 00:31:16.679 { 00:31:16.679 "name": "90cefb7c-7033-410a-a4ec-a8c8ea881b06", 00:31:16.679 "aliases": [ 00:31:16.679 "lvs0/lv0" 00:31:16.679 ], 00:31:16.679 "product_name": "Logical Volume", 00:31:16.679 "block_size": 512, 00:31:16.679 "num_blocks": 204800, 00:31:16.679 "uuid": "90cefb7c-7033-410a-a4ec-a8c8ea881b06", 00:31:16.679 "assigned_rate_limits": { 00:31:16.679 "rw_ios_per_sec": 0, 00:31:16.679 "rw_mbytes_per_sec": 0, 00:31:16.679 "r_mbytes_per_sec": 0, 00:31:16.679 "w_mbytes_per_sec": 0 00:31:16.679 }, 00:31:16.679 "claimed": false, 00:31:16.679 "zoned": false, 00:31:16.679 "supported_io_types": { 00:31:16.679 "read": true, 00:31:16.679 "write": true, 00:31:16.679 "unmap": true, 00:31:16.679 "flush": false, 00:31:16.679 "reset": true, 00:31:16.679 "nvme_admin": false, 00:31:16.679 "nvme_io": false, 00:31:16.679 "nvme_io_md": false, 00:31:16.679 "write_zeroes": true, 00:31:16.679 "zcopy": false, 00:31:16.679 "get_zone_info": false, 00:31:16.679 "zone_management": false, 00:31:16.679 "zone_append": false, 00:31:16.679 "compare": false, 00:31:16.679 "compare_and_write": false, 00:31:16.679 "abort": false, 00:31:16.679 "seek_hole": true, 00:31:16.679 "seek_data": true, 00:31:16.679 "copy": false, 00:31:16.679 "nvme_iov_md": false 00:31:16.679 }, 00:31:16.679 "driver_specific": { 00:31:16.679 "lvol": { 00:31:16.679 "lvol_store_uuid": "ad267a27-d1c9-495b-9d14-8646ed067c01", 00:31:16.679 "base_bdev": "Nvme0n1", 00:31:16.679 "thin_provision": true, 00:31:16.679 "num_allocated_clusters": 0, 00:31:16.679 "snapshot": false, 00:31:16.679 "clone": false, 00:31:16.679 "esnap_clone": false 00:31:16.679 } 00:31:16.679 } 00:31:16.679 } 00:31:16.679 ] 00:31:16.679 11:42:00 compress_isal -- common/autotest_common.sh@905 -- # return 0 00:31:16.679 11:42:00 compress_isal -- compress/compress.sh@41 -- # '[' -z '' ']' 00:31:16.679 11:42:00 compress_isal -- compress/compress.sh@42 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_compress_create -b lvs0/lv0 -p /tmp/pmem 00:31:16.938 [2024-07-15 11:42:00.468341] vbdev_compress.c:1016:vbdev_compress_claim: *NOTICE*: registered io_device and virtual bdev for: COMP_lvs0/lv0 00:31:16.938 COMP_lvs0/lv0 00:31:16.938 11:42:00 compress_isal -- compress/compress.sh@46 -- # waitforbdev COMP_lvs0/lv0 00:31:16.938 11:42:00 compress_isal -- common/autotest_common.sh@897 -- # local bdev_name=COMP_lvs0/lv0 00:31:16.938 11:42:00 compress_isal -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:31:16.938 11:42:00 compress_isal -- common/autotest_common.sh@899 -- # local i 00:31:16.938 11:42:00 compress_isal -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:31:16.938 11:42:00 compress_isal -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:31:16.938 11:42:00 compress_isal -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:31:17.196 11:42:00 compress_isal -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b COMP_lvs0/lv0 -t 2000 00:31:17.454 [ 00:31:17.454 { 00:31:17.454 "name": "COMP_lvs0/lv0", 00:31:17.454 "aliases": [ 00:31:17.454 "f566abd0-4833-5bfd-9dab-8a27996910db" 00:31:17.454 ], 00:31:17.454 "product_name": "compress", 00:31:17.454 "block_size": 512, 00:31:17.454 "num_blocks": 200704, 00:31:17.454 "uuid": "f566abd0-4833-5bfd-9dab-8a27996910db", 00:31:17.454 "assigned_rate_limits": { 00:31:17.454 "rw_ios_per_sec": 0, 00:31:17.454 "rw_mbytes_per_sec": 0, 00:31:17.454 "r_mbytes_per_sec": 0, 00:31:17.454 "w_mbytes_per_sec": 0 00:31:17.454 }, 00:31:17.454 "claimed": false, 00:31:17.454 "zoned": false, 00:31:17.454 "supported_io_types": { 00:31:17.454 "read": true, 00:31:17.454 "write": true, 00:31:17.454 "unmap": false, 00:31:17.454 "flush": false, 00:31:17.454 "reset": false, 00:31:17.454 "nvme_admin": false, 00:31:17.454 "nvme_io": false, 00:31:17.454 "nvme_io_md": false, 00:31:17.454 "write_zeroes": true, 00:31:17.454 "zcopy": false, 00:31:17.454 "get_zone_info": false, 00:31:17.454 "zone_management": false, 00:31:17.454 "zone_append": false, 00:31:17.454 "compare": false, 00:31:17.454 "compare_and_write": false, 00:31:17.454 "abort": false, 00:31:17.454 "seek_hole": false, 00:31:17.454 "seek_data": false, 00:31:17.454 "copy": false, 00:31:17.454 "nvme_iov_md": false 00:31:17.454 }, 00:31:17.454 "driver_specific": { 00:31:17.454 "compress": { 00:31:17.454 "name": "COMP_lvs0/lv0", 00:31:17.454 "base_bdev_name": "90cefb7c-7033-410a-a4ec-a8c8ea881b06" 00:31:17.454 } 00:31:17.454 } 00:31:17.454 } 00:31:17.454 ] 00:31:17.454 11:42:00 compress_isal -- common/autotest_common.sh@905 -- # return 0 00:31:17.454 11:42:00 compress_isal -- compress/compress.sh@59 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevio/tests.py perform_tests 00:31:17.714 I/O targets: 00:31:17.714 COMP_lvs0/lv0: 200704 blocks of 512 bytes (98 MiB) 00:31:17.714 00:31:17.714 00:31:17.714 CUnit - A unit testing framework for C - Version 2.1-3 00:31:17.714 http://cunit.sourceforge.net/ 00:31:17.714 00:31:17.714 00:31:17.714 Suite: bdevio tests on: COMP_lvs0/lv0 00:31:17.714 Test: blockdev write read block ...passed 00:31:17.714 Test: blockdev write zeroes read block ...passed 00:31:17.714 Test: blockdev write zeroes read no split ...passed 00:31:17.714 Test: blockdev write zeroes read split ...passed 00:31:17.714 Test: blockdev write zeroes read split partial ...passed 00:31:17.714 Test: blockdev reset ...[2024-07-15 11:42:01.123564] vbdev_compress.c: 252:vbdev_compress_submit_request: *ERROR*: Unknown I/O type 5 00:31:17.714 passed 00:31:17.714 Test: blockdev write read 8 blocks ...passed 00:31:17.714 Test: blockdev write read size > 128k ...passed 00:31:17.714 Test: blockdev write read invalid size ...passed 00:31:17.714 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:31:17.714 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:31:17.714 Test: blockdev write read max offset ...passed 00:31:17.714 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:31:17.714 Test: blockdev writev readv 8 blocks ...passed 00:31:17.714 Test: blockdev writev readv 30 x 1block ...passed 00:31:17.714 Test: blockdev writev readv block ...passed 00:31:17.714 Test: blockdev writev readv size > 128k ...passed 00:31:17.714 Test: blockdev writev readv size > 128k in two iovs ...passed 00:31:17.714 Test: blockdev comparev and writev ...passed 00:31:17.714 Test: blockdev nvme passthru rw ...passed 00:31:17.714 Test: blockdev nvme passthru vendor specific ...passed 00:31:17.714 Test: blockdev nvme admin passthru ...passed 00:31:17.714 Test: blockdev copy ...passed 00:31:17.714 00:31:17.714 Run Summary: Type Total Ran Passed Failed Inactive 00:31:17.714 suites 1 1 n/a 0 0 00:31:17.714 tests 23 23 23 0 0 00:31:17.714 asserts 130 130 130 0 n/a 00:31:17.714 00:31:17.714 Elapsed time = 0.108 seconds 00:31:17.714 0 00:31:17.714 11:42:01 compress_isal -- compress/compress.sh@60 -- # destroy_vols 00:31:17.714 11:42:01 compress_isal -- compress/compress.sh@29 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_compress_delete COMP_lvs0/lv0 00:31:17.973 11:42:01 compress_isal -- compress/compress.sh@30 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -l lvs0 00:31:18.233 11:42:01 compress_isal -- compress/compress.sh@61 -- # trap - SIGINT SIGTERM EXIT 00:31:18.233 11:42:01 compress_isal -- compress/compress.sh@62 -- # killprocess 1040935 00:31:18.233 11:42:01 compress_isal -- common/autotest_common.sh@948 -- # '[' -z 1040935 ']' 00:31:18.233 11:42:01 compress_isal -- common/autotest_common.sh@952 -- # kill -0 1040935 00:31:18.233 11:42:01 compress_isal -- common/autotest_common.sh@953 -- # uname 00:31:18.233 11:42:01 compress_isal -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:31:18.233 11:42:01 compress_isal -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1040935 00:31:18.233 11:42:01 compress_isal -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:31:18.233 11:42:01 compress_isal -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:31:18.233 11:42:01 compress_isal -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1040935' 00:31:18.233 killing process with pid 1040935 00:31:18.233 11:42:01 compress_isal -- common/autotest_common.sh@967 -- # kill 1040935 00:31:18.233 11:42:01 compress_isal -- common/autotest_common.sh@972 -- # wait 1040935 00:31:21.522 11:42:04 compress_isal -- compress/compress.sh@91 -- # '[' 0 -eq 1 ']' 00:31:21.522 11:42:04 compress_isal -- compress/compress.sh@120 -- # rm -rf /tmp/pmem 00:31:21.522 00:31:21.522 real 0m48.404s 00:31:21.522 user 1m53.712s 00:31:21.522 sys 0m4.284s 00:31:21.522 11:42:04 compress_isal -- common/autotest_common.sh@1124 -- # xtrace_disable 00:31:21.522 11:42:04 compress_isal -- common/autotest_common.sh@10 -- # set +x 00:31:21.522 ************************************ 00:31:21.522 END TEST compress_isal 00:31:21.522 ************************************ 00:31:21.522 11:42:04 -- common/autotest_common.sh@1142 -- # return 0 00:31:21.522 11:42:04 -- spdk/autotest.sh@352 -- # '[' 0 -eq 1 ']' 00:31:21.522 11:42:04 -- spdk/autotest.sh@356 -- # '[' 1 -eq 1 ']' 00:31:21.522 11:42:04 -- spdk/autotest.sh@357 -- # run_test blockdev_crypto_aesni /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/blockdev.sh crypto_aesni 00:31:21.522 11:42:04 -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:31:21.522 11:42:04 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:31:21.522 11:42:04 -- common/autotest_common.sh@10 -- # set +x 00:31:21.522 ************************************ 00:31:21.522 START TEST blockdev_crypto_aesni 00:31:21.522 ************************************ 00:31:21.522 11:42:04 blockdev_crypto_aesni -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/blockdev.sh crypto_aesni 00:31:21.522 * Looking for test storage... 00:31:21.522 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev 00:31:21.522 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@10 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbd_common.sh 00:31:21.522 11:42:04 blockdev_crypto_aesni -- bdev/nbd_common.sh@6 -- # set -e 00:31:21.522 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@12 -- # rpc_py=rpc_cmd 00:31:21.522 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@13 -- # conf_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json 00:31:21.522 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@14 -- # nonenclosed_conf_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonenclosed.json 00:31:21.522 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@15 -- # nonarray_conf_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonarray.json 00:31:21.522 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@17 -- # export RPC_PIPE_TIMEOUT=30 00:31:21.522 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@17 -- # RPC_PIPE_TIMEOUT=30 00:31:21.522 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@20 -- # : 00:31:21.522 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@670 -- # QOS_DEV_1=Malloc_0 00:31:21.522 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@671 -- # QOS_DEV_2=Null_1 00:31:21.522 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@672 -- # QOS_RUN_TIME=5 00:31:21.522 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@674 -- # uname -s 00:31:21.523 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@674 -- # '[' Linux = Linux ']' 00:31:21.523 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@676 -- # PRE_RESERVED_MEM=0 00:31:21.523 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@682 -- # test_type=crypto_aesni 00:31:21.523 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@683 -- # crypto_device= 00:31:21.523 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@684 -- # dek= 00:31:21.523 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@685 -- # env_ctx= 00:31:21.523 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@686 -- # wait_for_rpc= 00:31:21.523 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@687 -- # '[' -n '' ']' 00:31:21.523 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@690 -- # [[ crypto_aesni == bdev ]] 00:31:21.523 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@690 -- # [[ crypto_aesni == crypto_* ]] 00:31:21.523 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@691 -- # wait_for_rpc=--wait-for-rpc 00:31:21.523 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@693 -- # start_spdk_tgt 00:31:21.523 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@47 -- # spdk_tgt_pid=1042522 00:31:21.523 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@48 -- # trap 'killprocess "$spdk_tgt_pid"; exit 1' SIGINT SIGTERM EXIT 00:31:21.523 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@49 -- # waitforlisten 1042522 00:31:21.523 11:42:04 blockdev_crypto_aesni -- bdev/blockdev.sh@46 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt '' --wait-for-rpc 00:31:21.523 11:42:04 blockdev_crypto_aesni -- common/autotest_common.sh@829 -- # '[' -z 1042522 ']' 00:31:21.523 11:42:04 blockdev_crypto_aesni -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:31:21.523 11:42:04 blockdev_crypto_aesni -- common/autotest_common.sh@834 -- # local max_retries=100 00:31:21.523 11:42:04 blockdev_crypto_aesni -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:31:21.523 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:31:21.523 11:42:04 blockdev_crypto_aesni -- common/autotest_common.sh@838 -- # xtrace_disable 00:31:21.523 11:42:04 blockdev_crypto_aesni -- common/autotest_common.sh@10 -- # set +x 00:31:21.523 [2024-07-15 11:42:04.939818] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:31:21.523 [2024-07-15 11:42:04.939893] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1042522 ] 00:31:21.523 [2024-07-15 11:42:05.057156] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:31:21.783 [2024-07-15 11:42:05.155576] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:31:22.351 11:42:05 blockdev_crypto_aesni -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:31:22.351 11:42:05 blockdev_crypto_aesni -- common/autotest_common.sh@862 -- # return 0 00:31:22.351 11:42:05 blockdev_crypto_aesni -- bdev/blockdev.sh@694 -- # case "$test_type" in 00:31:22.351 11:42:05 blockdev_crypto_aesni -- bdev/blockdev.sh@705 -- # setup_crypto_aesni_conf 00:31:22.351 11:42:05 blockdev_crypto_aesni -- bdev/blockdev.sh@146 -- # rpc_cmd 00:31:22.351 11:42:05 blockdev_crypto_aesni -- common/autotest_common.sh@559 -- # xtrace_disable 00:31:22.351 11:42:05 blockdev_crypto_aesni -- common/autotest_common.sh@10 -- # set +x 00:31:22.351 [2024-07-15 11:42:05.809667] accel_dpdk_cryptodev.c: 223:accel_dpdk_cryptodev_set_driver: *NOTICE*: Using driver crypto_aesni_mb 00:31:22.351 [2024-07-15 11:42:05.817699] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation encrypt will be assigned to module dpdk_cryptodev 00:31:22.351 [2024-07-15 11:42:05.825717] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation decrypt will be assigned to module dpdk_cryptodev 00:31:22.351 [2024-07-15 11:42:05.897083] accel_dpdk_cryptodev.c:1178:accel_dpdk_cryptodev_init: *NOTICE*: Found crypto devices: 97 00:31:24.883 true 00:31:24.883 true 00:31:24.883 true 00:31:24.883 true 00:31:24.883 Malloc0 00:31:24.883 Malloc1 00:31:24.883 Malloc2 00:31:24.883 Malloc3 00:31:24.883 [2024-07-15 11:42:08.313479] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_1" 00:31:24.883 crypto_ram 00:31:24.883 [2024-07-15 11:42:08.321492] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_2" 00:31:24.883 crypto_ram2 00:31:24.884 [2024-07-15 11:42:08.329515] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_3" 00:31:24.884 crypto_ram3 00:31:24.884 [2024-07-15 11:42:08.337536] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_4" 00:31:24.884 crypto_ram4 00:31:24.884 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:31:24.884 11:42:08 blockdev_crypto_aesni -- bdev/blockdev.sh@737 -- # rpc_cmd bdev_wait_for_examine 00:31:24.884 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@559 -- # xtrace_disable 00:31:24.884 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@10 -- # set +x 00:31:24.884 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:31:24.884 11:42:08 blockdev_crypto_aesni -- bdev/blockdev.sh@740 -- # cat 00:31:24.884 11:42:08 blockdev_crypto_aesni -- bdev/blockdev.sh@740 -- # rpc_cmd save_subsystem_config -n accel 00:31:24.884 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@559 -- # xtrace_disable 00:31:24.884 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@10 -- # set +x 00:31:24.884 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:31:24.884 11:42:08 blockdev_crypto_aesni -- bdev/blockdev.sh@740 -- # rpc_cmd save_subsystem_config -n bdev 00:31:24.884 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@559 -- # xtrace_disable 00:31:24.884 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@10 -- # set +x 00:31:24.884 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:31:24.884 11:42:08 blockdev_crypto_aesni -- bdev/blockdev.sh@740 -- # rpc_cmd save_subsystem_config -n iobuf 00:31:24.884 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@559 -- # xtrace_disable 00:31:24.884 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@10 -- # set +x 00:31:24.884 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:31:24.884 11:42:08 blockdev_crypto_aesni -- bdev/blockdev.sh@748 -- # mapfile -t bdevs 00:31:24.884 11:42:08 blockdev_crypto_aesni -- bdev/blockdev.sh@748 -- # rpc_cmd bdev_get_bdevs 00:31:24.884 11:42:08 blockdev_crypto_aesni -- bdev/blockdev.sh@748 -- # jq -r '.[] | select(.claimed == false)' 00:31:24.884 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@559 -- # xtrace_disable 00:31:24.884 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@10 -- # set +x 00:31:25.143 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:31:25.143 11:42:08 blockdev_crypto_aesni -- bdev/blockdev.sh@749 -- # mapfile -t bdevs_name 00:31:25.143 11:42:08 blockdev_crypto_aesni -- bdev/blockdev.sh@749 -- # jq -r .name 00:31:25.143 11:42:08 blockdev_crypto_aesni -- bdev/blockdev.sh@749 -- # printf '%s\n' '{' ' "name": "crypto_ram",' ' "aliases": [' ' "009cf00b-f018-57da-adb8-e3fc376b9f78"' ' ],' ' "product_name": "crypto",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "009cf00b-f018-57da-adb8-e3fc376b9f78",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc0",' ' "name": "crypto_ram",' ' "key_name": "test_dek_aesni_cbc_1"' ' }' ' }' '}' '{' ' "name": "crypto_ram2",' ' "aliases": [' ' "e010df0d-beb5-5585-a68c-53d917a2a68c"' ' ],' ' "product_name": "crypto",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "e010df0d-beb5-5585-a68c-53d917a2a68c",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc1",' ' "name": "crypto_ram2",' ' "key_name": "test_dek_aesni_cbc_2"' ' }' ' }' '}' '{' ' "name": "crypto_ram3",' ' "aliases": [' ' "a2ff8065-d168-5d41-b165-63ed344581fd"' ' ],' ' "product_name": "crypto",' ' "block_size": 4096,' ' "num_blocks": 8192,' ' "uuid": "a2ff8065-d168-5d41-b165-63ed344581fd",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc2",' ' "name": "crypto_ram3",' ' "key_name": "test_dek_aesni_cbc_3"' ' }' ' }' '}' '{' ' "name": "crypto_ram4",' ' "aliases": [' ' "d3452518-0bbe-55c3-86ce-84884cd1f575"' ' ],' ' "product_name": "crypto",' ' "block_size": 4096,' ' "num_blocks": 8192,' ' "uuid": "d3452518-0bbe-55c3-86ce-84884cd1f575",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc3",' ' "name": "crypto_ram4",' ' "key_name": "test_dek_aesni_cbc_4"' ' }' ' }' '}' 00:31:25.143 11:42:08 blockdev_crypto_aesni -- bdev/blockdev.sh@750 -- # bdev_list=("${bdevs_name[@]}") 00:31:25.143 11:42:08 blockdev_crypto_aesni -- bdev/blockdev.sh@752 -- # hello_world_bdev=crypto_ram 00:31:25.143 11:42:08 blockdev_crypto_aesni -- bdev/blockdev.sh@753 -- # trap - SIGINT SIGTERM EXIT 00:31:25.143 11:42:08 blockdev_crypto_aesni -- bdev/blockdev.sh@754 -- # killprocess 1042522 00:31:25.143 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@948 -- # '[' -z 1042522 ']' 00:31:25.143 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@952 -- # kill -0 1042522 00:31:25.143 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@953 -- # uname 00:31:25.143 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:31:25.143 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1042522 00:31:25.143 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:31:25.143 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:31:25.143 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1042522' 00:31:25.143 killing process with pid 1042522 00:31:25.143 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@967 -- # kill 1042522 00:31:25.143 11:42:08 blockdev_crypto_aesni -- common/autotest_common.sh@972 -- # wait 1042522 00:31:25.710 11:42:09 blockdev_crypto_aesni -- bdev/blockdev.sh@758 -- # trap cleanup SIGINT SIGTERM EXIT 00:31:25.710 11:42:09 blockdev_crypto_aesni -- bdev/blockdev.sh@760 -- # run_test bdev_hello_world /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/hello_bdev --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -b crypto_ram '' 00:31:25.710 11:42:09 blockdev_crypto_aesni -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:31:25.710 11:42:09 blockdev_crypto_aesni -- common/autotest_common.sh@1105 -- # xtrace_disable 00:31:25.710 11:42:09 blockdev_crypto_aesni -- common/autotest_common.sh@10 -- # set +x 00:31:25.710 ************************************ 00:31:25.710 START TEST bdev_hello_world 00:31:25.710 ************************************ 00:31:25.710 11:42:09 blockdev_crypto_aesni.bdev_hello_world -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/hello_bdev --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -b crypto_ram '' 00:31:25.710 [2024-07-15 11:42:09.231507] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:31:25.710 [2024-07-15 11:42:09.231568] [ DPDK EAL parameters: hello_bdev --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1043288 ] 00:31:25.969 [2024-07-15 11:42:09.360528] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:31:25.969 [2024-07-15 11:42:09.456921] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:31:25.969 [2024-07-15 11:42:09.478207] accel_dpdk_cryptodev.c: 223:accel_dpdk_cryptodev_set_driver: *NOTICE*: Using driver crypto_aesni_mb 00:31:25.969 [2024-07-15 11:42:09.486234] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation encrypt will be assigned to module dpdk_cryptodev 00:31:25.969 [2024-07-15 11:42:09.494260] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation decrypt will be assigned to module dpdk_cryptodev 00:31:26.228 [2024-07-15 11:42:09.602385] accel_dpdk_cryptodev.c:1178:accel_dpdk_cryptodev_init: *NOTICE*: Found crypto devices: 97 00:31:28.756 [2024-07-15 11:42:11.819187] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_1" 00:31:28.756 [2024-07-15 11:42:11.819262] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:31:28.756 [2024-07-15 11:42:11.819276] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:31:28.756 [2024-07-15 11:42:11.827205] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_2" 00:31:28.756 [2024-07-15 11:42:11.827226] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:31:28.756 [2024-07-15 11:42:11.827243] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:31:28.756 [2024-07-15 11:42:11.835227] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_3" 00:31:28.756 [2024-07-15 11:42:11.835246] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:31:28.756 [2024-07-15 11:42:11.835258] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:31:28.756 [2024-07-15 11:42:11.843248] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_4" 00:31:28.756 [2024-07-15 11:42:11.843266] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc3 00:31:28.756 [2024-07-15 11:42:11.843278] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:31:28.756 [2024-07-15 11:42:11.915802] hello_bdev.c: 222:hello_start: *NOTICE*: Successfully started the application 00:31:28.756 [2024-07-15 11:42:11.915843] hello_bdev.c: 231:hello_start: *NOTICE*: Opening the bdev crypto_ram 00:31:28.756 [2024-07-15 11:42:11.915862] hello_bdev.c: 244:hello_start: *NOTICE*: Opening io channel 00:31:28.756 [2024-07-15 11:42:11.917132] hello_bdev.c: 138:hello_write: *NOTICE*: Writing to the bdev 00:31:28.756 [2024-07-15 11:42:11.917205] hello_bdev.c: 117:write_complete: *NOTICE*: bdev io write completed successfully 00:31:28.756 [2024-07-15 11:42:11.917223] hello_bdev.c: 84:hello_read: *NOTICE*: Reading io 00:31:28.756 [2024-07-15 11:42:11.917269] hello_bdev.c: 65:read_complete: *NOTICE*: Read string from bdev : Hello World! 00:31:28.756 00:31:28.756 [2024-07-15 11:42:11.917288] hello_bdev.c: 74:read_complete: *NOTICE*: Stopping app 00:31:28.756 00:31:28.756 real 0m3.155s 00:31:28.756 user 0m2.763s 00:31:28.756 sys 0m0.353s 00:31:28.756 11:42:12 blockdev_crypto_aesni.bdev_hello_world -- common/autotest_common.sh@1124 -- # xtrace_disable 00:31:28.756 11:42:12 blockdev_crypto_aesni.bdev_hello_world -- common/autotest_common.sh@10 -- # set +x 00:31:28.756 ************************************ 00:31:28.756 END TEST bdev_hello_world 00:31:28.756 ************************************ 00:31:29.014 11:42:12 blockdev_crypto_aesni -- common/autotest_common.sh@1142 -- # return 0 00:31:29.014 11:42:12 blockdev_crypto_aesni -- bdev/blockdev.sh@761 -- # run_test bdev_bounds bdev_bounds '' 00:31:29.014 11:42:12 blockdev_crypto_aesni -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:31:29.014 11:42:12 blockdev_crypto_aesni -- common/autotest_common.sh@1105 -- # xtrace_disable 00:31:29.014 11:42:12 blockdev_crypto_aesni -- common/autotest_common.sh@10 -- # set +x 00:31:29.014 ************************************ 00:31:29.014 START TEST bdev_bounds 00:31:29.014 ************************************ 00:31:29.014 11:42:12 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@1123 -- # bdev_bounds '' 00:31:29.014 11:42:12 blockdev_crypto_aesni.bdev_bounds -- bdev/blockdev.sh@290 -- # bdevio_pid=1043694 00:31:29.014 11:42:12 blockdev_crypto_aesni.bdev_bounds -- bdev/blockdev.sh@289 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevio/bdevio -w -s 0 --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json '' 00:31:29.014 11:42:12 blockdev_crypto_aesni.bdev_bounds -- bdev/blockdev.sh@291 -- # trap 'cleanup; killprocess $bdevio_pid; exit 1' SIGINT SIGTERM EXIT 00:31:29.014 11:42:12 blockdev_crypto_aesni.bdev_bounds -- bdev/blockdev.sh@292 -- # echo 'Process bdevio pid: 1043694' 00:31:29.014 Process bdevio pid: 1043694 00:31:29.014 11:42:12 blockdev_crypto_aesni.bdev_bounds -- bdev/blockdev.sh@293 -- # waitforlisten 1043694 00:31:29.014 11:42:12 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@829 -- # '[' -z 1043694 ']' 00:31:29.014 11:42:12 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:31:29.014 11:42:12 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@834 -- # local max_retries=100 00:31:29.014 11:42:12 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:31:29.014 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:31:29.014 11:42:12 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@838 -- # xtrace_disable 00:31:29.014 11:42:12 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:31:29.014 [2024-07-15 11:42:12.459671] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:31:29.014 [2024-07-15 11:42:12.459742] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 0 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1043694 ] 00:31:29.014 [2024-07-15 11:42:12.590590] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:31:29.272 [2024-07-15 11:42:12.700003] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:31:29.272 [2024-07-15 11:42:12.700091] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:31:29.272 [2024-07-15 11:42:12.700096] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:31:29.272 [2024-07-15 11:42:12.721428] accel_dpdk_cryptodev.c: 223:accel_dpdk_cryptodev_set_driver: *NOTICE*: Using driver crypto_aesni_mb 00:31:29.272 [2024-07-15 11:42:12.729454] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation encrypt will be assigned to module dpdk_cryptodev 00:31:29.272 [2024-07-15 11:42:12.737475] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation decrypt will be assigned to module dpdk_cryptodev 00:31:29.272 [2024-07-15 11:42:12.843874] accel_dpdk_cryptodev.c:1178:accel_dpdk_cryptodev_init: *NOTICE*: Found crypto devices: 97 00:31:31.802 [2024-07-15 11:42:15.066891] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_1" 00:31:31.802 [2024-07-15 11:42:15.066972] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:31:31.802 [2024-07-15 11:42:15.066987] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:31:31.802 [2024-07-15 11:42:15.074911] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_2" 00:31:31.802 [2024-07-15 11:42:15.074939] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:31:31.802 [2024-07-15 11:42:15.074951] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:31:31.802 [2024-07-15 11:42:15.082938] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_3" 00:31:31.802 [2024-07-15 11:42:15.082961] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:31:31.802 [2024-07-15 11:42:15.082973] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:31:31.802 [2024-07-15 11:42:15.090968] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_4" 00:31:31.802 [2024-07-15 11:42:15.090988] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc3 00:31:31.802 [2024-07-15 11:42:15.090999] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:31:31.802 11:42:15 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:31:31.802 11:42:15 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@862 -- # return 0 00:31:31.802 11:42:15 blockdev_crypto_aesni.bdev_bounds -- bdev/blockdev.sh@294 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevio/tests.py perform_tests 00:31:31.802 I/O targets: 00:31:31.802 crypto_ram: 65536 blocks of 512 bytes (32 MiB) 00:31:31.802 crypto_ram2: 65536 blocks of 512 bytes (32 MiB) 00:31:31.802 crypto_ram3: 8192 blocks of 4096 bytes (32 MiB) 00:31:31.802 crypto_ram4: 8192 blocks of 4096 bytes (32 MiB) 00:31:31.802 00:31:31.802 00:31:31.802 CUnit - A unit testing framework for C - Version 2.1-3 00:31:31.802 http://cunit.sourceforge.net/ 00:31:31.802 00:31:31.802 00:31:31.802 Suite: bdevio tests on: crypto_ram4 00:31:31.802 Test: blockdev write read block ...passed 00:31:31.802 Test: blockdev write zeroes read block ...passed 00:31:31.802 Test: blockdev write zeroes read no split ...passed 00:31:31.802 Test: blockdev write zeroes read split ...passed 00:31:31.802 Test: blockdev write zeroes read split partial ...passed 00:31:31.802 Test: blockdev reset ...passed 00:31:31.802 Test: blockdev write read 8 blocks ...passed 00:31:31.802 Test: blockdev write read size > 128k ...passed 00:31:31.802 Test: blockdev write read invalid size ...passed 00:31:31.802 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:31:31.802 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:31:31.802 Test: blockdev write read max offset ...passed 00:31:31.802 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:31:31.802 Test: blockdev writev readv 8 blocks ...passed 00:31:31.802 Test: blockdev writev readv 30 x 1block ...passed 00:31:31.802 Test: blockdev writev readv block ...passed 00:31:31.802 Test: blockdev writev readv size > 128k ...passed 00:31:31.802 Test: blockdev writev readv size > 128k in two iovs ...passed 00:31:31.802 Test: blockdev comparev and writev ...passed 00:31:31.802 Test: blockdev nvme passthru rw ...passed 00:31:31.802 Test: blockdev nvme passthru vendor specific ...passed 00:31:31.802 Test: blockdev nvme admin passthru ...passed 00:31:31.802 Test: blockdev copy ...passed 00:31:31.802 Suite: bdevio tests on: crypto_ram3 00:31:31.802 Test: blockdev write read block ...passed 00:31:31.802 Test: blockdev write zeroes read block ...passed 00:31:31.802 Test: blockdev write zeroes read no split ...passed 00:31:31.802 Test: blockdev write zeroes read split ...passed 00:31:31.802 Test: blockdev write zeroes read split partial ...passed 00:31:31.802 Test: blockdev reset ...passed 00:31:31.802 Test: blockdev write read 8 blocks ...passed 00:31:31.802 Test: blockdev write read size > 128k ...passed 00:31:31.802 Test: blockdev write read invalid size ...passed 00:31:31.802 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:31:31.802 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:31:31.802 Test: blockdev write read max offset ...passed 00:31:31.802 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:31:31.802 Test: blockdev writev readv 8 blocks ...passed 00:31:31.802 Test: blockdev writev readv 30 x 1block ...passed 00:31:31.802 Test: blockdev writev readv block ...passed 00:31:31.803 Test: blockdev writev readv size > 128k ...passed 00:31:31.803 Test: blockdev writev readv size > 128k in two iovs ...passed 00:31:31.803 Test: blockdev comparev and writev ...passed 00:31:31.803 Test: blockdev nvme passthru rw ...passed 00:31:31.803 Test: blockdev nvme passthru vendor specific ...passed 00:31:31.803 Test: blockdev nvme admin passthru ...passed 00:31:31.803 Test: blockdev copy ...passed 00:31:31.803 Suite: bdevio tests on: crypto_ram2 00:31:31.803 Test: blockdev write read block ...passed 00:31:31.803 Test: blockdev write zeroes read block ...passed 00:31:31.803 Test: blockdev write zeroes read no split ...passed 00:31:32.061 Test: blockdev write zeroes read split ...passed 00:31:32.061 Test: blockdev write zeroes read split partial ...passed 00:31:32.061 Test: blockdev reset ...passed 00:31:32.061 Test: blockdev write read 8 blocks ...passed 00:31:32.061 Test: blockdev write read size > 128k ...passed 00:31:32.061 Test: blockdev write read invalid size ...passed 00:31:32.061 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:31:32.061 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:31:32.061 Test: blockdev write read max offset ...passed 00:31:32.061 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:31:32.061 Test: blockdev writev readv 8 blocks ...passed 00:31:32.061 Test: blockdev writev readv 30 x 1block ...passed 00:31:32.061 Test: blockdev writev readv block ...passed 00:31:32.061 Test: blockdev writev readv size > 128k ...passed 00:31:32.061 Test: blockdev writev readv size > 128k in two iovs ...passed 00:31:32.061 Test: blockdev comparev and writev ...passed 00:31:32.061 Test: blockdev nvme passthru rw ...passed 00:31:32.061 Test: blockdev nvme passthru vendor specific ...passed 00:31:32.061 Test: blockdev nvme admin passthru ...passed 00:31:32.061 Test: blockdev copy ...passed 00:31:32.061 Suite: bdevio tests on: crypto_ram 00:31:32.061 Test: blockdev write read block ...passed 00:31:32.061 Test: blockdev write zeroes read block ...passed 00:31:32.061 Test: blockdev write zeroes read no split ...passed 00:31:32.061 Test: blockdev write zeroes read split ...passed 00:31:32.061 Test: blockdev write zeroes read split partial ...passed 00:31:32.061 Test: blockdev reset ...passed 00:31:32.061 Test: blockdev write read 8 blocks ...passed 00:31:32.061 Test: blockdev write read size > 128k ...passed 00:31:32.061 Test: blockdev write read invalid size ...passed 00:31:32.061 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:31:32.061 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:31:32.061 Test: blockdev write read max offset ...passed 00:31:32.061 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:31:32.061 Test: blockdev writev readv 8 blocks ...passed 00:31:32.061 Test: blockdev writev readv 30 x 1block ...passed 00:31:32.061 Test: blockdev writev readv block ...passed 00:31:32.061 Test: blockdev writev readv size > 128k ...passed 00:31:32.061 Test: blockdev writev readv size > 128k in two iovs ...passed 00:31:32.061 Test: blockdev comparev and writev ...passed 00:31:32.061 Test: blockdev nvme passthru rw ...passed 00:31:32.061 Test: blockdev nvme passthru vendor specific ...passed 00:31:32.061 Test: blockdev nvme admin passthru ...passed 00:31:32.061 Test: blockdev copy ...passed 00:31:32.061 00:31:32.061 Run Summary: Type Total Ran Passed Failed Inactive 00:31:32.061 suites 4 4 n/a 0 0 00:31:32.061 tests 92 92 92 0 0 00:31:32.061 asserts 520 520 520 0 n/a 00:31:32.061 00:31:32.061 Elapsed time = 0.530 seconds 00:31:32.061 0 00:31:32.061 11:42:15 blockdev_crypto_aesni.bdev_bounds -- bdev/blockdev.sh@295 -- # killprocess 1043694 00:31:32.061 11:42:15 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@948 -- # '[' -z 1043694 ']' 00:31:32.061 11:42:15 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@952 -- # kill -0 1043694 00:31:32.061 11:42:15 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@953 -- # uname 00:31:32.061 11:42:15 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:31:32.061 11:42:15 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1043694 00:31:32.061 11:42:15 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:31:32.061 11:42:15 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:31:32.061 11:42:15 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1043694' 00:31:32.061 killing process with pid 1043694 00:31:32.061 11:42:15 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@967 -- # kill 1043694 00:31:32.061 11:42:15 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@972 -- # wait 1043694 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_bounds -- bdev/blockdev.sh@296 -- # trap - SIGINT SIGTERM EXIT 00:31:32.630 00:31:32.630 real 0m3.627s 00:31:32.630 user 0m9.990s 00:31:32.630 sys 0m0.568s 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@1124 -- # xtrace_disable 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:31:32.630 ************************************ 00:31:32.630 END TEST bdev_bounds 00:31:32.630 ************************************ 00:31:32.630 11:42:16 blockdev_crypto_aesni -- common/autotest_common.sh@1142 -- # return 0 00:31:32.630 11:42:16 blockdev_crypto_aesni -- bdev/blockdev.sh@762 -- # run_test bdev_nbd nbd_function_test /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json 'crypto_ram crypto_ram2 crypto_ram3 crypto_ram4' '' 00:31:32.630 11:42:16 blockdev_crypto_aesni -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:31:32.630 11:42:16 blockdev_crypto_aesni -- common/autotest_common.sh@1105 -- # xtrace_disable 00:31:32.630 11:42:16 blockdev_crypto_aesni -- common/autotest_common.sh@10 -- # set +x 00:31:32.630 ************************************ 00:31:32.630 START TEST bdev_nbd 00:31:32.630 ************************************ 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@1123 -- # nbd_function_test /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json 'crypto_ram crypto_ram2 crypto_ram3 crypto_ram4' '' 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@300 -- # uname -s 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@300 -- # [[ Linux == Linux ]] 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@302 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@303 -- # local conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@304 -- # bdev_all=('crypto_ram' 'crypto_ram2' 'crypto_ram3' 'crypto_ram4') 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@304 -- # local bdev_all 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@305 -- # local bdev_num=4 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@309 -- # [[ -e /sys/module/nbd ]] 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@311 -- # nbd_all=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@311 -- # local nbd_all 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@312 -- # bdev_num=4 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@314 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11') 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@314 -- # local nbd_list 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@315 -- # bdev_list=('crypto_ram' 'crypto_ram2' 'crypto_ram3' 'crypto_ram4') 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@315 -- # local bdev_list 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@318 -- # nbd_pid=1044212 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@319 -- # trap 'cleanup; killprocess $nbd_pid' SIGINT SIGTERM EXIT 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@317 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-nbd.sock -i 0 --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json '' 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@320 -- # waitforlisten 1044212 /var/tmp/spdk-nbd.sock 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@829 -- # '[' -z 1044212 ']' 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@834 -- # local max_retries=100 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:31:32.630 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@838 -- # xtrace_disable 00:31:32.630 11:42:16 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:31:32.630 [2024-07-15 11:42:16.182186] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:31:32.630 [2024-07-15 11:42:16.182251] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:31:32.889 [2024-07-15 11:42:16.312778] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:31:32.889 [2024-07-15 11:42:16.414834] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:31:32.889 [2024-07-15 11:42:16.436113] accel_dpdk_cryptodev.c: 223:accel_dpdk_cryptodev_set_driver: *NOTICE*: Using driver crypto_aesni_mb 00:31:32.889 [2024-07-15 11:42:16.444133] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation encrypt will be assigned to module dpdk_cryptodev 00:31:32.889 [2024-07-15 11:42:16.452153] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation decrypt will be assigned to module dpdk_cryptodev 00:31:33.148 [2024-07-15 11:42:16.563663] accel_dpdk_cryptodev.c:1178:accel_dpdk_cryptodev_init: *NOTICE*: Found crypto devices: 97 00:31:35.711 [2024-07-15 11:42:18.784161] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_1" 00:31:35.711 [2024-07-15 11:42:18.784227] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:31:35.711 [2024-07-15 11:42:18.784242] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:31:35.712 [2024-07-15 11:42:18.792179] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_2" 00:31:35.712 [2024-07-15 11:42:18.792199] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:31:35.712 [2024-07-15 11:42:18.792211] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:31:35.712 [2024-07-15 11:42:18.800199] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_3" 00:31:35.712 [2024-07-15 11:42:18.800218] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:31:35.712 [2024-07-15 11:42:18.800229] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:31:35.712 [2024-07-15 11:42:18.808218] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_4" 00:31:35.712 [2024-07-15 11:42:18.808236] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc3 00:31:35.712 [2024-07-15 11:42:18.808252] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:31:35.712 11:42:18 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:31:35.712 11:42:18 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@862 -- # return 0 00:31:35.712 11:42:18 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@322 -- # nbd_rpc_start_stop_verify /var/tmp/spdk-nbd.sock 'crypto_ram crypto_ram2 crypto_ram3 crypto_ram4' 00:31:35.712 11:42:18 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@113 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:31:35.712 11:42:18 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@114 -- # bdev_list=('crypto_ram' 'crypto_ram2' 'crypto_ram3' 'crypto_ram4') 00:31:35.712 11:42:18 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@114 -- # local bdev_list 00:31:35.712 11:42:18 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@116 -- # nbd_start_disks_without_nbd_idx /var/tmp/spdk-nbd.sock 'crypto_ram crypto_ram2 crypto_ram3 crypto_ram4' 00:31:35.712 11:42:18 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@22 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:31:35.712 11:42:18 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@23 -- # bdev_list=('crypto_ram' 'crypto_ram2' 'crypto_ram3' 'crypto_ram4') 00:31:35.712 11:42:18 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@23 -- # local bdev_list 00:31:35.712 11:42:18 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@24 -- # local i 00:31:35.712 11:42:18 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@25 -- # local nbd_device 00:31:35.712 11:42:18 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i = 0 )) 00:31:35.712 11:42:18 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 4 )) 00:31:35.712 11:42:18 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk crypto_ram 00:31:35.712 11:42:19 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd0 00:31:35.712 11:42:19 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd0 00:31:35.712 11:42:19 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd0 00:31:35.712 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:31:35.712 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:31:35.712 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:31:35.712 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:31:35.712 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:31:35.712 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:31:35.712 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:31:35.712 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:31:35.712 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:35.712 1+0 records in 00:31:35.712 1+0 records out 00:31:35.712 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00027533 s, 14.9 MB/s 00:31:35.712 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:31:35.712 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:31:35.712 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:31:35.712 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:31:35.712 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:31:35.712 11:42:19 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:31:35.712 11:42:19 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 4 )) 00:31:35.712 11:42:19 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk crypto_ram2 00:31:35.970 11:42:19 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd1 00:31:35.970 11:42:19 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd1 00:31:35.970 11:42:19 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd1 00:31:35.970 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:31:35.970 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:31:35.970 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:31:35.970 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:31:35.970 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:31:35.970 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:31:35.970 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:31:35.970 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:31:35.970 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:35.970 1+0 records in 00:31:35.970 1+0 records out 00:31:35.970 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000296553 s, 13.8 MB/s 00:31:35.970 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:31:35.970 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:31:35.970 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:31:35.970 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:31:35.970 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:31:35.970 11:42:19 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:31:35.971 11:42:19 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 4 )) 00:31:35.971 11:42:19 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk crypto_ram3 00:31:36.229 11:42:19 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd2 00:31:36.229 11:42:19 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd2 00:31:36.229 11:42:19 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd2 00:31:36.229 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd2 00:31:36.229 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:31:36.229 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:31:36.229 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:31:36.229 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd2 /proc/partitions 00:31:36.229 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:31:36.229 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:31:36.229 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:31:36.229 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd2 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:36.229 1+0 records in 00:31:36.229 1+0 records out 00:31:36.229 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000341068 s, 12.0 MB/s 00:31:36.229 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:31:36.229 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:31:36.229 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:31:36.229 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:31:36.229 11:42:19 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:31:36.229 11:42:19 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:31:36.229 11:42:19 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 4 )) 00:31:36.229 11:42:19 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk crypto_ram4 00:31:36.487 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd3 00:31:36.487 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd3 00:31:36.487 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd3 00:31:36.487 11:42:20 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd3 00:31:36.487 11:42:20 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:31:36.487 11:42:20 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:31:36.487 11:42:20 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:31:36.487 11:42:20 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd3 /proc/partitions 00:31:36.487 11:42:20 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:31:36.487 11:42:20 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:31:36.487 11:42:20 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:31:36.487 11:42:20 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd3 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:36.487 1+0 records in 00:31:36.487 1+0 records out 00:31:36.487 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000325271 s, 12.6 MB/s 00:31:36.487 11:42:20 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:31:36.487 11:42:20 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:31:36.487 11:42:20 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:31:36.487 11:42:20 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:31:36.487 11:42:20 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:31:36.487 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:31:36.487 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 4 )) 00:31:36.487 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@118 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:31:36.746 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@118 -- # nbd_disks_json='[ 00:31:36.746 { 00:31:36.746 "nbd_device": "/dev/nbd0", 00:31:36.746 "bdev_name": "crypto_ram" 00:31:36.746 }, 00:31:36.746 { 00:31:36.746 "nbd_device": "/dev/nbd1", 00:31:36.746 "bdev_name": "crypto_ram2" 00:31:36.746 }, 00:31:36.746 { 00:31:36.746 "nbd_device": "/dev/nbd2", 00:31:36.746 "bdev_name": "crypto_ram3" 00:31:36.746 }, 00:31:36.746 { 00:31:36.746 "nbd_device": "/dev/nbd3", 00:31:36.746 "bdev_name": "crypto_ram4" 00:31:36.746 } 00:31:36.746 ]' 00:31:36.746 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@119 -- # nbd_disks_name=($(echo "${nbd_disks_json}" | jq -r '.[] | .nbd_device')) 00:31:36.746 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@119 -- # echo '[ 00:31:36.746 { 00:31:36.746 "nbd_device": "/dev/nbd0", 00:31:36.746 "bdev_name": "crypto_ram" 00:31:36.746 }, 00:31:36.746 { 00:31:36.746 "nbd_device": "/dev/nbd1", 00:31:36.746 "bdev_name": "crypto_ram2" 00:31:36.746 }, 00:31:36.746 { 00:31:36.746 "nbd_device": "/dev/nbd2", 00:31:36.746 "bdev_name": "crypto_ram3" 00:31:36.746 }, 00:31:36.746 { 00:31:36.746 "nbd_device": "/dev/nbd3", 00:31:36.746 "bdev_name": "crypto_ram4" 00:31:36.746 } 00:31:36.746 ]' 00:31:36.746 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@119 -- # jq -r '.[] | .nbd_device' 00:31:37.005 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@120 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1 /dev/nbd2 /dev/nbd3' 00:31:37.005 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:31:37.005 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd2' '/dev/nbd3') 00:31:37.005 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:31:37.005 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:31:37.005 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:37.005 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:31:37.264 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:31:37.264 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:31:37.264 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:31:37.264 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:37.264 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:37.264 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:31:37.264 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:31:37.264 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:31:37.264 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:37.264 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:31:37.524 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:31:37.524 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:31:37.524 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:31:37.524 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:37.524 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:37.524 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:31:37.524 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:31:37.524 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:31:37.524 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:37.524 11:42:20 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd2 00:31:37.783 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd2 00:31:37.783 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd2 00:31:37.783 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd2 00:31:37.783 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:37.783 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:37.783 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd2 /proc/partitions 00:31:37.783 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:31:37.783 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:31:37.783 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:37.783 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd3 00:31:38.042 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd3 00:31:38.042 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd3 00:31:38.042 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd3 00:31:38.042 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:38.043 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:38.043 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd3 /proc/partitions 00:31:38.043 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:31:38.043 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:31:38.043 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@122 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:31:38.043 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:31:38.043 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@122 -- # count=0 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@123 -- # '[' 0 -ne 0 ']' 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@127 -- # return 0 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@323 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'crypto_ram crypto_ram2 crypto_ram3 crypto_ram4' '/dev/nbd0 /dev/nbd1 /dev/nbd10 /dev/nbd11' 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@91 -- # bdev_list=('crypto_ram' 'crypto_ram2' 'crypto_ram3' 'crypto_ram4') 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@91 -- # local bdev_list 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11') 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@92 -- # local nbd_list 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'crypto_ram crypto_ram2 crypto_ram3 crypto_ram4' '/dev/nbd0 /dev/nbd1 /dev/nbd10 /dev/nbd11' 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@10 -- # bdev_list=('crypto_ram' 'crypto_ram2' 'crypto_ram3' 'crypto_ram4') 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@10 -- # local bdev_list 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11') 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@11 -- # local nbd_list 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@12 -- # local i 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 4 )) 00:31:38.302 11:42:21 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk crypto_ram /dev/nbd0 00:31:38.562 /dev/nbd0 00:31:38.562 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:31:38.562 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:31:38.562 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:31:38.562 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:31:38.562 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:31:38.562 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:31:38.562 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:31:38.562 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:31:38.562 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:31:38.562 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:31:38.562 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:38.562 1+0 records in 00:31:38.562 1+0 records out 00:31:38.562 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000281114 s, 14.6 MB/s 00:31:38.562 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:31:38.562 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:31:38.562 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:31:38.562 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:31:38.562 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:31:38.562 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:31:38.562 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 4 )) 00:31:38.562 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk crypto_ram2 /dev/nbd1 00:31:38.822 /dev/nbd1 00:31:38.822 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:31:38.822 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:31:38.822 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:31:38.822 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:31:38.822 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:31:38.822 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:31:38.822 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:31:38.822 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:31:38.822 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:31:38.822 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:31:38.822 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:38.822 1+0 records in 00:31:38.822 1+0 records out 00:31:38.822 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000501326 s, 8.2 MB/s 00:31:38.822 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:31:38.822 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:31:38.822 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:31:38.822 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:31:38.822 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:31:38.822 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:31:38.822 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 4 )) 00:31:38.822 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk crypto_ram3 /dev/nbd10 00:31:39.082 /dev/nbd10 00:31:39.082 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd10 00:31:39.082 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd10 00:31:39.082 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd10 00:31:39.082 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:31:39.082 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:31:39.082 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:31:39.082 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd10 /proc/partitions 00:31:39.082 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:31:39.082 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:31:39.082 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:31:39.082 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd10 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:39.082 1+0 records in 00:31:39.082 1+0 records out 00:31:39.082 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00031063 s, 13.2 MB/s 00:31:39.082 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:31:39.082 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:31:39.082 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:31:39.082 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:31:39.082 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:31:39.082 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:31:39.082 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 4 )) 00:31:39.082 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk crypto_ram4 /dev/nbd11 00:31:39.341 /dev/nbd11 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd11 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd11 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd11 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd11 /proc/partitions 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd11 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:39.341 1+0 records in 00:31:39.341 1+0 records out 00:31:39.341 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000290804 s, 14.1 MB/s 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 4 )) 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:31:39.341 11:42:22 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:31:39.600 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:31:39.601 { 00:31:39.601 "nbd_device": "/dev/nbd0", 00:31:39.601 "bdev_name": "crypto_ram" 00:31:39.601 }, 00:31:39.601 { 00:31:39.601 "nbd_device": "/dev/nbd1", 00:31:39.601 "bdev_name": "crypto_ram2" 00:31:39.601 }, 00:31:39.601 { 00:31:39.601 "nbd_device": "/dev/nbd10", 00:31:39.601 "bdev_name": "crypto_ram3" 00:31:39.601 }, 00:31:39.601 { 00:31:39.601 "nbd_device": "/dev/nbd11", 00:31:39.601 "bdev_name": "crypto_ram4" 00:31:39.601 } 00:31:39.601 ]' 00:31:39.601 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[ 00:31:39.601 { 00:31:39.601 "nbd_device": "/dev/nbd0", 00:31:39.601 "bdev_name": "crypto_ram" 00:31:39.601 }, 00:31:39.601 { 00:31:39.601 "nbd_device": "/dev/nbd1", 00:31:39.601 "bdev_name": "crypto_ram2" 00:31:39.601 }, 00:31:39.601 { 00:31:39.601 "nbd_device": "/dev/nbd10", 00:31:39.601 "bdev_name": "crypto_ram3" 00:31:39.601 }, 00:31:39.601 { 00:31:39.601 "nbd_device": "/dev/nbd11", 00:31:39.601 "bdev_name": "crypto_ram4" 00:31:39.601 } 00:31:39.601 ]' 00:31:39.601 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:31:39.601 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:31:39.601 /dev/nbd1 00:31:39.601 /dev/nbd10 00:31:39.601 /dev/nbd11' 00:31:39.601 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:31:39.601 /dev/nbd1 00:31:39.601 /dev/nbd10 00:31:39.601 /dev/nbd11' 00:31:39.601 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:31:39.601 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=4 00:31:39.601 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 4 00:31:39.601 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@95 -- # count=4 00:31:39.601 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@96 -- # '[' 4 -ne 4 ']' 00:31:39.601 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1 /dev/nbd10 /dev/nbd11' write 00:31:39.601 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11') 00:31:39.601 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:31:39.601 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=write 00:31:39.601 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest 00:31:39.601 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:31:39.601 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest bs=4096 count=256 00:31:39.601 256+0 records in 00:31:39.601 256+0 records out 00:31:39.601 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0103424 s, 101 MB/s 00:31:39.601 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:31:39.860 256+0 records in 00:31:39.860 256+0 records out 00:31:39.860 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0524966 s, 20.0 MB/s 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:31:39.860 256+0 records in 00:31:39.860 256+0 records out 00:31:39.860 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0663658 s, 15.8 MB/s 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd10 bs=4096 count=256 oflag=direct 00:31:39.860 256+0 records in 00:31:39.860 256+0 records out 00:31:39.860 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0405516 s, 25.9 MB/s 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd11 bs=4096 count=256 oflag=direct 00:31:39.860 256+0 records in 00:31:39.860 256+0 records out 00:31:39.860 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0378119 s, 27.7 MB/s 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1 /dev/nbd10 /dev/nbd11' verify 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11') 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=verify 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd0 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd1 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd10 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:31:39.860 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd11 00:31:40.118 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest 00:31:40.118 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1 /dev/nbd10 /dev/nbd11' 00:31:40.118 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:31:40.118 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11') 00:31:40.118 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:31:40.118 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:31:40.118 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:40.118 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:31:40.118 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:31:40.118 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:31:40.118 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:31:40.376 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:40.376 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:40.376 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:31:40.376 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:31:40.376 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:31:40.376 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:40.376 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:31:40.635 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:31:40.635 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:31:40.635 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:31:40.635 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:40.635 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:40.635 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:31:40.635 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:31:40.635 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:31:40.635 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:40.635 11:42:23 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd10 00:31:40.894 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd10 00:31:40.894 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd10 00:31:40.894 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd10 00:31:40.894 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:40.894 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:40.894 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd10 /proc/partitions 00:31:40.894 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:31:40.894 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:31:40.894 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:40.894 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd11 00:31:41.153 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd11 00:31:41.153 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd11 00:31:41.153 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd11 00:31:41.153 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:41.153 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:41.153 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd11 /proc/partitions 00:31:41.153 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:31:41.153 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:31:41.153 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:31:41.153 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:31:41.153 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:31:41.411 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:31:41.411 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:31:41.411 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:31:41.411 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:31:41.411 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:31:41.411 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:31:41.411 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:31:41.411 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:31:41.411 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:31:41.411 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@104 -- # count=0 00:31:41.411 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:31:41.411 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@109 -- # return 0 00:31:41.411 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@324 -- # nbd_with_lvol_verify /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1 /dev/nbd10 /dev/nbd11' 00:31:41.411 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@131 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:31:41.411 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@132 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11') 00:31:41.411 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@132 -- # local nbd_list 00:31:41.411 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@133 -- # local mkfs_ret 00:31:41.411 11:42:24 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@135 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create -b malloc_lvol_verify 16 512 00:31:41.668 malloc_lvol_verify 00:31:41.668 11:42:25 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@136 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create_lvstore malloc_lvol_verify lvs 00:31:41.926 6b88359b-58a3-4462-b1f2-39c63f189b17 00:31:41.926 11:42:25 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@137 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create lvol 4 -l lvs 00:31:42.184 9764f7c4-0cce-4e08-b327-a08f6a07707a 00:31:42.184 11:42:25 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@138 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk lvs/lvol /dev/nbd0 00:31:42.443 /dev/nbd0 00:31:42.443 11:42:25 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@140 -- # mkfs.ext4 /dev/nbd0 00:31:42.443 mke2fs 1.46.5 (30-Dec-2021) 00:31:42.443 Discarding device blocks: 0/4096 done 00:31:42.443 Creating filesystem with 4096 1k blocks and 1024 inodes 00:31:42.443 00:31:42.443 Allocating group tables: 0/1 done 00:31:42.443 Writing inode tables: 0/1 done 00:31:42.443 Creating journal (1024 blocks): done 00:31:42.443 Writing superblocks and filesystem accounting information: 0/1 done 00:31:42.443 00:31:42.443 11:42:25 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@141 -- # mkfs_ret=0 00:31:42.443 11:42:25 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@142 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:31:42.443 11:42:25 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:31:42.443 11:42:25 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:31:42.443 11:42:25 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:31:42.443 11:42:25 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:31:42.443 11:42:25 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:42.443 11:42:25 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@143 -- # '[' 0 -ne 0 ']' 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- bdev/nbd_common.sh@147 -- # return 0 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@326 -- # killprocess 1044212 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@948 -- # '[' -z 1044212 ']' 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@952 -- # kill -0 1044212 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@953 -- # uname 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1044212 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1044212' 00:31:42.702 killing process with pid 1044212 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@967 -- # kill 1044212 00:31:42.702 11:42:26 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@972 -- # wait 1044212 00:31:43.270 11:42:26 blockdev_crypto_aesni.bdev_nbd -- bdev/blockdev.sh@327 -- # trap - SIGINT SIGTERM EXIT 00:31:43.270 00:31:43.270 real 0m10.452s 00:31:43.270 user 0m13.690s 00:31:43.270 sys 0m4.180s 00:31:43.270 11:42:26 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@1124 -- # xtrace_disable 00:31:43.270 11:42:26 blockdev_crypto_aesni.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:31:43.270 ************************************ 00:31:43.270 END TEST bdev_nbd 00:31:43.270 ************************************ 00:31:43.270 11:42:26 blockdev_crypto_aesni -- common/autotest_common.sh@1142 -- # return 0 00:31:43.270 11:42:26 blockdev_crypto_aesni -- bdev/blockdev.sh@763 -- # [[ y == y ]] 00:31:43.270 11:42:26 blockdev_crypto_aesni -- bdev/blockdev.sh@764 -- # '[' crypto_aesni = nvme ']' 00:31:43.270 11:42:26 blockdev_crypto_aesni -- bdev/blockdev.sh@764 -- # '[' crypto_aesni = gpt ']' 00:31:43.270 11:42:26 blockdev_crypto_aesni -- bdev/blockdev.sh@768 -- # run_test bdev_fio fio_test_suite '' 00:31:43.270 11:42:26 blockdev_crypto_aesni -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:31:43.270 11:42:26 blockdev_crypto_aesni -- common/autotest_common.sh@1105 -- # xtrace_disable 00:31:43.270 11:42:26 blockdev_crypto_aesni -- common/autotest_common.sh@10 -- # set +x 00:31:43.270 ************************************ 00:31:43.270 START TEST bdev_fio 00:31:43.270 ************************************ 00:31:43.270 11:42:26 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1123 -- # fio_test_suite '' 00:31:43.270 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@331 -- # local env_context 00:31:43.270 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@335 -- # pushd /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev 00:31:43.270 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev /var/jenkins/workspace/crypto-phy-autotest/spdk 00:31:43.270 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@336 -- # trap 'rm -f ./*.state; popd; exit 1' SIGINT SIGTERM EXIT 00:31:43.270 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@339 -- # echo '' 00:31:43.270 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@339 -- # sed s/--env-context=// 00:31:43.270 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@339 -- # env_context= 00:31:43.270 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@340 -- # fio_config_gen /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio verify AIO '' 00:31:43.270 11:42:26 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1280 -- # local config_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:31:43.270 11:42:26 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1281 -- # local workload=verify 00:31:43.270 11:42:26 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1282 -- # local bdev_type=AIO 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1283 -- # local env_context= 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1284 -- # local fio_dir=/usr/src/fio 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1286 -- # '[' -e /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio ']' 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1291 -- # '[' -z verify ']' 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -n '' ']' 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1299 -- # touch /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1301 -- # cat 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1313 -- # '[' verify == verify ']' 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1314 -- # cat 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1323 -- # '[' AIO == AIO ']' 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1324 -- # /usr/src/fio/fio --version 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1324 -- # [[ fio-3.35 == *\f\i\o\-\3* ]] 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1325 -- # echo serialize_overlap=1 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_crypto_ram]' 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=crypto_ram 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_crypto_ram2]' 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=crypto_ram2 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_crypto_ram3]' 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=crypto_ram3 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_crypto_ram4]' 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=crypto_ram4 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@347 -- # local 'fio_params=--ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json' 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@349 -- # run_test bdev_fio_rw_verify fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1099 -- # '[' 11 -le 1 ']' 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1105 -- # xtrace_disable 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:31:43.271 ************************************ 00:31:43.271 START TEST bdev_fio_rw_verify 00:31:43.271 ************************************ 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1123 -- # fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1339 -- # local sanitizers 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1341 -- # shift 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # local asan_lib= 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # grep libasan 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # asan_lib= 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # asan_lib= 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev' 00:31:43.271 11:42:26 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:31:43.837 job_crypto_ram: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:31:43.837 job_crypto_ram2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:31:43.837 job_crypto_ram3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:31:43.837 job_crypto_ram4: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:31:43.837 fio-3.35 00:31:43.837 Starting 4 threads 00:31:58.709 00:31:58.709 job_crypto_ram: (groupid=0, jobs=4): err= 0: pid=1046254: Mon Jul 15 11:42:39 2024 00:31:58.709 read: IOPS=18.3k, BW=71.5MiB/s (75.0MB/s)(715MiB/10001msec) 00:31:58.709 slat (usec): min=17, max=1310, avg=72.95, stdev=37.75 00:31:58.709 clat (usec): min=16, max=2183, avg=388.54, stdev=233.51 00:31:58.709 lat (usec): min=60, max=2227, avg=461.49, stdev=252.77 00:31:58.709 clat percentiles (usec): 00:31:58.709 | 50.000th=[ 334], 99.000th=[ 1074], 99.900th=[ 1254], 99.990th=[ 1401], 00:31:58.709 | 99.999th=[ 2073] 00:31:58.709 write: IOPS=20.2k, BW=79.0MiB/s (82.9MB/s)(771MiB/9757msec); 0 zone resets 00:31:58.709 slat (usec): min=23, max=509, avg=87.96, stdev=37.56 00:31:58.709 clat (usec): min=29, max=1738, avg=472.25, stdev=273.30 00:31:58.709 lat (usec): min=93, max=1908, avg=560.21, stdev=292.07 00:31:58.709 clat percentiles (usec): 00:31:58.709 | 50.000th=[ 420], 99.000th=[ 1303], 99.900th=[ 1532], 99.990th=[ 1631], 00:31:58.709 | 99.999th=[ 1696] 00:31:58.709 bw ( KiB/s): min=62912, max=108389, per=97.73%, avg=79073.84, stdev=2968.64, samples=76 00:31:58.709 iops : min=15728, max=27097, avg=19768.21, stdev=742.13, samples=76 00:31:58.709 lat (usec) : 20=0.01%, 50=0.01%, 100=4.31%, 250=23.28%, 500=39.57% 00:31:58.709 lat (usec) : 750=20.03%, 1000=9.48% 00:31:58.709 lat (msec) : 2=3.33%, 4=0.01% 00:31:58.709 cpu : usr=99.54%, sys=0.00%, ctx=72, majf=0, minf=267 00:31:58.709 IO depths : 1=9.6%, 2=25.7%, 4=51.4%, 8=13.2%, 16=0.0%, 32=0.0%, >=64=0.0% 00:31:58.709 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:31:58.709 complete : 0=0.0%, 4=88.6%, 8=11.4%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:31:58.709 issued rwts: total=183157,197362,0,0 short=0,0,0,0 dropped=0,0,0,0 00:31:58.709 latency : target=0, window=0, percentile=100.00%, depth=8 00:31:58.709 00:31:58.709 Run status group 0 (all jobs): 00:31:58.709 READ: bw=71.5MiB/s (75.0MB/s), 71.5MiB/s-71.5MiB/s (75.0MB/s-75.0MB/s), io=715MiB (750MB), run=10001-10001msec 00:31:58.709 WRITE: bw=79.0MiB/s (82.9MB/s), 79.0MiB/s-79.0MiB/s (82.9MB/s-82.9MB/s), io=771MiB (808MB), run=9757-9757msec 00:31:58.709 00:31:58.709 real 0m13.513s 00:31:58.709 user 0m45.489s 00:31:58.709 sys 0m0.492s 00:31:58.709 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1124 -- # xtrace_disable 00:31:58.709 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@10 -- # set +x 00:31:58.709 ************************************ 00:31:58.709 END TEST bdev_fio_rw_verify 00:31:58.709 ************************************ 00:31:58.709 11:42:40 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1142 -- # return 0 00:31:58.709 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@350 -- # rm -f 00:31:58.709 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@351 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:31:58.709 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@354 -- # fio_config_gen /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio trim '' '' 00:31:58.709 11:42:40 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1280 -- # local config_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:31:58.709 11:42:40 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1281 -- # local workload=trim 00:31:58.709 11:42:40 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1282 -- # local bdev_type= 00:31:58.709 11:42:40 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1283 -- # local env_context= 00:31:58.709 11:42:40 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1284 -- # local fio_dir=/usr/src/fio 00:31:58.709 11:42:40 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1286 -- # '[' -e /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio ']' 00:31:58.709 11:42:40 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1291 -- # '[' -z trim ']' 00:31:58.709 11:42:40 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -n '' ']' 00:31:58.709 11:42:40 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1299 -- # touch /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:31:58.709 11:42:40 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1301 -- # cat 00:31:58.709 11:42:40 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1313 -- # '[' trim == verify ']' 00:31:58.709 11:42:40 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1328 -- # '[' trim == trim ']' 00:31:58.709 11:42:40 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1329 -- # echo rw=trimwrite 00:31:58.709 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@355 -- # printf '%s\n' '{' ' "name": "crypto_ram",' ' "aliases": [' ' "009cf00b-f018-57da-adb8-e3fc376b9f78"' ' ],' ' "product_name": "crypto",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "009cf00b-f018-57da-adb8-e3fc376b9f78",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc0",' ' "name": "crypto_ram",' ' "key_name": "test_dek_aesni_cbc_1"' ' }' ' }' '}' '{' ' "name": "crypto_ram2",' ' "aliases": [' ' "e010df0d-beb5-5585-a68c-53d917a2a68c"' ' ],' ' "product_name": "crypto",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "e010df0d-beb5-5585-a68c-53d917a2a68c",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc1",' ' "name": "crypto_ram2",' ' "key_name": "test_dek_aesni_cbc_2"' ' }' ' }' '}' '{' ' "name": "crypto_ram3",' ' "aliases": [' ' "a2ff8065-d168-5d41-b165-63ed344581fd"' ' ],' ' "product_name": "crypto",' ' "block_size": 4096,' ' "num_blocks": 8192,' ' "uuid": "a2ff8065-d168-5d41-b165-63ed344581fd",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc2",' ' "name": "crypto_ram3",' ' "key_name": "test_dek_aesni_cbc_3"' ' }' ' }' '}' '{' ' "name": "crypto_ram4",' ' "aliases": [' ' "d3452518-0bbe-55c3-86ce-84884cd1f575"' ' ],' ' "product_name": "crypto",' ' "block_size": 4096,' ' "num_blocks": 8192,' ' "uuid": "d3452518-0bbe-55c3-86ce-84884cd1f575",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc3",' ' "name": "crypto_ram4",' ' "key_name": "test_dek_aesni_cbc_4"' ' }' ' }' '}' 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@355 -- # jq -r 'select(.supported_io_types.unmap == true) | .name' 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@355 -- # [[ -n crypto_ram 00:31:58.710 crypto_ram2 00:31:58.710 crypto_ram3 00:31:58.710 crypto_ram4 ]] 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@356 -- # jq -r 'select(.supported_io_types.unmap == true) | .name' 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@356 -- # printf '%s\n' '{' ' "name": "crypto_ram",' ' "aliases": [' ' "009cf00b-f018-57da-adb8-e3fc376b9f78"' ' ],' ' "product_name": "crypto",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "009cf00b-f018-57da-adb8-e3fc376b9f78",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc0",' ' "name": "crypto_ram",' ' "key_name": "test_dek_aesni_cbc_1"' ' }' ' }' '}' '{' ' "name": "crypto_ram2",' ' "aliases": [' ' "e010df0d-beb5-5585-a68c-53d917a2a68c"' ' ],' ' "product_name": "crypto",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "e010df0d-beb5-5585-a68c-53d917a2a68c",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc1",' ' "name": "crypto_ram2",' ' "key_name": "test_dek_aesni_cbc_2"' ' }' ' }' '}' '{' ' "name": "crypto_ram3",' ' "aliases": [' ' "a2ff8065-d168-5d41-b165-63ed344581fd"' ' ],' ' "product_name": "crypto",' ' "block_size": 4096,' ' "num_blocks": 8192,' ' "uuid": "a2ff8065-d168-5d41-b165-63ed344581fd",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc2",' ' "name": "crypto_ram3",' ' "key_name": "test_dek_aesni_cbc_3"' ' }' ' }' '}' '{' ' "name": "crypto_ram4",' ' "aliases": [' ' "d3452518-0bbe-55c3-86ce-84884cd1f575"' ' ],' ' "product_name": "crypto",' ' "block_size": 4096,' ' "num_blocks": 8192,' ' "uuid": "d3452518-0bbe-55c3-86ce-84884cd1f575",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc3",' ' "name": "crypto_ram4",' ' "key_name": "test_dek_aesni_cbc_4"' ' }' ' }' '}' 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_crypto_ram]' 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=crypto_ram 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_crypto_ram2]' 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=crypto_ram2 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_crypto_ram3]' 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=crypto_ram3 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_crypto_ram4]' 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=crypto_ram4 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@367 -- # run_test bdev_fio_trim fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --verify_state_save=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1099 -- # '[' 11 -le 1 ']' 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1105 -- # xtrace_disable 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:31:58.710 ************************************ 00:31:58.710 START TEST bdev_fio_trim 00:31:58.710 ************************************ 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1123 -- # fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --verify_state_save=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --verify_state_save=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1339 -- # local sanitizers 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1341 -- # shift 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1343 -- # local asan_lib= 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # grep libasan 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # asan_lib= 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # asan_lib= 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev' 00:31:58.710 11:42:40 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --verify_state_save=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:31:58.710 job_crypto_ram: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:31:58.710 job_crypto_ram2: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:31:58.710 job_crypto_ram3: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:31:58.710 job_crypto_ram4: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:31:58.710 fio-3.35 00:31:58.710 Starting 4 threads 00:32:10.987 00:32:10.987 job_crypto_ram: (groupid=0, jobs=4): err= 0: pid=1048103: Mon Jul 15 11:42:53 2024 00:32:10.987 write: IOPS=35.5k, BW=139MiB/s (146MB/s)(1388MiB/10001msec); 0 zone resets 00:32:10.987 slat (usec): min=17, max=439, avg=63.43, stdev=34.02 00:32:10.987 clat (usec): min=53, max=1844, avg=288.20, stdev=184.10 00:32:10.987 lat (usec): min=70, max=1881, avg=351.64, stdev=207.38 00:32:10.987 clat percentiles (usec): 00:32:10.987 | 50.000th=[ 241], 99.000th=[ 922], 99.900th=[ 1045], 99.990th=[ 1090], 00:32:10.987 | 99.999th=[ 1303] 00:32:10.987 bw ( KiB/s): min=130352, max=205704, per=100.00%, avg=142826.79, stdev=6202.80, samples=76 00:32:10.987 iops : min=32588, max=51426, avg=35706.68, stdev=1550.69, samples=76 00:32:10.987 trim: IOPS=35.5k, BW=139MiB/s (146MB/s)(1388MiB/10001msec); 0 zone resets 00:32:10.987 slat (usec): min=6, max=1409, avg=17.69, stdev= 7.51 00:32:10.987 clat (usec): min=65, max=1881, avg=271.57, stdev=120.98 00:32:10.987 lat (usec): min=74, max=1895, avg=289.26, stdev=123.47 00:32:10.987 clat percentiles (usec): 00:32:10.987 | 50.000th=[ 251], 99.000th=[ 635], 99.900th=[ 709], 99.990th=[ 758], 00:32:10.987 | 99.999th=[ 955] 00:32:10.987 bw ( KiB/s): min=130344, max=205704, per=100.00%, avg=142828.05, stdev=6203.36, samples=76 00:32:10.987 iops : min=32586, max=51426, avg=35707.00, stdev=1550.83, samples=76 00:32:10.987 lat (usec) : 100=5.28%, 250=45.70%, 500=40.34%, 750=6.68%, 1000=1.85% 00:32:10.987 lat (msec) : 2=0.16% 00:32:10.987 cpu : usr=99.58%, sys=0.01%, ctx=68, majf=0, minf=119 00:32:10.987 IO depths : 1=7.9%, 2=26.3%, 4=52.6%, 8=13.2%, 16=0.0%, 32=0.0%, >=64=0.0% 00:32:10.987 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:32:10.987 complete : 0=0.0%, 4=88.4%, 8=11.6%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:32:10.987 issued rwts: total=0,355389,355389,0 short=0,0,0,0 dropped=0,0,0,0 00:32:10.987 latency : target=0, window=0, percentile=100.00%, depth=8 00:32:10.987 00:32:10.987 Run status group 0 (all jobs): 00:32:10.987 WRITE: bw=139MiB/s (146MB/s), 139MiB/s-139MiB/s (146MB/s-146MB/s), io=1388MiB (1456MB), run=10001-10001msec 00:32:10.987 TRIM: bw=139MiB/s (146MB/s), 139MiB/s-139MiB/s (146MB/s-146MB/s), io=1388MiB (1456MB), run=10001-10001msec 00:32:10.987 00:32:10.987 real 0m13.564s 00:32:10.987 user 0m45.735s 00:32:10.987 sys 0m0.504s 00:32:10.987 11:42:54 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1124 -- # xtrace_disable 00:32:10.987 11:42:54 blockdev_crypto_aesni.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@10 -- # set +x 00:32:10.987 ************************************ 00:32:10.987 END TEST bdev_fio_trim 00:32:10.987 ************************************ 00:32:10.987 11:42:54 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1142 -- # return 0 00:32:10.987 11:42:54 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@368 -- # rm -f 00:32:10.987 11:42:54 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@369 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:32:10.987 11:42:54 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@370 -- # popd 00:32:10.987 /var/jenkins/workspace/crypto-phy-autotest/spdk 00:32:10.987 11:42:54 blockdev_crypto_aesni.bdev_fio -- bdev/blockdev.sh@371 -- # trap - SIGINT SIGTERM EXIT 00:32:10.987 00:32:10.987 real 0m27.443s 00:32:10.987 user 1m31.413s 00:32:10.987 sys 0m1.197s 00:32:10.987 11:42:54 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@1124 -- # xtrace_disable 00:32:10.987 11:42:54 blockdev_crypto_aesni.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:32:10.987 ************************************ 00:32:10.987 END TEST bdev_fio 00:32:10.987 ************************************ 00:32:10.987 11:42:54 blockdev_crypto_aesni -- common/autotest_common.sh@1142 -- # return 0 00:32:10.987 11:42:54 blockdev_crypto_aesni -- bdev/blockdev.sh@775 -- # trap cleanup SIGINT SIGTERM EXIT 00:32:10.987 11:42:54 blockdev_crypto_aesni -- bdev/blockdev.sh@777 -- # run_test bdev_verify /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:32:10.987 11:42:54 blockdev_crypto_aesni -- common/autotest_common.sh@1099 -- # '[' 16 -le 1 ']' 00:32:10.987 11:42:54 blockdev_crypto_aesni -- common/autotest_common.sh@1105 -- # xtrace_disable 00:32:10.987 11:42:54 blockdev_crypto_aesni -- common/autotest_common.sh@10 -- # set +x 00:32:10.987 ************************************ 00:32:10.987 START TEST bdev_verify 00:32:10.987 ************************************ 00:32:10.987 11:42:54 blockdev_crypto_aesni.bdev_verify -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:32:10.987 [2024-07-15 11:42:54.227872] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:32:10.987 [2024-07-15 11:42:54.227943] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1049523 ] 00:32:10.987 [2024-07-15 11:42:54.356696] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:32:10.987 [2024-07-15 11:42:54.454600] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:32:10.987 [2024-07-15 11:42:54.454605] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:32:10.987 [2024-07-15 11:42:54.475999] accel_dpdk_cryptodev.c: 223:accel_dpdk_cryptodev_set_driver: *NOTICE*: Using driver crypto_aesni_mb 00:32:10.987 [2024-07-15 11:42:54.484032] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation encrypt will be assigned to module dpdk_cryptodev 00:32:10.987 [2024-07-15 11:42:54.492059] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation decrypt will be assigned to module dpdk_cryptodev 00:32:11.244 [2024-07-15 11:42:54.604984] accel_dpdk_cryptodev.c:1178:accel_dpdk_cryptodev_init: *NOTICE*: Found crypto devices: 97 00:32:13.770 [2024-07-15 11:42:56.822138] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_1" 00:32:13.770 [2024-07-15 11:42:56.822224] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:32:13.770 [2024-07-15 11:42:56.822239] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:13.770 [2024-07-15 11:42:56.830153] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_2" 00:32:13.770 [2024-07-15 11:42:56.830175] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:32:13.770 [2024-07-15 11:42:56.830187] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:13.770 [2024-07-15 11:42:56.838177] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_3" 00:32:13.770 [2024-07-15 11:42:56.838195] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:32:13.770 [2024-07-15 11:42:56.838207] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:13.770 [2024-07-15 11:42:56.846199] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_4" 00:32:13.770 [2024-07-15 11:42:56.846217] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc3 00:32:13.770 [2024-07-15 11:42:56.846229] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:13.770 Running I/O for 5 seconds... 00:32:19.031 00:32:19.031 Latency(us) 00:32:19.031 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:19.031 Job: crypto_ram (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:32:19.031 Verification LBA range: start 0x0 length 0x1000 00:32:19.031 crypto_ram : 5.07 485.77 1.90 0.00 0.00 262167.20 1567.17 186008.26 00:32:19.031 Job: crypto_ram (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:32:19.031 Verification LBA range: start 0x1000 length 0x1000 00:32:19.031 crypto_ram : 5.07 491.74 1.92 0.00 0.00 258957.08 2322.25 185096.46 00:32:19.031 Job: crypto_ram2 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:32:19.031 Verification LBA range: start 0x0 length 0x1000 00:32:19.031 crypto_ram2 : 5.07 487.22 1.90 0.00 0.00 260880.18 2023.07 174154.80 00:32:19.031 Job: crypto_ram2 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:32:19.031 Verification LBA range: start 0x1000 length 0x1000 00:32:19.031 crypto_ram2 : 5.08 493.19 1.93 0.00 0.00 257696.34 2877.89 174154.80 00:32:19.031 Job: crypto_ram3 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:32:19.031 Verification LBA range: start 0x0 length 0x1000 00:32:19.031 crypto_ram3 : 5.05 3809.56 14.88 0.00 0.00 33260.17 3675.71 28949.82 00:32:19.031 Job: crypto_ram3 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:32:19.031 Verification LBA range: start 0x1000 length 0x1000 00:32:19.031 crypto_ram3 : 5.05 3834.88 14.98 0.00 0.00 33039.49 3675.71 28835.84 00:32:19.031 Job: crypto_ram4 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:32:19.031 Verification LBA range: start 0x0 length 0x1000 00:32:19.031 crypto_ram4 : 5.06 3818.98 14.92 0.00 0.00 33127.81 2535.96 28835.84 00:32:19.031 Job: crypto_ram4 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:32:19.031 Verification LBA range: start 0x1000 length 0x1000 00:32:19.031 crypto_ram4 : 5.06 3843.90 15.02 0.00 0.00 32910.59 3048.85 28607.89 00:32:19.031 =================================================================================================================== 00:32:19.031 Total : 17265.22 67.44 0.00 0.00 58873.02 1567.17 186008.26 00:32:19.031 00:32:19.031 real 0m8.272s 00:32:19.031 user 0m15.673s 00:32:19.031 sys 0m0.382s 00:32:19.031 11:43:02 blockdev_crypto_aesni.bdev_verify -- common/autotest_common.sh@1124 -- # xtrace_disable 00:32:19.031 11:43:02 blockdev_crypto_aesni.bdev_verify -- common/autotest_common.sh@10 -- # set +x 00:32:19.031 ************************************ 00:32:19.031 END TEST bdev_verify 00:32:19.031 ************************************ 00:32:19.031 11:43:02 blockdev_crypto_aesni -- common/autotest_common.sh@1142 -- # return 0 00:32:19.032 11:43:02 blockdev_crypto_aesni -- bdev/blockdev.sh@778 -- # run_test bdev_verify_big_io /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:32:19.032 11:43:02 blockdev_crypto_aesni -- common/autotest_common.sh@1099 -- # '[' 16 -le 1 ']' 00:32:19.032 11:43:02 blockdev_crypto_aesni -- common/autotest_common.sh@1105 -- # xtrace_disable 00:32:19.032 11:43:02 blockdev_crypto_aesni -- common/autotest_common.sh@10 -- # set +x 00:32:19.032 ************************************ 00:32:19.032 START TEST bdev_verify_big_io 00:32:19.032 ************************************ 00:32:19.032 11:43:02 blockdev_crypto_aesni.bdev_verify_big_io -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:32:19.032 [2024-07-15 11:43:02.577353] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:32:19.032 [2024-07-15 11:43:02.577412] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1050587 ] 00:32:19.289 [2024-07-15 11:43:02.705989] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:32:19.289 [2024-07-15 11:43:02.807601] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:32:19.289 [2024-07-15 11:43:02.807606] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:32:19.289 [2024-07-15 11:43:02.828988] accel_dpdk_cryptodev.c: 223:accel_dpdk_cryptodev_set_driver: *NOTICE*: Using driver crypto_aesni_mb 00:32:19.289 [2024-07-15 11:43:02.837017] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation encrypt will be assigned to module dpdk_cryptodev 00:32:19.289 [2024-07-15 11:43:02.845045] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation decrypt will be assigned to module dpdk_cryptodev 00:32:19.547 [2024-07-15 11:43:02.948689] accel_dpdk_cryptodev.c:1178:accel_dpdk_cryptodev_init: *NOTICE*: Found crypto devices: 97 00:32:22.075 [2024-07-15 11:43:05.172816] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_1" 00:32:22.075 [2024-07-15 11:43:05.172898] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:32:22.075 [2024-07-15 11:43:05.172913] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:22.075 [2024-07-15 11:43:05.180833] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_2" 00:32:22.075 [2024-07-15 11:43:05.180856] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:32:22.075 [2024-07-15 11:43:05.180868] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:22.075 [2024-07-15 11:43:05.188857] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_3" 00:32:22.075 [2024-07-15 11:43:05.188876] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:32:22.075 [2024-07-15 11:43:05.188888] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:22.075 [2024-07-15 11:43:05.196879] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_4" 00:32:22.075 [2024-07-15 11:43:05.196899] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc3 00:32:22.075 [2024-07-15 11:43:05.196910] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:22.075 Running I/O for 5 seconds... 00:32:28.628 00:32:28.628 Latency(us) 00:32:28.628 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:28.628 Job: crypto_ram (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:32:28.628 Verification LBA range: start 0x0 length 0x100 00:32:28.628 crypto_ram : 5.89 43.49 2.72 0.00 0.00 2855285.98 76135.74 2728121.21 00:32:28.628 Job: crypto_ram (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:32:28.628 Verification LBA range: start 0x100 length 0x100 00:32:28.628 crypto_ram : 5.88 43.53 2.72 0.00 0.00 2853065.24 78871.15 2713532.33 00:32:28.628 Job: crypto_ram2 (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:32:28.628 Verification LBA range: start 0x0 length 0x100 00:32:28.628 crypto_ram2 : 5.89 43.49 2.72 0.00 0.00 2749609.18 75679.83 2728121.21 00:32:28.628 Job: crypto_ram2 (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:32:28.628 Verification LBA range: start 0x100 length 0x100 00:32:28.628 crypto_ram2 : 5.88 43.52 2.72 0.00 0.00 2747489.95 77959.35 2713532.33 00:32:28.628 Job: crypto_ram3 (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:32:28.628 Verification LBA range: start 0x0 length 0x100 00:32:28.628 crypto_ram3 : 5.58 266.23 16.64 0.00 0.00 428429.18 37156.06 598144.22 00:32:28.628 Job: crypto_ram3 (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:32:28.628 Verification LBA range: start 0x100 length 0x100 00:32:28.628 crypto_ram3 : 5.59 267.25 16.70 0.00 0.00 426510.70 49465.43 590849.78 00:32:28.628 Job: crypto_ram4 (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:32:28.628 Verification LBA range: start 0x0 length 0x100 00:32:28.628 crypto_ram4 : 5.71 284.14 17.76 0.00 0.00 389480.31 17438.27 554377.57 00:32:28.628 Job: crypto_ram4 (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:32:28.628 Verification LBA range: start 0x100 length 0x100 00:32:28.628 crypto_ram4 : 5.71 285.24 17.83 0.00 0.00 388007.67 18578.03 550730.35 00:32:28.628 =================================================================================================================== 00:32:28.628 Total : 1276.89 79.81 0.00 0.00 745224.92 17438.27 2728121.21 00:32:28.628 00:32:28.628 real 0m9.107s 00:32:28.628 user 0m17.313s 00:32:28.628 sys 0m0.409s 00:32:28.628 11:43:11 blockdev_crypto_aesni.bdev_verify_big_io -- common/autotest_common.sh@1124 -- # xtrace_disable 00:32:28.628 11:43:11 blockdev_crypto_aesni.bdev_verify_big_io -- common/autotest_common.sh@10 -- # set +x 00:32:28.628 ************************************ 00:32:28.628 END TEST bdev_verify_big_io 00:32:28.629 ************************************ 00:32:28.629 11:43:11 blockdev_crypto_aesni -- common/autotest_common.sh@1142 -- # return 0 00:32:28.629 11:43:11 blockdev_crypto_aesni -- bdev/blockdev.sh@779 -- # run_test bdev_write_zeroes /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:32:28.629 11:43:11 blockdev_crypto_aesni -- common/autotest_common.sh@1099 -- # '[' 13 -le 1 ']' 00:32:28.629 11:43:11 blockdev_crypto_aesni -- common/autotest_common.sh@1105 -- # xtrace_disable 00:32:28.629 11:43:11 blockdev_crypto_aesni -- common/autotest_common.sh@10 -- # set +x 00:32:28.629 ************************************ 00:32:28.629 START TEST bdev_write_zeroes 00:32:28.629 ************************************ 00:32:28.629 11:43:11 blockdev_crypto_aesni.bdev_write_zeroes -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:32:28.629 [2024-07-15 11:43:11.773740] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:32:28.629 [2024-07-15 11:43:11.773810] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1051797 ] 00:32:28.629 [2024-07-15 11:43:11.903412] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:28.629 [2024-07-15 11:43:12.007605] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:32:28.629 [2024-07-15 11:43:12.028903] accel_dpdk_cryptodev.c: 223:accel_dpdk_cryptodev_set_driver: *NOTICE*: Using driver crypto_aesni_mb 00:32:28.629 [2024-07-15 11:43:12.036935] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation encrypt will be assigned to module dpdk_cryptodev 00:32:28.629 [2024-07-15 11:43:12.044952] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation decrypt will be assigned to module dpdk_cryptodev 00:32:28.629 [2024-07-15 11:43:12.156061] accel_dpdk_cryptodev.c:1178:accel_dpdk_cryptodev_init: *NOTICE*: Found crypto devices: 97 00:32:31.152 [2024-07-15 11:43:14.385178] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_1" 00:32:31.152 [2024-07-15 11:43:14.385251] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:32:31.152 [2024-07-15 11:43:14.385267] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:31.152 [2024-07-15 11:43:14.393195] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_2" 00:32:31.152 [2024-07-15 11:43:14.393216] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:32:31.152 [2024-07-15 11:43:14.393228] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:31.152 [2024-07-15 11:43:14.401216] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_3" 00:32:31.152 [2024-07-15 11:43:14.401234] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:32:31.152 [2024-07-15 11:43:14.401246] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:31.152 [2024-07-15 11:43:14.409235] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_aesni_cbc_4" 00:32:31.152 [2024-07-15 11:43:14.409253] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc3 00:32:31.152 [2024-07-15 11:43:14.409265] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:31.152 Running I/O for 1 seconds... 00:32:32.087 00:32:32.087 Latency(us) 00:32:32.087 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:32.087 Job: crypto_ram (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:32:32.087 crypto_ram : 1.03 1972.21 7.70 0.00 0.00 64367.95 5442.34 77503.44 00:32:32.087 Job: crypto_ram2 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:32:32.087 crypto_ram2 : 1.03 1977.94 7.73 0.00 0.00 63837.06 5413.84 72032.61 00:32:32.087 Job: crypto_ram3 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:32:32.087 crypto_ram3 : 1.02 15170.70 59.26 0.00 0.00 8311.49 2464.72 10770.70 00:32:32.087 Job: crypto_ram4 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:32:32.087 crypto_ram4 : 1.02 15207.83 59.41 0.00 0.00 8265.78 2464.72 8719.14 00:32:32.087 =================================================================================================================== 00:32:32.087 Total : 34328.68 134.10 0.00 0.00 14738.63 2464.72 77503.44 00:32:32.345 00:32:32.345 real 0m4.208s 00:32:32.345 user 0m3.786s 00:32:32.345 sys 0m0.381s 00:32:32.345 11:43:15 blockdev_crypto_aesni.bdev_write_zeroes -- common/autotest_common.sh@1124 -- # xtrace_disable 00:32:32.345 11:43:15 blockdev_crypto_aesni.bdev_write_zeroes -- common/autotest_common.sh@10 -- # set +x 00:32:32.345 ************************************ 00:32:32.345 END TEST bdev_write_zeroes 00:32:32.345 ************************************ 00:32:32.602 11:43:15 blockdev_crypto_aesni -- common/autotest_common.sh@1142 -- # return 0 00:32:32.602 11:43:15 blockdev_crypto_aesni -- bdev/blockdev.sh@782 -- # run_test bdev_json_nonenclosed /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:32:32.602 11:43:15 blockdev_crypto_aesni -- common/autotest_common.sh@1099 -- # '[' 13 -le 1 ']' 00:32:32.602 11:43:15 blockdev_crypto_aesni -- common/autotest_common.sh@1105 -- # xtrace_disable 00:32:32.602 11:43:15 blockdev_crypto_aesni -- common/autotest_common.sh@10 -- # set +x 00:32:32.602 ************************************ 00:32:32.602 START TEST bdev_json_nonenclosed 00:32:32.602 ************************************ 00:32:32.602 11:43:16 blockdev_crypto_aesni.bdev_json_nonenclosed -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:32:32.602 [2024-07-15 11:43:16.063663] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:32:32.602 [2024-07-15 11:43:16.063723] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1052356 ] 00:32:32.602 [2024-07-15 11:43:16.189488] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:32.860 [2024-07-15 11:43:16.291075] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:32:32.860 [2024-07-15 11:43:16.291147] json_config.c: 608:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: not enclosed in {}. 00:32:32.860 [2024-07-15 11:43:16.291168] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:32:32.860 [2024-07-15 11:43:16.291181] app.c:1053:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:32:32.860 00:32:32.860 real 0m0.393s 00:32:32.860 user 0m0.231s 00:32:32.860 sys 0m0.160s 00:32:32.860 11:43:16 blockdev_crypto_aesni.bdev_json_nonenclosed -- common/autotest_common.sh@1123 -- # es=234 00:32:32.860 11:43:16 blockdev_crypto_aesni.bdev_json_nonenclosed -- common/autotest_common.sh@1124 -- # xtrace_disable 00:32:32.860 11:43:16 blockdev_crypto_aesni.bdev_json_nonenclosed -- common/autotest_common.sh@10 -- # set +x 00:32:32.860 ************************************ 00:32:32.860 END TEST bdev_json_nonenclosed 00:32:32.860 ************************************ 00:32:32.860 11:43:16 blockdev_crypto_aesni -- common/autotest_common.sh@1142 -- # return 234 00:32:32.860 11:43:16 blockdev_crypto_aesni -- bdev/blockdev.sh@782 -- # true 00:32:32.860 11:43:16 blockdev_crypto_aesni -- bdev/blockdev.sh@785 -- # run_test bdev_json_nonarray /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:32:32.860 11:43:16 blockdev_crypto_aesni -- common/autotest_common.sh@1099 -- # '[' 13 -le 1 ']' 00:32:32.860 11:43:16 blockdev_crypto_aesni -- common/autotest_common.sh@1105 -- # xtrace_disable 00:32:32.860 11:43:16 blockdev_crypto_aesni -- common/autotest_common.sh@10 -- # set +x 00:32:33.120 ************************************ 00:32:33.120 START TEST bdev_json_nonarray 00:32:33.120 ************************************ 00:32:33.120 11:43:16 blockdev_crypto_aesni.bdev_json_nonarray -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:32:33.120 [2024-07-15 11:43:16.535275] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:32:33.120 [2024-07-15 11:43:16.535334] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1052379 ] 00:32:33.120 [2024-07-15 11:43:16.662717] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:33.407 [2024-07-15 11:43:16.764724] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:32:33.407 [2024-07-15 11:43:16.764798] json_config.c: 614:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: 'subsystems' should be an array. 00:32:33.407 [2024-07-15 11:43:16.764819] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:32:33.407 [2024-07-15 11:43:16.764832] app.c:1053:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:32:33.407 00:32:33.407 real 0m0.399s 00:32:33.407 user 0m0.237s 00:32:33.407 sys 0m0.159s 00:32:33.407 11:43:16 blockdev_crypto_aesni.bdev_json_nonarray -- common/autotest_common.sh@1123 -- # es=234 00:32:33.407 11:43:16 blockdev_crypto_aesni.bdev_json_nonarray -- common/autotest_common.sh@1124 -- # xtrace_disable 00:32:33.407 11:43:16 blockdev_crypto_aesni.bdev_json_nonarray -- common/autotest_common.sh@10 -- # set +x 00:32:33.407 ************************************ 00:32:33.407 END TEST bdev_json_nonarray 00:32:33.407 ************************************ 00:32:33.407 11:43:16 blockdev_crypto_aesni -- common/autotest_common.sh@1142 -- # return 234 00:32:33.407 11:43:16 blockdev_crypto_aesni -- bdev/blockdev.sh@785 -- # true 00:32:33.407 11:43:16 blockdev_crypto_aesni -- bdev/blockdev.sh@787 -- # [[ crypto_aesni == bdev ]] 00:32:33.407 11:43:16 blockdev_crypto_aesni -- bdev/blockdev.sh@794 -- # [[ crypto_aesni == gpt ]] 00:32:33.407 11:43:16 blockdev_crypto_aesni -- bdev/blockdev.sh@798 -- # [[ crypto_aesni == crypto_sw ]] 00:32:33.407 11:43:16 blockdev_crypto_aesni -- bdev/blockdev.sh@810 -- # trap - SIGINT SIGTERM EXIT 00:32:33.407 11:43:16 blockdev_crypto_aesni -- bdev/blockdev.sh@811 -- # cleanup 00:32:33.407 11:43:16 blockdev_crypto_aesni -- bdev/blockdev.sh@23 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/aiofile 00:32:33.407 11:43:16 blockdev_crypto_aesni -- bdev/blockdev.sh@24 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json 00:32:33.407 11:43:16 blockdev_crypto_aesni -- bdev/blockdev.sh@26 -- # [[ crypto_aesni == rbd ]] 00:32:33.407 11:43:16 blockdev_crypto_aesni -- bdev/blockdev.sh@30 -- # [[ crypto_aesni == daos ]] 00:32:33.407 11:43:16 blockdev_crypto_aesni -- bdev/blockdev.sh@34 -- # [[ crypto_aesni = \g\p\t ]] 00:32:33.407 11:43:16 blockdev_crypto_aesni -- bdev/blockdev.sh@40 -- # [[ crypto_aesni == xnvme ]] 00:32:33.407 00:32:33.407 real 1m12.178s 00:32:33.407 user 2m39.692s 00:32:33.407 sys 0m9.056s 00:32:33.407 11:43:16 blockdev_crypto_aesni -- common/autotest_common.sh@1124 -- # xtrace_disable 00:32:33.407 11:43:16 blockdev_crypto_aesni -- common/autotest_common.sh@10 -- # set +x 00:32:33.407 ************************************ 00:32:33.407 END TEST blockdev_crypto_aesni 00:32:33.407 ************************************ 00:32:33.407 11:43:16 -- common/autotest_common.sh@1142 -- # return 0 00:32:33.407 11:43:16 -- spdk/autotest.sh@358 -- # run_test blockdev_crypto_sw /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/blockdev.sh crypto_sw 00:32:33.407 11:43:16 -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:32:33.407 11:43:16 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:32:33.407 11:43:16 -- common/autotest_common.sh@10 -- # set +x 00:32:33.665 ************************************ 00:32:33.665 START TEST blockdev_crypto_sw 00:32:33.665 ************************************ 00:32:33.665 11:43:17 blockdev_crypto_sw -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/blockdev.sh crypto_sw 00:32:33.665 * Looking for test storage... 00:32:33.665 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev 00:32:33.665 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@10 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbd_common.sh 00:32:33.665 11:43:17 blockdev_crypto_sw -- bdev/nbd_common.sh@6 -- # set -e 00:32:33.665 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@12 -- # rpc_py=rpc_cmd 00:32:33.665 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@13 -- # conf_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json 00:32:33.665 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@14 -- # nonenclosed_conf_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonenclosed.json 00:32:33.665 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@15 -- # nonarray_conf_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonarray.json 00:32:33.665 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@17 -- # export RPC_PIPE_TIMEOUT=30 00:32:33.665 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@17 -- # RPC_PIPE_TIMEOUT=30 00:32:33.665 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@20 -- # : 00:32:33.665 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@670 -- # QOS_DEV_1=Malloc_0 00:32:33.665 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@671 -- # QOS_DEV_2=Null_1 00:32:33.665 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@672 -- # QOS_RUN_TIME=5 00:32:33.665 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@674 -- # uname -s 00:32:33.665 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@674 -- # '[' Linux = Linux ']' 00:32:33.665 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@676 -- # PRE_RESERVED_MEM=0 00:32:33.666 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@682 -- # test_type=crypto_sw 00:32:33.666 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@683 -- # crypto_device= 00:32:33.666 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@684 -- # dek= 00:32:33.666 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@685 -- # env_ctx= 00:32:33.666 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@686 -- # wait_for_rpc= 00:32:33.666 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@687 -- # '[' -n '' ']' 00:32:33.666 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@690 -- # [[ crypto_sw == bdev ]] 00:32:33.666 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@690 -- # [[ crypto_sw == crypto_* ]] 00:32:33.666 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@691 -- # wait_for_rpc=--wait-for-rpc 00:32:33.666 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@693 -- # start_spdk_tgt 00:32:33.666 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@47 -- # spdk_tgt_pid=1052562 00:32:33.666 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@48 -- # trap 'killprocess "$spdk_tgt_pid"; exit 1' SIGINT SIGTERM EXIT 00:32:33.666 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@49 -- # waitforlisten 1052562 00:32:33.666 11:43:17 blockdev_crypto_sw -- common/autotest_common.sh@829 -- # '[' -z 1052562 ']' 00:32:33.666 11:43:17 blockdev_crypto_sw -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:33.666 11:43:17 blockdev_crypto_sw -- common/autotest_common.sh@834 -- # local max_retries=100 00:32:33.666 11:43:17 blockdev_crypto_sw -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:33.666 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:33.666 11:43:17 blockdev_crypto_sw -- common/autotest_common.sh@838 -- # xtrace_disable 00:32:33.666 11:43:17 blockdev_crypto_sw -- common/autotest_common.sh@10 -- # set +x 00:32:33.666 11:43:17 blockdev_crypto_sw -- bdev/blockdev.sh@46 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt '' --wait-for-rpc 00:32:33.666 [2024-07-15 11:43:17.195557] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:32:33.666 [2024-07-15 11:43:17.195629] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1052562 ] 00:32:33.924 [2024-07-15 11:43:17.323737] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:33.924 [2024-07-15 11:43:17.432471] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:32:34.856 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:32:34.856 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@862 -- # return 0 00:32:34.856 11:43:18 blockdev_crypto_sw -- bdev/blockdev.sh@694 -- # case "$test_type" in 00:32:34.856 11:43:18 blockdev_crypto_sw -- bdev/blockdev.sh@711 -- # setup_crypto_sw_conf 00:32:34.856 11:43:18 blockdev_crypto_sw -- bdev/blockdev.sh@193 -- # rpc_cmd 00:32:34.856 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@559 -- # xtrace_disable 00:32:34.856 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@10 -- # set +x 00:32:34.856 Malloc0 00:32:34.856 Malloc1 00:32:34.856 true 00:32:34.856 true 00:32:34.856 true 00:32:34.856 [2024-07-15 11:43:18.409109] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw" 00:32:34.856 crypto_ram 00:32:34.856 [2024-07-15 11:43:18.417140] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw2" 00:32:34.856 crypto_ram2 00:32:34.856 [2024-07-15 11:43:18.425163] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw3" 00:32:34.856 crypto_ram3 00:32:34.856 [ 00:32:34.856 { 00:32:34.856 "name": "Malloc1", 00:32:34.856 "aliases": [ 00:32:34.856 "482f37fa-3357-41eb-9aae-d66847f992a6" 00:32:34.856 ], 00:32:34.856 "product_name": "Malloc disk", 00:32:34.856 "block_size": 4096, 00:32:34.856 "num_blocks": 4096, 00:32:34.856 "uuid": "482f37fa-3357-41eb-9aae-d66847f992a6", 00:32:34.856 "assigned_rate_limits": { 00:32:34.856 "rw_ios_per_sec": 0, 00:32:34.856 "rw_mbytes_per_sec": 0, 00:32:34.856 "r_mbytes_per_sec": 0, 00:32:34.856 "w_mbytes_per_sec": 0 00:32:34.856 }, 00:32:34.856 "claimed": true, 00:32:34.856 "claim_type": "exclusive_write", 00:32:34.856 "zoned": false, 00:32:34.856 "supported_io_types": { 00:32:34.856 "read": true, 00:32:34.856 "write": true, 00:32:34.856 "unmap": true, 00:32:34.856 "flush": true, 00:32:34.856 "reset": true, 00:32:34.856 "nvme_admin": false, 00:32:34.856 "nvme_io": false, 00:32:34.856 "nvme_io_md": false, 00:32:34.856 "write_zeroes": true, 00:32:34.856 "zcopy": true, 00:32:34.856 "get_zone_info": false, 00:32:34.856 "zone_management": false, 00:32:34.856 "zone_append": false, 00:32:34.856 "compare": false, 00:32:34.856 "compare_and_write": false, 00:32:34.856 "abort": true, 00:32:34.856 "seek_hole": false, 00:32:34.856 "seek_data": false, 00:32:34.856 "copy": true, 00:32:34.856 "nvme_iov_md": false 00:32:34.856 }, 00:32:34.856 "memory_domains": [ 00:32:34.856 { 00:32:34.856 "dma_device_id": "system", 00:32:34.856 "dma_device_type": 1 00:32:34.856 }, 00:32:34.856 { 00:32:34.856 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:34.856 "dma_device_type": 2 00:32:34.856 } 00:32:34.856 ], 00:32:34.856 "driver_specific": {} 00:32:34.856 } 00:32:34.856 ] 00:32:34.856 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:32:34.856 11:43:18 blockdev_crypto_sw -- bdev/blockdev.sh@737 -- # rpc_cmd bdev_wait_for_examine 00:32:34.856 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@559 -- # xtrace_disable 00:32:34.856 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@10 -- # set +x 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:32:35.115 11:43:18 blockdev_crypto_sw -- bdev/blockdev.sh@740 -- # cat 00:32:35.115 11:43:18 blockdev_crypto_sw -- bdev/blockdev.sh@740 -- # rpc_cmd save_subsystem_config -n accel 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@559 -- # xtrace_disable 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@10 -- # set +x 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:32:35.115 11:43:18 blockdev_crypto_sw -- bdev/blockdev.sh@740 -- # rpc_cmd save_subsystem_config -n bdev 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@559 -- # xtrace_disable 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@10 -- # set +x 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:32:35.115 11:43:18 blockdev_crypto_sw -- bdev/blockdev.sh@740 -- # rpc_cmd save_subsystem_config -n iobuf 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@559 -- # xtrace_disable 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@10 -- # set +x 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:32:35.115 11:43:18 blockdev_crypto_sw -- bdev/blockdev.sh@748 -- # mapfile -t bdevs 00:32:35.115 11:43:18 blockdev_crypto_sw -- bdev/blockdev.sh@748 -- # rpc_cmd bdev_get_bdevs 00:32:35.115 11:43:18 blockdev_crypto_sw -- bdev/blockdev.sh@748 -- # jq -r '.[] | select(.claimed == false)' 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@559 -- # xtrace_disable 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@10 -- # set +x 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:32:35.115 11:43:18 blockdev_crypto_sw -- bdev/blockdev.sh@749 -- # mapfile -t bdevs_name 00:32:35.115 11:43:18 blockdev_crypto_sw -- bdev/blockdev.sh@749 -- # jq -r .name 00:32:35.115 11:43:18 blockdev_crypto_sw -- bdev/blockdev.sh@749 -- # printf '%s\n' '{' ' "name": "crypto_ram",' ' "aliases": [' ' "9c5927c8-212c-5d34-9980-c71117547c38"' ' ],' ' "product_name": "crypto",' ' "block_size": 512,' ' "num_blocks": 32768,' ' "uuid": "9c5927c8-212c-5d34-9980-c71117547c38",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc0",' ' "name": "crypto_ram",' ' "key_name": "test_dek_sw"' ' }' ' }' '}' '{' ' "name": "crypto_ram3",' ' "aliases": [' ' "fff24f32-a580-51a1-bf7c-e9fdb68a924b"' ' ],' ' "product_name": "crypto",' ' "block_size": 4096,' ' "num_blocks": 4096,' ' "uuid": "fff24f32-a580-51a1-bf7c-e9fdb68a924b",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "crypto_ram2",' ' "name": "crypto_ram3",' ' "key_name": "test_dek_sw3"' ' }' ' }' '}' 00:32:35.115 11:43:18 blockdev_crypto_sw -- bdev/blockdev.sh@750 -- # bdev_list=("${bdevs_name[@]}") 00:32:35.115 11:43:18 blockdev_crypto_sw -- bdev/blockdev.sh@752 -- # hello_world_bdev=crypto_ram 00:32:35.115 11:43:18 blockdev_crypto_sw -- bdev/blockdev.sh@753 -- # trap - SIGINT SIGTERM EXIT 00:32:35.115 11:43:18 blockdev_crypto_sw -- bdev/blockdev.sh@754 -- # killprocess 1052562 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@948 -- # '[' -z 1052562 ']' 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@952 -- # kill -0 1052562 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@953 -- # uname 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1052562 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1052562' 00:32:35.115 killing process with pid 1052562 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@967 -- # kill 1052562 00:32:35.115 11:43:18 blockdev_crypto_sw -- common/autotest_common.sh@972 -- # wait 1052562 00:32:35.681 11:43:19 blockdev_crypto_sw -- bdev/blockdev.sh@758 -- # trap cleanup SIGINT SIGTERM EXIT 00:32:35.681 11:43:19 blockdev_crypto_sw -- bdev/blockdev.sh@760 -- # run_test bdev_hello_world /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/hello_bdev --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -b crypto_ram '' 00:32:35.681 11:43:19 blockdev_crypto_sw -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:32:35.681 11:43:19 blockdev_crypto_sw -- common/autotest_common.sh@1105 -- # xtrace_disable 00:32:35.681 11:43:19 blockdev_crypto_sw -- common/autotest_common.sh@10 -- # set +x 00:32:35.681 ************************************ 00:32:35.681 START TEST bdev_hello_world 00:32:35.681 ************************************ 00:32:35.681 11:43:19 blockdev_crypto_sw.bdev_hello_world -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/hello_bdev --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -b crypto_ram '' 00:32:35.681 [2024-07-15 11:43:19.188668] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:32:35.681 [2024-07-15 11:43:19.188729] [ DPDK EAL parameters: hello_bdev --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1052814 ] 00:32:35.939 [2024-07-15 11:43:19.316614] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:35.939 [2024-07-15 11:43:19.419489] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:32:36.197 [2024-07-15 11:43:19.593404] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw" 00:32:36.197 [2024-07-15 11:43:19.593474] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:32:36.197 [2024-07-15 11:43:19.593490] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:36.197 [2024-07-15 11:43:19.601422] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw2" 00:32:36.197 [2024-07-15 11:43:19.601443] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:32:36.197 [2024-07-15 11:43:19.601455] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:36.197 [2024-07-15 11:43:19.609444] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw3" 00:32:36.197 [2024-07-15 11:43:19.609464] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: crypto_ram2 00:32:36.197 [2024-07-15 11:43:19.609475] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:36.197 [2024-07-15 11:43:19.650962] hello_bdev.c: 222:hello_start: *NOTICE*: Successfully started the application 00:32:36.197 [2024-07-15 11:43:19.651003] hello_bdev.c: 231:hello_start: *NOTICE*: Opening the bdev crypto_ram 00:32:36.197 [2024-07-15 11:43:19.651021] hello_bdev.c: 244:hello_start: *NOTICE*: Opening io channel 00:32:36.197 [2024-07-15 11:43:19.653039] hello_bdev.c: 138:hello_write: *NOTICE*: Writing to the bdev 00:32:36.197 [2024-07-15 11:43:19.653120] hello_bdev.c: 117:write_complete: *NOTICE*: bdev io write completed successfully 00:32:36.197 [2024-07-15 11:43:19.653136] hello_bdev.c: 84:hello_read: *NOTICE*: Reading io 00:32:36.197 [2024-07-15 11:43:19.653170] hello_bdev.c: 65:read_complete: *NOTICE*: Read string from bdev : Hello World! 00:32:36.197 00:32:36.197 [2024-07-15 11:43:19.653188] hello_bdev.c: 74:read_complete: *NOTICE*: Stopping app 00:32:36.455 00:32:36.455 real 0m0.734s 00:32:36.455 user 0m0.480s 00:32:36.455 sys 0m0.242s 00:32:36.455 11:43:19 blockdev_crypto_sw.bdev_hello_world -- common/autotest_common.sh@1124 -- # xtrace_disable 00:32:36.455 11:43:19 blockdev_crypto_sw.bdev_hello_world -- common/autotest_common.sh@10 -- # set +x 00:32:36.455 ************************************ 00:32:36.455 END TEST bdev_hello_world 00:32:36.455 ************************************ 00:32:36.455 11:43:19 blockdev_crypto_sw -- common/autotest_common.sh@1142 -- # return 0 00:32:36.455 11:43:19 blockdev_crypto_sw -- bdev/blockdev.sh@761 -- # run_test bdev_bounds bdev_bounds '' 00:32:36.455 11:43:19 blockdev_crypto_sw -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:32:36.455 11:43:19 blockdev_crypto_sw -- common/autotest_common.sh@1105 -- # xtrace_disable 00:32:36.455 11:43:19 blockdev_crypto_sw -- common/autotest_common.sh@10 -- # set +x 00:32:36.455 ************************************ 00:32:36.455 START TEST bdev_bounds 00:32:36.455 ************************************ 00:32:36.455 11:43:19 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@1123 -- # bdev_bounds '' 00:32:36.455 11:43:19 blockdev_crypto_sw.bdev_bounds -- bdev/blockdev.sh@290 -- # bdevio_pid=1052977 00:32:36.455 11:43:19 blockdev_crypto_sw.bdev_bounds -- bdev/blockdev.sh@291 -- # trap 'cleanup; killprocess $bdevio_pid; exit 1' SIGINT SIGTERM EXIT 00:32:36.455 11:43:19 blockdev_crypto_sw.bdev_bounds -- bdev/blockdev.sh@292 -- # echo 'Process bdevio pid: 1052977' 00:32:36.455 Process bdevio pid: 1052977 00:32:36.455 11:43:19 blockdev_crypto_sw.bdev_bounds -- bdev/blockdev.sh@293 -- # waitforlisten 1052977 00:32:36.455 11:43:19 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@829 -- # '[' -z 1052977 ']' 00:32:36.455 11:43:19 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:36.455 11:43:19 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@834 -- # local max_retries=100 00:32:36.455 11:43:19 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:36.455 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:36.455 11:43:19 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@838 -- # xtrace_disable 00:32:36.455 11:43:19 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:32:36.455 11:43:19 blockdev_crypto_sw.bdev_bounds -- bdev/blockdev.sh@289 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevio/bdevio -w -s 0 --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json '' 00:32:36.455 [2024-07-15 11:43:19.998003] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:32:36.455 [2024-07-15 11:43:19.998070] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 0 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1052977 ] 00:32:36.713 [2024-07-15 11:43:20.128949] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:32:36.713 [2024-07-15 11:43:20.233663] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:32:36.713 [2024-07-15 11:43:20.233750] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:32:36.713 [2024-07-15 11:43:20.233754] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:32:36.971 [2024-07-15 11:43:20.400944] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw" 00:32:36.971 [2024-07-15 11:43:20.401012] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:32:36.971 [2024-07-15 11:43:20.401027] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:36.971 [2024-07-15 11:43:20.408967] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw2" 00:32:36.971 [2024-07-15 11:43:20.408988] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:32:36.971 [2024-07-15 11:43:20.408999] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:36.971 [2024-07-15 11:43:20.417000] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw3" 00:32:36.971 [2024-07-15 11:43:20.417019] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: crypto_ram2 00:32:36.971 [2024-07-15 11:43:20.417030] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:37.537 11:43:20 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:32:37.537 11:43:20 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@862 -- # return 0 00:32:37.537 11:43:20 blockdev_crypto_sw.bdev_bounds -- bdev/blockdev.sh@294 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevio/tests.py perform_tests 00:32:37.537 I/O targets: 00:32:37.537 crypto_ram: 32768 blocks of 512 bytes (16 MiB) 00:32:37.537 crypto_ram3: 4096 blocks of 4096 bytes (16 MiB) 00:32:37.537 00:32:37.537 00:32:37.537 CUnit - A unit testing framework for C - Version 2.1-3 00:32:37.537 http://cunit.sourceforge.net/ 00:32:37.537 00:32:37.537 00:32:37.537 Suite: bdevio tests on: crypto_ram3 00:32:37.537 Test: blockdev write read block ...passed 00:32:37.537 Test: blockdev write zeroes read block ...passed 00:32:37.537 Test: blockdev write zeroes read no split ...passed 00:32:37.537 Test: blockdev write zeroes read split ...passed 00:32:37.537 Test: blockdev write zeroes read split partial ...passed 00:32:37.537 Test: blockdev reset ...passed 00:32:37.537 Test: blockdev write read 8 blocks ...passed 00:32:37.537 Test: blockdev write read size > 128k ...passed 00:32:37.537 Test: blockdev write read invalid size ...passed 00:32:37.537 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:32:37.537 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:32:37.537 Test: blockdev write read max offset ...passed 00:32:37.537 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:32:37.537 Test: blockdev writev readv 8 blocks ...passed 00:32:37.537 Test: blockdev writev readv 30 x 1block ...passed 00:32:37.537 Test: blockdev writev readv block ...passed 00:32:37.537 Test: blockdev writev readv size > 128k ...passed 00:32:37.537 Test: blockdev writev readv size > 128k in two iovs ...passed 00:32:37.537 Test: blockdev comparev and writev ...passed 00:32:37.537 Test: blockdev nvme passthru rw ...passed 00:32:37.537 Test: blockdev nvme passthru vendor specific ...passed 00:32:37.537 Test: blockdev nvme admin passthru ...passed 00:32:37.537 Test: blockdev copy ...passed 00:32:37.537 Suite: bdevio tests on: crypto_ram 00:32:37.537 Test: blockdev write read block ...passed 00:32:37.537 Test: blockdev write zeroes read block ...passed 00:32:37.537 Test: blockdev write zeroes read no split ...passed 00:32:37.537 Test: blockdev write zeroes read split ...passed 00:32:37.537 Test: blockdev write zeroes read split partial ...passed 00:32:37.537 Test: blockdev reset ...passed 00:32:37.537 Test: blockdev write read 8 blocks ...passed 00:32:37.537 Test: blockdev write read size > 128k ...passed 00:32:37.537 Test: blockdev write read invalid size ...passed 00:32:37.537 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:32:37.537 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:32:37.537 Test: blockdev write read max offset ...passed 00:32:37.537 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:32:37.537 Test: blockdev writev readv 8 blocks ...passed 00:32:37.537 Test: blockdev writev readv 30 x 1block ...passed 00:32:37.537 Test: blockdev writev readv block ...passed 00:32:37.537 Test: blockdev writev readv size > 128k ...passed 00:32:37.537 Test: blockdev writev readv size > 128k in two iovs ...passed 00:32:37.537 Test: blockdev comparev and writev ...passed 00:32:37.537 Test: blockdev nvme passthru rw ...passed 00:32:37.537 Test: blockdev nvme passthru vendor specific ...passed 00:32:37.537 Test: blockdev nvme admin passthru ...passed 00:32:37.537 Test: blockdev copy ...passed 00:32:37.537 00:32:37.537 Run Summary: Type Total Ran Passed Failed Inactive 00:32:37.537 suites 2 2 n/a 0 0 00:32:37.537 tests 46 46 46 0 0 00:32:37.537 asserts 260 260 260 0 n/a 00:32:37.537 00:32:37.537 Elapsed time = 0.086 seconds 00:32:37.537 0 00:32:37.537 11:43:21 blockdev_crypto_sw.bdev_bounds -- bdev/blockdev.sh@295 -- # killprocess 1052977 00:32:37.537 11:43:21 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@948 -- # '[' -z 1052977 ']' 00:32:37.537 11:43:21 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@952 -- # kill -0 1052977 00:32:37.537 11:43:21 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@953 -- # uname 00:32:37.537 11:43:21 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:32:37.537 11:43:21 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1052977 00:32:37.796 11:43:21 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:32:37.796 11:43:21 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:32:37.796 11:43:21 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1052977' 00:32:37.796 killing process with pid 1052977 00:32:37.796 11:43:21 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@967 -- # kill 1052977 00:32:37.796 11:43:21 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@972 -- # wait 1052977 00:32:37.796 11:43:21 blockdev_crypto_sw.bdev_bounds -- bdev/blockdev.sh@296 -- # trap - SIGINT SIGTERM EXIT 00:32:37.796 00:32:37.796 real 0m1.435s 00:32:37.796 user 0m3.728s 00:32:37.796 sys 0m0.366s 00:32:37.796 11:43:21 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@1124 -- # xtrace_disable 00:32:37.796 11:43:21 blockdev_crypto_sw.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:32:37.796 ************************************ 00:32:37.796 END TEST bdev_bounds 00:32:37.796 ************************************ 00:32:38.054 11:43:21 blockdev_crypto_sw -- common/autotest_common.sh@1142 -- # return 0 00:32:38.054 11:43:21 blockdev_crypto_sw -- bdev/blockdev.sh@762 -- # run_test bdev_nbd nbd_function_test /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json 'crypto_ram crypto_ram3' '' 00:32:38.054 11:43:21 blockdev_crypto_sw -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:32:38.054 11:43:21 blockdev_crypto_sw -- common/autotest_common.sh@1105 -- # xtrace_disable 00:32:38.054 11:43:21 blockdev_crypto_sw -- common/autotest_common.sh@10 -- # set +x 00:32:38.054 ************************************ 00:32:38.054 START TEST bdev_nbd 00:32:38.054 ************************************ 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@1123 -- # nbd_function_test /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json 'crypto_ram crypto_ram3' '' 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@300 -- # uname -s 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@300 -- # [[ Linux == Linux ]] 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@302 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@303 -- # local conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@304 -- # bdev_all=('crypto_ram' 'crypto_ram3') 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@304 -- # local bdev_all 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@305 -- # local bdev_num=2 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@309 -- # [[ -e /sys/module/nbd ]] 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@311 -- # nbd_all=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@311 -- # local nbd_all 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@312 -- # bdev_num=2 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@314 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@314 -- # local nbd_list 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@315 -- # bdev_list=('crypto_ram' 'crypto_ram3') 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@315 -- # local bdev_list 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@318 -- # nbd_pid=1053208 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@319 -- # trap 'cleanup; killprocess $nbd_pid' SIGINT SIGTERM EXIT 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@317 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-nbd.sock -i 0 --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json '' 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@320 -- # waitforlisten 1053208 /var/tmp/spdk-nbd.sock 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@829 -- # '[' -z 1053208 ']' 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@834 -- # local max_retries=100 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:32:38.054 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@838 -- # xtrace_disable 00:32:38.054 11:43:21 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:32:38.054 [2024-07-15 11:43:21.523214] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:32:38.054 [2024-07-15 11:43:21.523283] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:32:38.312 [2024-07-15 11:43:21.653440] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:38.312 [2024-07-15 11:43:21.759764] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:32:38.570 [2024-07-15 11:43:21.940576] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw" 00:32:38.570 [2024-07-15 11:43:21.940642] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:32:38.570 [2024-07-15 11:43:21.940657] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:38.570 [2024-07-15 11:43:21.948595] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw2" 00:32:38.570 [2024-07-15 11:43:21.948617] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:32:38.570 [2024-07-15 11:43:21.948628] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:38.570 [2024-07-15 11:43:21.956617] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw3" 00:32:38.570 [2024-07-15 11:43:21.956636] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: crypto_ram2 00:32:38.570 [2024-07-15 11:43:21.956648] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:32:39.138 11:43:22 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:32:39.138 11:43:22 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@862 -- # return 0 00:32:39.138 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@322 -- # nbd_rpc_start_stop_verify /var/tmp/spdk-nbd.sock 'crypto_ram crypto_ram3' 00:32:39.138 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@113 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:32:39.138 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@114 -- # bdev_list=('crypto_ram' 'crypto_ram3') 00:32:39.138 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@114 -- # local bdev_list 00:32:39.138 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@116 -- # nbd_start_disks_without_nbd_idx /var/tmp/spdk-nbd.sock 'crypto_ram crypto_ram3' 00:32:39.138 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@22 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:32:39.138 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@23 -- # bdev_list=('crypto_ram' 'crypto_ram3') 00:32:39.138 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@23 -- # local bdev_list 00:32:39.138 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@24 -- # local i 00:32:39.138 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@25 -- # local nbd_device 00:32:39.138 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i = 0 )) 00:32:39.138 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 2 )) 00:32:39.139 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk crypto_ram 00:32:39.139 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd0 00:32:39.139 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd0 00:32:39.139 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd0 00:32:39.139 11:43:22 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:32:39.139 11:43:22 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:32:39.139 11:43:22 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:32:39.139 11:43:22 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:32:39.139 11:43:22 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:32:39.139 11:43:22 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:32:39.139 11:43:22 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:32:39.139 11:43:22 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:32:39.139 11:43:22 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:32:39.139 1+0 records in 00:32:39.139 1+0 records out 00:32:39.139 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000258989 s, 15.8 MB/s 00:32:39.139 11:43:22 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:32:39.397 11:43:22 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:32:39.397 11:43:22 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:32:39.397 11:43:22 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:32:39.397 11:43:22 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:32:39.397 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:32:39.397 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 2 )) 00:32:39.397 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk crypto_ram3 00:32:39.656 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd1 00:32:39.656 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd1 00:32:39.656 11:43:22 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd1 00:32:39.656 11:43:23 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:32:39.656 11:43:23 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:32:39.656 11:43:23 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:32:39.656 11:43:23 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:32:39.656 11:43:23 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:32:39.656 11:43:23 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:32:39.656 11:43:23 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:32:39.656 11:43:23 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:32:39.656 11:43:23 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:32:39.656 1+0 records in 00:32:39.656 1+0 records out 00:32:39.656 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000343927 s, 11.9 MB/s 00:32:39.656 11:43:23 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:32:39.656 11:43:23 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:32:39.656 11:43:23 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:32:39.656 11:43:23 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:32:39.656 11:43:23 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:32:39.656 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:32:39.656 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 2 )) 00:32:39.656 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@118 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:32:39.913 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@118 -- # nbd_disks_json='[ 00:32:39.913 { 00:32:39.913 "nbd_device": "/dev/nbd0", 00:32:39.913 "bdev_name": "crypto_ram" 00:32:39.913 }, 00:32:39.913 { 00:32:39.913 "nbd_device": "/dev/nbd1", 00:32:39.913 "bdev_name": "crypto_ram3" 00:32:39.913 } 00:32:39.913 ]' 00:32:39.913 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@119 -- # nbd_disks_name=($(echo "${nbd_disks_json}" | jq -r '.[] | .nbd_device')) 00:32:39.913 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@119 -- # echo '[ 00:32:39.913 { 00:32:39.913 "nbd_device": "/dev/nbd0", 00:32:39.913 "bdev_name": "crypto_ram" 00:32:39.913 }, 00:32:39.913 { 00:32:39.913 "nbd_device": "/dev/nbd1", 00:32:39.913 "bdev_name": "crypto_ram3" 00:32:39.913 } 00:32:39.913 ]' 00:32:39.913 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@119 -- # jq -r '.[] | .nbd_device' 00:32:39.913 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@120 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:32:39.913 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:32:39.913 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:32:39.913 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:32:39.913 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:32:39.913 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:32:39.913 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:32:40.170 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:32:40.170 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:32:40.170 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:32:40.170 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:32:40.170 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:32:40.170 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:32:40.170 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@39 -- # sleep 0.1 00:32:40.170 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i++ )) 00:32:40.170 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:32:40.170 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:32:40.170 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:32:40.170 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:32:40.170 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:32:40.170 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:32:40.428 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:32:40.428 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:32:40.428 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:32:40.428 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:32:40.428 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:32:40.428 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:32:40.428 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:32:40.428 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:32:40.428 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@122 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:32:40.428 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:32:40.428 11:43:23 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:32:40.684 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:32:40.684 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:32:40.684 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:32:40.684 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:32:40.684 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:32:40.684 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:32:40.684 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:32:40.684 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:32:40.942 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:32:40.942 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@122 -- # count=0 00:32:40.942 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@123 -- # '[' 0 -ne 0 ']' 00:32:40.942 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@127 -- # return 0 00:32:40.942 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@323 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'crypto_ram crypto_ram3' '/dev/nbd0 /dev/nbd1' 00:32:40.942 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:32:40.942 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@91 -- # bdev_list=('crypto_ram' 'crypto_ram3') 00:32:40.942 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@91 -- # local bdev_list 00:32:40.942 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:32:40.942 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@92 -- # local nbd_list 00:32:40.942 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'crypto_ram crypto_ram3' '/dev/nbd0 /dev/nbd1' 00:32:40.942 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:32:40.942 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@10 -- # bdev_list=('crypto_ram' 'crypto_ram3') 00:32:40.942 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@10 -- # local bdev_list 00:32:40.942 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:32:40.942 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@11 -- # local nbd_list 00:32:40.942 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@12 -- # local i 00:32:40.942 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:32:40.942 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:32:40.942 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk crypto_ram /dev/nbd0 00:32:40.942 /dev/nbd0 00:32:41.199 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:32:41.199 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:32:41.199 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:32:41.200 1+0 records in 00:32:41.200 1+0 records out 00:32:41.200 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000277566 s, 14.8 MB/s 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk crypto_ram3 /dev/nbd1 00:32:41.200 /dev/nbd1 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:32:41.200 1+0 records in 00:32:41.200 1+0 records out 00:32:41.200 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000316676 s, 12.9 MB/s 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:32:41.200 11:43:24 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:32:41.457 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:32:41.457 { 00:32:41.457 "nbd_device": "/dev/nbd0", 00:32:41.457 "bdev_name": "crypto_ram" 00:32:41.457 }, 00:32:41.457 { 00:32:41.457 "nbd_device": "/dev/nbd1", 00:32:41.457 "bdev_name": "crypto_ram3" 00:32:41.457 } 00:32:41.457 ]' 00:32:41.457 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[ 00:32:41.457 { 00:32:41.457 "nbd_device": "/dev/nbd0", 00:32:41.457 "bdev_name": "crypto_ram" 00:32:41.457 }, 00:32:41.457 { 00:32:41.457 "nbd_device": "/dev/nbd1", 00:32:41.457 "bdev_name": "crypto_ram3" 00:32:41.457 } 00:32:41.457 ]' 00:32:41.457 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:32:41.714 /dev/nbd1' 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:32:41.714 /dev/nbd1' 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=2 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 2 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@95 -- # count=2 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=write 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest bs=4096 count=256 00:32:41.714 256+0 records in 00:32:41.714 256+0 records out 00:32:41.714 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0107583 s, 97.5 MB/s 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:32:41.714 256+0 records in 00:32:41.714 256+0 records out 00:32:41.714 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0312891 s, 33.5 MB/s 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:32:41.714 256+0 records in 00:32:41.714 256+0 records out 00:32:41.714 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0417634 s, 25.1 MB/s 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=verify 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd0 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd1 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:32:41.714 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:32:41.971 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:32:41.971 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:32:41.971 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:32:41.971 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:32:41.971 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:32:41.971 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:32:41.971 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:32:41.971 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:32:41.971 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:32:41.971 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:32:42.228 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:32:42.228 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:32:42.228 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:32:42.228 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:32:42.228 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:32:42.228 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:32:42.228 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:32:42.228 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:32:42.228 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:32:42.228 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:32:42.228 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:32:42.486 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:32:42.486 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:32:42.486 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:32:42.486 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:32:42.486 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:32:42.486 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:32:42.486 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:32:42.486 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:32:42.486 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:32:42.486 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@104 -- # count=0 00:32:42.486 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:32:42.486 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@109 -- # return 0 00:32:42.486 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@324 -- # nbd_with_lvol_verify /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:32:42.486 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@131 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:32:42.486 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@132 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:32:42.486 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@132 -- # local nbd_list 00:32:42.486 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@133 -- # local mkfs_ret 00:32:42.486 11:43:25 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@135 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create -b malloc_lvol_verify 16 512 00:32:42.743 malloc_lvol_verify 00:32:42.743 11:43:26 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@136 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create_lvstore malloc_lvol_verify lvs 00:32:43.000 2793b9b1-b47b-48da-9a4a-a21c84f4e704 00:32:43.000 11:43:26 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@137 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create lvol 4 -l lvs 00:32:43.258 ddf27d9f-3448-4872-9fc2-d59db4c09dde 00:32:43.258 11:43:26 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@138 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk lvs/lvol /dev/nbd0 00:32:43.258 /dev/nbd0 00:32:43.258 11:43:26 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@140 -- # mkfs.ext4 /dev/nbd0 00:32:43.258 mke2fs 1.46.5 (30-Dec-2021) 00:32:43.258 Discarding device blocks: 0/4096 done 00:32:43.258 Creating filesystem with 4096 1k blocks and 1024 inodes 00:32:43.258 00:32:43.258 Allocating group tables: 0/1 done 00:32:43.258 Writing inode tables: 0/1 done 00:32:43.258 Creating journal (1024 blocks): done 00:32:43.258 Writing superblocks and filesystem accounting information: 0/1 done 00:32:43.258 00:32:43.258 11:43:26 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@141 -- # mkfs_ret=0 00:32:43.258 11:43:26 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@142 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:32:43.258 11:43:26 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:32:43.258 11:43:26 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:32:43.258 11:43:26 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:32:43.258 11:43:26 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:32:43.258 11:43:26 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:32:43.258 11:43:26 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:32:43.516 11:43:27 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:32:43.516 11:43:27 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:32:43.516 11:43:27 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:32:43.516 11:43:27 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:32:43.516 11:43:27 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:32:43.516 11:43:27 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:32:43.774 11:43:27 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:32:43.774 11:43:27 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:32:43.774 11:43:27 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@143 -- # '[' 0 -ne 0 ']' 00:32:43.774 11:43:27 blockdev_crypto_sw.bdev_nbd -- bdev/nbd_common.sh@147 -- # return 0 00:32:43.774 11:43:27 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@326 -- # killprocess 1053208 00:32:43.774 11:43:27 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@948 -- # '[' -z 1053208 ']' 00:32:43.774 11:43:27 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@952 -- # kill -0 1053208 00:32:43.774 11:43:27 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@953 -- # uname 00:32:43.774 11:43:27 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:32:43.774 11:43:27 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1053208 00:32:43.774 11:43:27 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:32:43.774 11:43:27 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:32:43.774 11:43:27 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1053208' 00:32:43.774 killing process with pid 1053208 00:32:43.774 11:43:27 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@967 -- # kill 1053208 00:32:43.774 11:43:27 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@972 -- # wait 1053208 00:32:44.033 11:43:27 blockdev_crypto_sw.bdev_nbd -- bdev/blockdev.sh@327 -- # trap - SIGINT SIGTERM EXIT 00:32:44.033 00:32:44.033 real 0m5.944s 00:32:44.033 user 0m8.267s 00:32:44.033 sys 0m2.505s 00:32:44.033 11:43:27 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@1124 -- # xtrace_disable 00:32:44.033 11:43:27 blockdev_crypto_sw.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:32:44.033 ************************************ 00:32:44.033 END TEST bdev_nbd 00:32:44.033 ************************************ 00:32:44.033 11:43:27 blockdev_crypto_sw -- common/autotest_common.sh@1142 -- # return 0 00:32:44.033 11:43:27 blockdev_crypto_sw -- bdev/blockdev.sh@763 -- # [[ y == y ]] 00:32:44.033 11:43:27 blockdev_crypto_sw -- bdev/blockdev.sh@764 -- # '[' crypto_sw = nvme ']' 00:32:44.033 11:43:27 blockdev_crypto_sw -- bdev/blockdev.sh@764 -- # '[' crypto_sw = gpt ']' 00:32:44.033 11:43:27 blockdev_crypto_sw -- bdev/blockdev.sh@768 -- # run_test bdev_fio fio_test_suite '' 00:32:44.033 11:43:27 blockdev_crypto_sw -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:32:44.033 11:43:27 blockdev_crypto_sw -- common/autotest_common.sh@1105 -- # xtrace_disable 00:32:44.034 11:43:27 blockdev_crypto_sw -- common/autotest_common.sh@10 -- # set +x 00:32:44.034 ************************************ 00:32:44.034 START TEST bdev_fio 00:32:44.034 ************************************ 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1123 -- # fio_test_suite '' 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@331 -- # local env_context 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@335 -- # pushd /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev 00:32:44.034 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev /var/jenkins/workspace/crypto-phy-autotest/spdk 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@336 -- # trap 'rm -f ./*.state; popd; exit 1' SIGINT SIGTERM EXIT 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@339 -- # echo '' 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@339 -- # sed s/--env-context=// 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@339 -- # env_context= 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@340 -- # fio_config_gen /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio verify AIO '' 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1280 -- # local config_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1281 -- # local workload=verify 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1282 -- # local bdev_type=AIO 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1283 -- # local env_context= 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1284 -- # local fio_dir=/usr/src/fio 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1286 -- # '[' -e /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio ']' 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1291 -- # '[' -z verify ']' 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -n '' ']' 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1299 -- # touch /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1301 -- # cat 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1313 -- # '[' verify == verify ']' 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1314 -- # cat 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1323 -- # '[' AIO == AIO ']' 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1324 -- # /usr/src/fio/fio --version 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1324 -- # [[ fio-3.35 == *\f\i\o\-\3* ]] 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1325 -- # echo serialize_overlap=1 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_crypto_ram]' 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=crypto_ram 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_crypto_ram3]' 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=crypto_ram3 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@347 -- # local 'fio_params=--ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json' 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@349 -- # run_test bdev_fio_rw_verify fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1099 -- # '[' 11 -le 1 ']' 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1105 -- # xtrace_disable 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:32:44.034 ************************************ 00:32:44.034 START TEST bdev_fio_rw_verify 00:32:44.034 ************************************ 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1123 -- # fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1339 -- # local sanitizers 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1341 -- # shift 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # local asan_lib= 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # grep libasan 00:32:44.034 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:32:44.293 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # asan_lib= 00:32:44.293 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:32:44.293 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:32:44.293 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:32:44.293 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:32:44.293 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:32:44.293 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # asan_lib= 00:32:44.293 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:32:44.294 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev' 00:32:44.294 11:43:27 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:32:44.552 job_crypto_ram: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:32:44.552 job_crypto_ram3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:32:44.552 fio-3.35 00:32:44.552 Starting 2 threads 00:32:56.795 00:32:56.795 job_crypto_ram: (groupid=0, jobs=2): err= 0: pid=1054318: Mon Jul 15 11:43:38 2024 00:32:56.795 read: IOPS=15.8k, BW=61.7MiB/s (64.7MB/s)(617MiB/10001msec) 00:32:56.795 slat (usec): min=14, max=2240, avg=31.15, stdev=12.74 00:32:56.795 clat (usec): min=8, max=2767, avg=210.99, stdev=98.79 00:32:56.795 lat (usec): min=32, max=2804, avg=242.14, stdev=103.93 00:32:56.795 clat percentiles (usec): 00:32:56.795 | 50.000th=[ 198], 99.000th=[ 445], 99.900th=[ 506], 99.990th=[ 701], 00:32:56.795 | 99.999th=[ 2737] 00:32:56.795 write: IOPS=19.0k, BW=74.2MiB/s (77.8MB/s)(705MiB/9494msec); 0 zone resets 00:32:56.795 slat (usec): min=14, max=765, avg=46.45, stdev=12.19 00:32:56.795 clat (usec): min=26, max=1284, avg=278.21, stdev=135.81 00:32:56.795 lat (usec): min=58, max=1336, avg=324.66, stdev=140.02 00:32:56.795 clat percentiles (usec): 00:32:56.795 | 50.000th=[ 265], 99.000th=[ 594], 99.900th=[ 676], 99.990th=[ 742], 00:32:56.795 | 99.999th=[ 1020] 00:32:56.795 bw ( KiB/s): min=56808, max=91560, per=94.74%, avg=72002.95, stdev=5011.72, samples=38 00:32:56.795 iops : min=14202, max=22890, avg=18000.74, stdev=1252.93, samples=38 00:32:56.795 lat (usec) : 10=0.01%, 20=0.01%, 50=0.88%, 100=9.21%, 250=45.96% 00:32:56.795 lat (usec) : 500=40.18%, 750=3.75%, 1000=0.01% 00:32:56.795 lat (msec) : 2=0.01%, 4=0.01% 00:32:56.795 cpu : usr=99.58%, sys=0.01%, ctx=45, majf=0, minf=430 00:32:56.795 IO depths : 1=12.5%, 2=25.0%, 4=50.0%, 8=12.5%, 16=0.0%, 32=0.0%, >=64=0.0% 00:32:56.795 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:32:56.795 complete : 0=0.0%, 4=88.9%, 8=11.1%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:32:56.795 issued rwts: total=158021,180394,0,0 short=0,0,0,0 dropped=0,0,0,0 00:32:56.795 latency : target=0, window=0, percentile=100.00%, depth=8 00:32:56.795 00:32:56.795 Run status group 0 (all jobs): 00:32:56.795 READ: bw=61.7MiB/s (64.7MB/s), 61.7MiB/s-61.7MiB/s (64.7MB/s-64.7MB/s), io=617MiB (647MB), run=10001-10001msec 00:32:56.795 WRITE: bw=74.2MiB/s (77.8MB/s), 74.2MiB/s-74.2MiB/s (77.8MB/s-77.8MB/s), io=705MiB (739MB), run=9494-9494msec 00:32:56.795 00:32:56.795 real 0m11.246s 00:32:56.795 user 0m23.523s 00:32:56.795 sys 0m0.370s 00:32:56.795 11:43:38 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1124 -- # xtrace_disable 00:32:56.795 11:43:38 blockdev_crypto_sw.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@10 -- # set +x 00:32:56.795 ************************************ 00:32:56.795 END TEST bdev_fio_rw_verify 00:32:56.795 ************************************ 00:32:56.795 11:43:38 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1142 -- # return 0 00:32:56.795 11:43:38 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@350 -- # rm -f 00:32:56.795 11:43:38 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@351 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:32:56.795 11:43:38 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@354 -- # fio_config_gen /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio trim '' '' 00:32:56.795 11:43:38 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1280 -- # local config_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:32:56.795 11:43:38 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1281 -- # local workload=trim 00:32:56.795 11:43:38 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1282 -- # local bdev_type= 00:32:56.795 11:43:38 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1283 -- # local env_context= 00:32:56.795 11:43:38 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1284 -- # local fio_dir=/usr/src/fio 00:32:56.795 11:43:38 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1286 -- # '[' -e /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio ']' 00:32:56.795 11:43:38 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1291 -- # '[' -z trim ']' 00:32:56.795 11:43:38 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -n '' ']' 00:32:56.795 11:43:38 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1299 -- # touch /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:32:56.795 11:43:38 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1301 -- # cat 00:32:56.795 11:43:38 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1313 -- # '[' trim == verify ']' 00:32:56.796 11:43:38 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1328 -- # '[' trim == trim ']' 00:32:56.796 11:43:38 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1329 -- # echo rw=trimwrite 00:32:56.796 11:43:38 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@355 -- # jq -r 'select(.supported_io_types.unmap == true) | .name' 00:32:56.796 11:43:38 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@355 -- # printf '%s\n' '{' ' "name": "crypto_ram",' ' "aliases": [' ' "9c5927c8-212c-5d34-9980-c71117547c38"' ' ],' ' "product_name": "crypto",' ' "block_size": 512,' ' "num_blocks": 32768,' ' "uuid": "9c5927c8-212c-5d34-9980-c71117547c38",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc0",' ' "name": "crypto_ram",' ' "key_name": "test_dek_sw"' ' }' ' }' '}' '{' ' "name": "crypto_ram3",' ' "aliases": [' ' "fff24f32-a580-51a1-bf7c-e9fdb68a924b"' ' ],' ' "product_name": "crypto",' ' "block_size": 4096,' ' "num_blocks": 4096,' ' "uuid": "fff24f32-a580-51a1-bf7c-e9fdb68a924b",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "crypto_ram2",' ' "name": "crypto_ram3",' ' "key_name": "test_dek_sw3"' ' }' ' }' '}' 00:32:56.796 11:43:38 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@355 -- # [[ -n crypto_ram 00:32:56.796 crypto_ram3 ]] 00:32:56.796 11:43:38 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@356 -- # jq -r 'select(.supported_io_types.unmap == true) | .name' 00:32:56.796 11:43:38 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@356 -- # printf '%s\n' '{' ' "name": "crypto_ram",' ' "aliases": [' ' "9c5927c8-212c-5d34-9980-c71117547c38"' ' ],' ' "product_name": "crypto",' ' "block_size": 512,' ' "num_blocks": 32768,' ' "uuid": "9c5927c8-212c-5d34-9980-c71117547c38",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc0",' ' "name": "crypto_ram",' ' "key_name": "test_dek_sw"' ' }' ' }' '}' '{' ' "name": "crypto_ram3",' ' "aliases": [' ' "fff24f32-a580-51a1-bf7c-e9fdb68a924b"' ' ],' ' "product_name": "crypto",' ' "block_size": 4096,' ' "num_blocks": 4096,' ' "uuid": "fff24f32-a580-51a1-bf7c-e9fdb68a924b",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "crypto_ram2",' ' "name": "crypto_ram3",' ' "key_name": "test_dek_sw3"' ' }' ' }' '}' 00:32:56.796 11:43:38 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:32:56.796 11:43:38 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_crypto_ram]' 00:32:56.796 11:43:38 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=crypto_ram 00:32:56.796 11:43:38 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:32:56.796 11:43:38 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_crypto_ram3]' 00:32:56.796 11:43:38 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=crypto_ram3 00:32:56.796 11:43:38 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@367 -- # run_test bdev_fio_trim fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --verify_state_save=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:32:56.796 11:43:38 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1099 -- # '[' 11 -le 1 ']' 00:32:56.796 11:43:38 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1105 -- # xtrace_disable 00:32:56.796 11:43:38 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:32:56.796 ************************************ 00:32:56.796 START TEST bdev_fio_trim 00:32:56.796 ************************************ 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1123 -- # fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --verify_state_save=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --verify_state_save=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1339 -- # local sanitizers 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1341 -- # shift 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1343 -- # local asan_lib= 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # grep libasan 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # asan_lib= 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # asan_lib= 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev' 00:32:56.796 11:43:39 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --verify_state_save=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:32:56.796 job_crypto_ram: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:32:56.796 job_crypto_ram3: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:32:56.796 fio-3.35 00:32:56.796 Starting 2 threads 00:33:06.765 00:33:06.765 job_crypto_ram: (groupid=0, jobs=2): err= 0: pid=1055833: Mon Jul 15 11:43:49 2024 00:33:06.765 write: IOPS=44.1k, BW=172MiB/s (181MB/s)(1723MiB/10001msec); 0 zone resets 00:33:06.765 slat (usec): min=11, max=126, avg=19.72, stdev= 2.91 00:33:06.765 clat (usec): min=15, max=2107, avg=148.48, stdev=57.75 00:33:06.765 lat (usec): min=38, max=2126, avg=168.20, stdev=57.47 00:33:06.765 clat percentiles (usec): 00:33:06.766 | 50.000th=[ 155], 99.000th=[ 249], 99.900th=[ 277], 99.990th=[ 553], 00:33:06.766 | 99.999th=[ 775] 00:33:06.766 bw ( KiB/s): min=167772, max=178416, per=100.00%, avg=176560.21, stdev=1183.55, samples=38 00:33:06.766 iops : min=41943, max=44604, avg=44140.05, stdev=295.89, samples=38 00:33:06.766 trim: IOPS=44.1k, BW=172MiB/s (181MB/s)(1723MiB/10001msec); 0 zone resets 00:33:06.766 slat (usec): min=5, max=1893, avg= 9.60, stdev= 3.47 00:33:06.766 clat (usec): min=29, max=2126, avg=98.39, stdev=39.07 00:33:06.766 lat (usec): min=36, max=2136, avg=107.99, stdev=39.62 00:33:06.766 clat percentiles (usec): 00:33:06.766 | 50.000th=[ 92], 99.000th=[ 182], 99.900th=[ 200], 99.990th=[ 314], 00:33:06.766 | 99.999th=[ 515] 00:33:06.766 bw ( KiB/s): min=167804, max=178408, per=100.00%, avg=176561.47, stdev=1179.56, samples=38 00:33:06.766 iops : min=41951, max=44602, avg=44140.37, stdev=294.89, samples=38 00:33:06.766 lat (usec) : 20=0.01%, 50=6.71%, 100=35.22%, 250=57.66%, 500=0.41% 00:33:06.766 lat (usec) : 750=0.01%, 1000=0.01% 00:33:06.766 lat (msec) : 4=0.01% 00:33:06.766 cpu : usr=99.61%, sys=0.01%, ctx=36, majf=0, minf=273 00:33:06.766 IO depths : 1=6.3%, 2=15.7%, 4=62.4%, 8=15.6%, 16=0.0%, 32=0.0%, >=64=0.0% 00:33:06.766 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:06.766 complete : 0=0.0%, 4=86.5%, 8=13.5%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:06.766 issued rwts: total=0,441042,441042,0 short=0,0,0,0 dropped=0,0,0,0 00:33:06.766 latency : target=0, window=0, percentile=100.00%, depth=8 00:33:06.766 00:33:06.766 Run status group 0 (all jobs): 00:33:06.766 WRITE: bw=172MiB/s (181MB/s), 172MiB/s-172MiB/s (181MB/s-181MB/s), io=1723MiB (1807MB), run=10001-10001msec 00:33:06.766 TRIM: bw=172MiB/s (181MB/s), 172MiB/s-172MiB/s (181MB/s-181MB/s), io=1723MiB (1807MB), run=10001-10001msec 00:33:06.766 00:33:06.766 real 0m11.131s 00:33:06.766 user 0m23.610s 00:33:06.766 sys 0m0.370s 00:33:06.766 11:43:50 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1124 -- # xtrace_disable 00:33:06.766 11:43:50 blockdev_crypto_sw.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@10 -- # set +x 00:33:06.766 ************************************ 00:33:06.766 END TEST bdev_fio_trim 00:33:06.766 ************************************ 00:33:06.766 11:43:50 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1142 -- # return 0 00:33:06.766 11:43:50 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@368 -- # rm -f 00:33:06.766 11:43:50 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@369 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:33:06.766 11:43:50 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@370 -- # popd 00:33:06.766 /var/jenkins/workspace/crypto-phy-autotest/spdk 00:33:06.766 11:43:50 blockdev_crypto_sw.bdev_fio -- bdev/blockdev.sh@371 -- # trap - SIGINT SIGTERM EXIT 00:33:06.766 00:33:06.766 real 0m22.724s 00:33:06.766 user 0m47.304s 00:33:06.766 sys 0m0.937s 00:33:06.766 11:43:50 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@1124 -- # xtrace_disable 00:33:06.766 11:43:50 blockdev_crypto_sw.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:33:06.766 ************************************ 00:33:06.766 END TEST bdev_fio 00:33:06.766 ************************************ 00:33:06.766 11:43:50 blockdev_crypto_sw -- common/autotest_common.sh@1142 -- # return 0 00:33:06.766 11:43:50 blockdev_crypto_sw -- bdev/blockdev.sh@775 -- # trap cleanup SIGINT SIGTERM EXIT 00:33:06.766 11:43:50 blockdev_crypto_sw -- bdev/blockdev.sh@777 -- # run_test bdev_verify /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:33:06.766 11:43:50 blockdev_crypto_sw -- common/autotest_common.sh@1099 -- # '[' 16 -le 1 ']' 00:33:06.766 11:43:50 blockdev_crypto_sw -- common/autotest_common.sh@1105 -- # xtrace_disable 00:33:06.766 11:43:50 blockdev_crypto_sw -- common/autotest_common.sh@10 -- # set +x 00:33:06.766 ************************************ 00:33:06.766 START TEST bdev_verify 00:33:06.766 ************************************ 00:33:06.766 11:43:50 blockdev_crypto_sw.bdev_verify -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:33:06.766 [2024-07-15 11:43:50.331746] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:33:06.766 [2024-07-15 11:43:50.331805] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1057249 ] 00:33:07.024 [2024-07-15 11:43:50.462898] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:33:07.024 [2024-07-15 11:43:50.565618] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:33:07.024 [2024-07-15 11:43:50.565623] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:33:07.281 [2024-07-15 11:43:50.741036] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw" 00:33:07.281 [2024-07-15 11:43:50.741107] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:33:07.281 [2024-07-15 11:43:50.741122] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:07.281 [2024-07-15 11:43:50.749057] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw2" 00:33:07.281 [2024-07-15 11:43:50.749077] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:33:07.281 [2024-07-15 11:43:50.749090] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:07.281 [2024-07-15 11:43:50.757080] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw3" 00:33:07.281 [2024-07-15 11:43:50.757099] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: crypto_ram2 00:33:07.281 [2024-07-15 11:43:50.757112] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:07.281 Running I/O for 5 seconds... 00:33:12.538 00:33:12.538 Latency(us) 00:33:12.538 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:33:12.538 Job: crypto_ram (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:33:12.538 Verification LBA range: start 0x0 length 0x800 00:33:12.538 crypto_ram : 5.04 5411.30 21.14 0.00 0.00 23546.55 1951.83 30317.52 00:33:12.538 Job: crypto_ram (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:33:12.538 Verification LBA range: start 0x800 length 0x800 00:33:12.538 crypto_ram : 5.04 5411.83 21.14 0.00 0.00 23543.93 2065.81 30545.47 00:33:12.538 Job: crypto_ram3 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:33:12.538 Verification LBA range: start 0x0 length 0x800 00:33:12.538 crypto_ram3 : 5.05 2713.40 10.60 0.00 0.00 46880.60 2336.50 34420.65 00:33:12.538 Job: crypto_ram3 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:33:12.538 Verification LBA range: start 0x800 length 0x800 00:33:12.538 crypto_ram3 : 5.05 2713.64 10.60 0.00 0.00 46871.74 2421.98 34420.65 00:33:12.538 =================================================================================================================== 00:33:12.538 Total : 16250.17 63.48 0.00 0.00 31346.52 1951.83 34420.65 00:33:12.538 00:33:12.538 real 0m5.833s 00:33:12.538 user 0m10.966s 00:33:12.538 sys 0m0.244s 00:33:12.538 11:43:56 blockdev_crypto_sw.bdev_verify -- common/autotest_common.sh@1124 -- # xtrace_disable 00:33:12.538 11:43:56 blockdev_crypto_sw.bdev_verify -- common/autotest_common.sh@10 -- # set +x 00:33:12.538 ************************************ 00:33:12.538 END TEST bdev_verify 00:33:12.538 ************************************ 00:33:12.795 11:43:56 blockdev_crypto_sw -- common/autotest_common.sh@1142 -- # return 0 00:33:12.795 11:43:56 blockdev_crypto_sw -- bdev/blockdev.sh@778 -- # run_test bdev_verify_big_io /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:33:12.795 11:43:56 blockdev_crypto_sw -- common/autotest_common.sh@1099 -- # '[' 16 -le 1 ']' 00:33:12.795 11:43:56 blockdev_crypto_sw -- common/autotest_common.sh@1105 -- # xtrace_disable 00:33:12.795 11:43:56 blockdev_crypto_sw -- common/autotest_common.sh@10 -- # set +x 00:33:12.795 ************************************ 00:33:12.795 START TEST bdev_verify_big_io 00:33:12.795 ************************************ 00:33:12.795 11:43:56 blockdev_crypto_sw.bdev_verify_big_io -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:33:12.795 [2024-07-15 11:43:56.230622] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:33:12.795 [2024-07-15 11:43:56.230681] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1057972 ] 00:33:12.795 [2024-07-15 11:43:56.360622] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:33:13.051 [2024-07-15 11:43:56.459132] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:33:13.051 [2024-07-15 11:43:56.459140] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:33:13.051 [2024-07-15 11:43:56.622959] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw" 00:33:13.051 [2024-07-15 11:43:56.623027] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:33:13.051 [2024-07-15 11:43:56.623042] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:13.051 [2024-07-15 11:43:56.630978] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw2" 00:33:13.051 [2024-07-15 11:43:56.630998] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:33:13.051 [2024-07-15 11:43:56.631009] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:13.051 [2024-07-15 11:43:56.638997] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw3" 00:33:13.051 [2024-07-15 11:43:56.639016] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: crypto_ram2 00:33:13.051 [2024-07-15 11:43:56.639028] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:13.308 Running I/O for 5 seconds... 00:33:18.563 00:33:18.563 Latency(us) 00:33:18.563 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:33:18.563 Job: crypto_ram (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:33:18.563 Verification LBA range: start 0x0 length 0x80 00:33:18.563 crypto_ram : 5.22 465.67 29.10 0.00 0.00 268801.72 6126.19 366545.70 00:33:18.563 Job: crypto_ram (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:33:18.563 Verification LBA range: start 0x80 length 0x80 00:33:18.563 crypto_ram : 5.20 467.47 29.22 0.00 0.00 267846.64 6097.70 364722.09 00:33:18.563 Job: crypto_ram3 (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:33:18.563 Verification LBA range: start 0x0 length 0x80 00:33:18.563 crypto_ram3 : 5.24 244.46 15.28 0.00 0.00 493582.00 5698.78 373840.14 00:33:18.563 Job: crypto_ram3 (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:33:18.563 Verification LBA range: start 0x80 length 0x80 00:33:18.563 crypto_ram3 : 5.22 245.31 15.33 0.00 0.00 491798.84 6468.12 373840.14 00:33:18.563 =================================================================================================================== 00:33:18.563 Total : 1422.90 88.93 0.00 0.00 345691.85 5698.78 373840.14 00:33:18.820 00:33:18.820 real 0m6.002s 00:33:18.820 user 0m11.328s 00:33:18.820 sys 0m0.226s 00:33:18.820 11:44:02 blockdev_crypto_sw.bdev_verify_big_io -- common/autotest_common.sh@1124 -- # xtrace_disable 00:33:18.820 11:44:02 blockdev_crypto_sw.bdev_verify_big_io -- common/autotest_common.sh@10 -- # set +x 00:33:18.820 ************************************ 00:33:18.820 END TEST bdev_verify_big_io 00:33:18.820 ************************************ 00:33:18.820 11:44:02 blockdev_crypto_sw -- common/autotest_common.sh@1142 -- # return 0 00:33:18.820 11:44:02 blockdev_crypto_sw -- bdev/blockdev.sh@779 -- # run_test bdev_write_zeroes /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:33:18.820 11:44:02 blockdev_crypto_sw -- common/autotest_common.sh@1099 -- # '[' 13 -le 1 ']' 00:33:18.820 11:44:02 blockdev_crypto_sw -- common/autotest_common.sh@1105 -- # xtrace_disable 00:33:18.820 11:44:02 blockdev_crypto_sw -- common/autotest_common.sh@10 -- # set +x 00:33:18.820 ************************************ 00:33:18.821 START TEST bdev_write_zeroes 00:33:18.821 ************************************ 00:33:18.821 11:44:02 blockdev_crypto_sw.bdev_write_zeroes -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:33:18.821 [2024-07-15 11:44:02.321687] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:33:18.821 [2024-07-15 11:44:02.321748] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1058758 ] 00:33:19.078 [2024-07-15 11:44:02.448061] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:19.078 [2024-07-15 11:44:02.549292] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:33:19.336 [2024-07-15 11:44:02.729489] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw" 00:33:19.336 [2024-07-15 11:44:02.729561] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:33:19.336 [2024-07-15 11:44:02.729577] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:19.336 [2024-07-15 11:44:02.737509] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw2" 00:33:19.336 [2024-07-15 11:44:02.737529] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:33:19.336 [2024-07-15 11:44:02.737541] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:19.336 [2024-07-15 11:44:02.745547] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw3" 00:33:19.336 [2024-07-15 11:44:02.745566] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: crypto_ram2 00:33:19.336 [2024-07-15 11:44:02.745578] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:19.336 Running I/O for 1 seconds... 00:33:20.268 00:33:20.268 Latency(us) 00:33:20.268 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:33:20.268 Job: crypto_ram (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:33:20.268 crypto_ram : 1.01 26522.53 103.60 0.00 0.00 4813.75 2080.06 6439.62 00:33:20.268 Job: crypto_ram3 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:33:20.268 crypto_ram3 : 1.01 13290.74 51.92 0.00 0.00 9562.49 3348.03 9744.92 00:33:20.268 =================================================================================================================== 00:33:20.268 Total : 39813.27 155.52 0.00 0.00 6401.70 2080.06 9744.92 00:33:20.525 00:33:20.525 real 0m1.761s 00:33:20.525 user 0m1.504s 00:33:20.525 sys 0m0.235s 00:33:20.525 11:44:04 blockdev_crypto_sw.bdev_write_zeroes -- common/autotest_common.sh@1124 -- # xtrace_disable 00:33:20.525 11:44:04 blockdev_crypto_sw.bdev_write_zeroes -- common/autotest_common.sh@10 -- # set +x 00:33:20.525 ************************************ 00:33:20.525 END TEST bdev_write_zeroes 00:33:20.525 ************************************ 00:33:20.525 11:44:04 blockdev_crypto_sw -- common/autotest_common.sh@1142 -- # return 0 00:33:20.525 11:44:04 blockdev_crypto_sw -- bdev/blockdev.sh@782 -- # run_test bdev_json_nonenclosed /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:33:20.525 11:44:04 blockdev_crypto_sw -- common/autotest_common.sh@1099 -- # '[' 13 -le 1 ']' 00:33:20.525 11:44:04 blockdev_crypto_sw -- common/autotest_common.sh@1105 -- # xtrace_disable 00:33:20.525 11:44:04 blockdev_crypto_sw -- common/autotest_common.sh@10 -- # set +x 00:33:20.525 ************************************ 00:33:20.525 START TEST bdev_json_nonenclosed 00:33:20.525 ************************************ 00:33:20.525 11:44:04 blockdev_crypto_sw.bdev_json_nonenclosed -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:33:20.840 [2024-07-15 11:44:04.156987] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:33:20.840 [2024-07-15 11:44:04.157052] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1059047 ] 00:33:20.840 [2024-07-15 11:44:04.283790] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:20.840 [2024-07-15 11:44:04.380847] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:33:20.840 [2024-07-15 11:44:04.380919] json_config.c: 608:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: not enclosed in {}. 00:33:20.840 [2024-07-15 11:44:04.380946] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:33:20.840 [2024-07-15 11:44:04.380960] app.c:1053:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:33:21.099 00:33:21.099 real 0m0.387s 00:33:21.099 user 0m0.243s 00:33:21.099 sys 0m0.142s 00:33:21.099 11:44:04 blockdev_crypto_sw.bdev_json_nonenclosed -- common/autotest_common.sh@1123 -- # es=234 00:33:21.099 11:44:04 blockdev_crypto_sw.bdev_json_nonenclosed -- common/autotest_common.sh@1124 -- # xtrace_disable 00:33:21.099 11:44:04 blockdev_crypto_sw.bdev_json_nonenclosed -- common/autotest_common.sh@10 -- # set +x 00:33:21.099 ************************************ 00:33:21.099 END TEST bdev_json_nonenclosed 00:33:21.099 ************************************ 00:33:21.099 11:44:04 blockdev_crypto_sw -- common/autotest_common.sh@1142 -- # return 234 00:33:21.099 11:44:04 blockdev_crypto_sw -- bdev/blockdev.sh@782 -- # true 00:33:21.099 11:44:04 blockdev_crypto_sw -- bdev/blockdev.sh@785 -- # run_test bdev_json_nonarray /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:33:21.099 11:44:04 blockdev_crypto_sw -- common/autotest_common.sh@1099 -- # '[' 13 -le 1 ']' 00:33:21.099 11:44:04 blockdev_crypto_sw -- common/autotest_common.sh@1105 -- # xtrace_disable 00:33:21.099 11:44:04 blockdev_crypto_sw -- common/autotest_common.sh@10 -- # set +x 00:33:21.099 ************************************ 00:33:21.099 START TEST bdev_json_nonarray 00:33:21.099 ************************************ 00:33:21.099 11:44:04 blockdev_crypto_sw.bdev_json_nonarray -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:33:21.099 [2024-07-15 11:44:04.629231] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:33:21.099 [2024-07-15 11:44:04.629290] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1059069 ] 00:33:21.357 [2024-07-15 11:44:04.756671] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:21.357 [2024-07-15 11:44:04.859761] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:33:21.357 [2024-07-15 11:44:04.859832] json_config.c: 614:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: 'subsystems' should be an array. 00:33:21.357 [2024-07-15 11:44:04.859853] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:33:21.357 [2024-07-15 11:44:04.859866] app.c:1053:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:33:21.615 00:33:21.615 real 0m0.397s 00:33:21.615 user 0m0.244s 00:33:21.615 sys 0m0.150s 00:33:21.615 11:44:04 blockdev_crypto_sw.bdev_json_nonarray -- common/autotest_common.sh@1123 -- # es=234 00:33:21.615 11:44:04 blockdev_crypto_sw.bdev_json_nonarray -- common/autotest_common.sh@1124 -- # xtrace_disable 00:33:21.615 11:44:04 blockdev_crypto_sw.bdev_json_nonarray -- common/autotest_common.sh@10 -- # set +x 00:33:21.615 ************************************ 00:33:21.615 END TEST bdev_json_nonarray 00:33:21.615 ************************************ 00:33:21.615 11:44:05 blockdev_crypto_sw -- common/autotest_common.sh@1142 -- # return 234 00:33:21.615 11:44:05 blockdev_crypto_sw -- bdev/blockdev.sh@785 -- # true 00:33:21.616 11:44:05 blockdev_crypto_sw -- bdev/blockdev.sh@787 -- # [[ crypto_sw == bdev ]] 00:33:21.616 11:44:05 blockdev_crypto_sw -- bdev/blockdev.sh@794 -- # [[ crypto_sw == gpt ]] 00:33:21.616 11:44:05 blockdev_crypto_sw -- bdev/blockdev.sh@798 -- # [[ crypto_sw == crypto_sw ]] 00:33:21.616 11:44:05 blockdev_crypto_sw -- bdev/blockdev.sh@799 -- # run_test bdev_crypto_enomem bdev_crypto_enomem 00:33:21.616 11:44:05 blockdev_crypto_sw -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:33:21.616 11:44:05 blockdev_crypto_sw -- common/autotest_common.sh@1105 -- # xtrace_disable 00:33:21.616 11:44:05 blockdev_crypto_sw -- common/autotest_common.sh@10 -- # set +x 00:33:21.616 ************************************ 00:33:21.616 START TEST bdev_crypto_enomem 00:33:21.616 ************************************ 00:33:21.616 11:44:05 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@1123 -- # bdev_crypto_enomem 00:33:21.616 11:44:05 blockdev_crypto_sw.bdev_crypto_enomem -- bdev/blockdev.sh@635 -- # local base_dev=base0 00:33:21.616 11:44:05 blockdev_crypto_sw.bdev_crypto_enomem -- bdev/blockdev.sh@636 -- # local test_dev=crypt0 00:33:21.616 11:44:05 blockdev_crypto_sw.bdev_crypto_enomem -- bdev/blockdev.sh@637 -- # local err_dev=EE_base0 00:33:21.616 11:44:05 blockdev_crypto_sw.bdev_crypto_enomem -- bdev/blockdev.sh@638 -- # local qd=32 00:33:21.616 11:44:05 blockdev_crypto_sw.bdev_crypto_enomem -- bdev/blockdev.sh@641 -- # ERR_PID=1059195 00:33:21.616 11:44:05 blockdev_crypto_sw.bdev_crypto_enomem -- bdev/blockdev.sh@642 -- # trap 'cleanup; killprocess $ERR_PID; exit 1' SIGINT SIGTERM EXIT 00:33:21.616 11:44:05 blockdev_crypto_sw.bdev_crypto_enomem -- bdev/blockdev.sh@640 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -z -m 0x2 -q 32 -o 4096 -w randwrite -t 5 -f '' 00:33:21.616 11:44:05 blockdev_crypto_sw.bdev_crypto_enomem -- bdev/blockdev.sh@643 -- # waitforlisten 1059195 00:33:21.616 11:44:05 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@829 -- # '[' -z 1059195 ']' 00:33:21.616 11:44:05 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:21.616 11:44:05 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@834 -- # local max_retries=100 00:33:21.616 11:44:05 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:21.616 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:21.616 11:44:05 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@838 -- # xtrace_disable 00:33:21.616 11:44:05 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@10 -- # set +x 00:33:21.616 [2024-07-15 11:44:05.115815] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:33:21.616 [2024-07-15 11:44:05.115889] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1059195 ] 00:33:21.874 [2024-07-15 11:44:05.235780] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:21.874 [2024-07-15 11:44:05.332505] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:33:22.440 11:44:05 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:33:22.440 11:44:05 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@862 -- # return 0 00:33:22.440 11:44:05 blockdev_crypto_sw.bdev_crypto_enomem -- bdev/blockdev.sh@645 -- # rpc_cmd 00:33:22.440 11:44:05 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@559 -- # xtrace_disable 00:33:22.440 11:44:05 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@10 -- # set +x 00:33:22.440 true 00:33:22.440 base0 00:33:22.440 true 00:33:22.440 [2024-07-15 11:44:05.999369] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_sw" 00:33:22.440 crypt0 00:33:22.440 11:44:06 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:33:22.440 11:44:06 blockdev_crypto_sw.bdev_crypto_enomem -- bdev/blockdev.sh@652 -- # waitforbdev crypt0 00:33:22.440 11:44:06 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@897 -- # local bdev_name=crypt0 00:33:22.440 11:44:06 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:33:22.440 11:44:06 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@899 -- # local i 00:33:22.440 11:44:06 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:33:22.440 11:44:06 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:33:22.440 11:44:06 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@902 -- # rpc_cmd bdev_wait_for_examine 00:33:22.440 11:44:06 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@559 -- # xtrace_disable 00:33:22.440 11:44:06 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@10 -- # set +x 00:33:22.440 11:44:06 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:33:22.440 11:44:06 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@904 -- # rpc_cmd bdev_get_bdevs -b crypt0 -t 2000 00:33:22.440 11:44:06 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@559 -- # xtrace_disable 00:33:22.440 11:44:06 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@10 -- # set +x 00:33:22.440 [ 00:33:22.440 { 00:33:22.440 "name": "crypt0", 00:33:22.440 "aliases": [ 00:33:22.440 "ff5cb90b-3224-5147-be47-a0d1b4737c68" 00:33:22.440 ], 00:33:22.440 "product_name": "crypto", 00:33:22.440 "block_size": 512, 00:33:22.440 "num_blocks": 2097152, 00:33:22.440 "uuid": "ff5cb90b-3224-5147-be47-a0d1b4737c68", 00:33:22.440 "assigned_rate_limits": { 00:33:22.440 "rw_ios_per_sec": 0, 00:33:22.440 "rw_mbytes_per_sec": 0, 00:33:22.440 "r_mbytes_per_sec": 0, 00:33:22.440 "w_mbytes_per_sec": 0 00:33:22.440 }, 00:33:22.440 "claimed": false, 00:33:22.440 "zoned": false, 00:33:22.440 "supported_io_types": { 00:33:22.440 "read": true, 00:33:22.440 "write": true, 00:33:22.440 "unmap": false, 00:33:22.440 "flush": false, 00:33:22.440 "reset": true, 00:33:22.440 "nvme_admin": false, 00:33:22.440 "nvme_io": false, 00:33:22.440 "nvme_io_md": false, 00:33:22.440 "write_zeroes": true, 00:33:22.440 "zcopy": false, 00:33:22.440 "get_zone_info": false, 00:33:22.440 "zone_management": false, 00:33:22.440 "zone_append": false, 00:33:22.440 "compare": false, 00:33:22.440 "compare_and_write": false, 00:33:22.440 "abort": false, 00:33:22.440 "seek_hole": false, 00:33:22.440 "seek_data": false, 00:33:22.440 "copy": false, 00:33:22.440 "nvme_iov_md": false 00:33:22.440 }, 00:33:22.440 "memory_domains": [ 00:33:22.440 { 00:33:22.440 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:22.440 "dma_device_type": 2 00:33:22.440 } 00:33:22.440 ], 00:33:22.440 "driver_specific": { 00:33:22.440 "crypto": { 00:33:22.440 "base_bdev_name": "EE_base0", 00:33:22.440 "name": "crypt0", 00:33:22.440 "key_name": "test_dek_sw" 00:33:22.440 } 00:33:22.440 } 00:33:22.440 } 00:33:22.440 ] 00:33:22.440 11:44:06 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:33:22.440 11:44:06 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@905 -- # return 0 00:33:22.440 11:44:06 blockdev_crypto_sw.bdev_crypto_enomem -- bdev/blockdev.sh@655 -- # rpcpid=1059274 00:33:22.440 11:44:06 blockdev_crypto_sw.bdev_crypto_enomem -- bdev/blockdev.sh@657 -- # sleep 1 00:33:22.440 11:44:06 blockdev_crypto_sw.bdev_crypto_enomem -- bdev/blockdev.sh@654 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:33:22.698 Running I/O for 5 seconds... 00:33:23.681 11:44:07 blockdev_crypto_sw.bdev_crypto_enomem -- bdev/blockdev.sh@658 -- # rpc_cmd bdev_error_inject_error EE_base0 -n 5 -q 31 write nomem 00:33:23.681 11:44:07 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@559 -- # xtrace_disable 00:33:23.681 11:44:07 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@10 -- # set +x 00:33:23.681 11:44:07 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:33:23.681 11:44:07 blockdev_crypto_sw.bdev_crypto_enomem -- bdev/blockdev.sh@660 -- # wait 1059274 00:33:27.869 00:33:27.869 Latency(us) 00:33:27.869 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:33:27.869 Job: crypt0 (Core Mask 0x2, workload: randwrite, depth: 32, IO size: 4096) 00:33:27.869 crypt0 : 5.00 36190.74 141.37 0.00 0.00 880.52 418.50 1417.57 00:33:27.869 =================================================================================================================== 00:33:27.869 Total : 36190.74 141.37 0.00 0.00 880.52 418.50 1417.57 00:33:27.869 0 00:33:27.869 11:44:11 blockdev_crypto_sw.bdev_crypto_enomem -- bdev/blockdev.sh@662 -- # rpc_cmd bdev_crypto_delete crypt0 00:33:27.869 11:44:11 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@559 -- # xtrace_disable 00:33:27.869 11:44:11 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@10 -- # set +x 00:33:27.869 11:44:11 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:33:27.869 11:44:11 blockdev_crypto_sw.bdev_crypto_enomem -- bdev/blockdev.sh@664 -- # killprocess 1059195 00:33:27.869 11:44:11 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@948 -- # '[' -z 1059195 ']' 00:33:27.869 11:44:11 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@952 -- # kill -0 1059195 00:33:27.869 11:44:11 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@953 -- # uname 00:33:27.869 11:44:11 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:33:27.869 11:44:11 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1059195 00:33:27.869 11:44:11 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:33:27.869 11:44:11 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:33:27.869 11:44:11 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1059195' 00:33:27.869 killing process with pid 1059195 00:33:27.869 11:44:11 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@967 -- # kill 1059195 00:33:27.869 Received shutdown signal, test time was about 5.000000 seconds 00:33:27.869 00:33:27.869 Latency(us) 00:33:27.869 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:33:27.869 =================================================================================================================== 00:33:27.869 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:33:27.869 11:44:11 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@972 -- # wait 1059195 00:33:27.869 11:44:11 blockdev_crypto_sw.bdev_crypto_enomem -- bdev/blockdev.sh@665 -- # trap - SIGINT SIGTERM EXIT 00:33:27.869 00:33:27.869 real 0m6.346s 00:33:27.869 user 0m6.503s 00:33:27.869 sys 0m0.376s 00:33:27.869 11:44:11 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@1124 -- # xtrace_disable 00:33:27.869 11:44:11 blockdev_crypto_sw.bdev_crypto_enomem -- common/autotest_common.sh@10 -- # set +x 00:33:27.869 ************************************ 00:33:27.869 END TEST bdev_crypto_enomem 00:33:27.869 ************************************ 00:33:27.869 11:44:11 blockdev_crypto_sw -- common/autotest_common.sh@1142 -- # return 0 00:33:27.869 11:44:11 blockdev_crypto_sw -- bdev/blockdev.sh@810 -- # trap - SIGINT SIGTERM EXIT 00:33:27.869 11:44:11 blockdev_crypto_sw -- bdev/blockdev.sh@811 -- # cleanup 00:33:27.869 11:44:11 blockdev_crypto_sw -- bdev/blockdev.sh@23 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/aiofile 00:33:27.869 11:44:11 blockdev_crypto_sw -- bdev/blockdev.sh@24 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json 00:33:27.869 11:44:11 blockdev_crypto_sw -- bdev/blockdev.sh@26 -- # [[ crypto_sw == rbd ]] 00:33:27.869 11:44:11 blockdev_crypto_sw -- bdev/blockdev.sh@30 -- # [[ crypto_sw == daos ]] 00:33:27.869 11:44:11 blockdev_crypto_sw -- bdev/blockdev.sh@34 -- # [[ crypto_sw = \g\p\t ]] 00:33:27.869 11:44:11 blockdev_crypto_sw -- bdev/blockdev.sh@40 -- # [[ crypto_sw == xnvme ]] 00:33:27.869 00:33:27.869 real 0m54.441s 00:33:27.869 user 1m33.006s 00:33:27.869 sys 0m6.630s 00:33:27.869 11:44:11 blockdev_crypto_sw -- common/autotest_common.sh@1124 -- # xtrace_disable 00:33:27.869 11:44:11 blockdev_crypto_sw -- common/autotest_common.sh@10 -- # set +x 00:33:27.869 ************************************ 00:33:27.869 END TEST blockdev_crypto_sw 00:33:27.869 ************************************ 00:33:28.129 11:44:11 -- common/autotest_common.sh@1142 -- # return 0 00:33:28.129 11:44:11 -- spdk/autotest.sh@359 -- # run_test blockdev_crypto_qat /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/blockdev.sh crypto_qat 00:33:28.129 11:44:11 -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:33:28.129 11:44:11 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:33:28.129 11:44:11 -- common/autotest_common.sh@10 -- # set +x 00:33:28.129 ************************************ 00:33:28.129 START TEST blockdev_crypto_qat 00:33:28.129 ************************************ 00:33:28.129 11:44:11 blockdev_crypto_qat -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/blockdev.sh crypto_qat 00:33:28.130 * Looking for test storage... 00:33:28.130 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@10 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbd_common.sh 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/nbd_common.sh@6 -- # set -e 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@12 -- # rpc_py=rpc_cmd 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@13 -- # conf_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@14 -- # nonenclosed_conf_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonenclosed.json 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@15 -- # nonarray_conf_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonarray.json 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@17 -- # export RPC_PIPE_TIMEOUT=30 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@17 -- # RPC_PIPE_TIMEOUT=30 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@20 -- # : 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@670 -- # QOS_DEV_1=Malloc_0 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@671 -- # QOS_DEV_2=Null_1 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@672 -- # QOS_RUN_TIME=5 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@674 -- # uname -s 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@674 -- # '[' Linux = Linux ']' 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@676 -- # PRE_RESERVED_MEM=0 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@682 -- # test_type=crypto_qat 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@683 -- # crypto_device= 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@684 -- # dek= 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@685 -- # env_ctx= 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@686 -- # wait_for_rpc= 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@687 -- # '[' -n '' ']' 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@690 -- # [[ crypto_qat == bdev ]] 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@690 -- # [[ crypto_qat == crypto_* ]] 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@691 -- # wait_for_rpc=--wait-for-rpc 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@693 -- # start_spdk_tgt 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@47 -- # spdk_tgt_pid=1060041 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@48 -- # trap 'killprocess "$spdk_tgt_pid"; exit 1' SIGINT SIGTERM EXIT 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@46 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_tgt '' --wait-for-rpc 00:33:28.130 11:44:11 blockdev_crypto_qat -- bdev/blockdev.sh@49 -- # waitforlisten 1060041 00:33:28.130 11:44:11 blockdev_crypto_qat -- common/autotest_common.sh@829 -- # '[' -z 1060041 ']' 00:33:28.130 11:44:11 blockdev_crypto_qat -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:28.130 11:44:11 blockdev_crypto_qat -- common/autotest_common.sh@834 -- # local max_retries=100 00:33:28.130 11:44:11 blockdev_crypto_qat -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:28.130 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:28.130 11:44:11 blockdev_crypto_qat -- common/autotest_common.sh@838 -- # xtrace_disable 00:33:28.130 11:44:11 blockdev_crypto_qat -- common/autotest_common.sh@10 -- # set +x 00:33:28.389 [2024-07-15 11:44:11.737911] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:33:28.389 [2024-07-15 11:44:11.737997] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1060041 ] 00:33:28.389 [2024-07-15 11:44:11.870132] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:28.389 [2024-07-15 11:44:11.974317] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:33:29.326 11:44:12 blockdev_crypto_qat -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:33:29.326 11:44:12 blockdev_crypto_qat -- common/autotest_common.sh@862 -- # return 0 00:33:29.326 11:44:12 blockdev_crypto_qat -- bdev/blockdev.sh@694 -- # case "$test_type" in 00:33:29.326 11:44:12 blockdev_crypto_qat -- bdev/blockdev.sh@708 -- # setup_crypto_qat_conf 00:33:29.326 11:44:12 blockdev_crypto_qat -- bdev/blockdev.sh@170 -- # rpc_cmd 00:33:29.326 11:44:12 blockdev_crypto_qat -- common/autotest_common.sh@559 -- # xtrace_disable 00:33:29.326 11:44:12 blockdev_crypto_qat -- common/autotest_common.sh@10 -- # set +x 00:33:29.326 [2024-07-15 11:44:12.660683] accel_dpdk_cryptodev.c: 223:accel_dpdk_cryptodev_set_driver: *NOTICE*: Using driver crypto_qat 00:33:29.326 [2024-07-15 11:44:12.668718] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation encrypt will be assigned to module dpdk_cryptodev 00:33:29.326 [2024-07-15 11:44:12.676736] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation decrypt will be assigned to module dpdk_cryptodev 00:33:29.326 [2024-07-15 11:44:12.742567] accel_dpdk_cryptodev.c:1178:accel_dpdk_cryptodev_init: *NOTICE*: Found crypto devices: 96 00:33:31.861 true 00:33:31.861 true 00:33:31.861 true 00:33:31.861 true 00:33:31.861 Malloc0 00:33:31.861 Malloc1 00:33:31.861 Malloc2 00:33:31.861 Malloc3 00:33:31.861 [2024-07-15 11:44:15.118372] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_cbc" 00:33:31.861 crypto_ram 00:33:31.861 [2024-07-15 11:44:15.126386] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_xts" 00:33:31.861 crypto_ram1 00:33:31.861 [2024-07-15 11:44:15.134405] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_cbc2" 00:33:31.861 crypto_ram2 00:33:31.861 [2024-07-15 11:44:15.142428] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_xts2" 00:33:31.861 crypto_ram3 00:33:31.861 [ 00:33:31.861 { 00:33:31.861 "name": "Malloc1", 00:33:31.861 "aliases": [ 00:33:31.862 "845d36ca-c435-49fe-968e-807d33bb255f" 00:33:31.862 ], 00:33:31.862 "product_name": "Malloc disk", 00:33:31.862 "block_size": 512, 00:33:31.862 "num_blocks": 65536, 00:33:31.862 "uuid": "845d36ca-c435-49fe-968e-807d33bb255f", 00:33:31.862 "assigned_rate_limits": { 00:33:31.862 "rw_ios_per_sec": 0, 00:33:31.862 "rw_mbytes_per_sec": 0, 00:33:31.862 "r_mbytes_per_sec": 0, 00:33:31.862 "w_mbytes_per_sec": 0 00:33:31.862 }, 00:33:31.862 "claimed": true, 00:33:31.862 "claim_type": "exclusive_write", 00:33:31.862 "zoned": false, 00:33:31.862 "supported_io_types": { 00:33:31.862 "read": true, 00:33:31.862 "write": true, 00:33:31.862 "unmap": true, 00:33:31.862 "flush": true, 00:33:31.862 "reset": true, 00:33:31.862 "nvme_admin": false, 00:33:31.862 "nvme_io": false, 00:33:31.862 "nvme_io_md": false, 00:33:31.862 "write_zeroes": true, 00:33:31.862 "zcopy": true, 00:33:31.862 "get_zone_info": false, 00:33:31.862 "zone_management": false, 00:33:31.862 "zone_append": false, 00:33:31.862 "compare": false, 00:33:31.862 "compare_and_write": false, 00:33:31.862 "abort": true, 00:33:31.862 "seek_hole": false, 00:33:31.862 "seek_data": false, 00:33:31.862 "copy": true, 00:33:31.862 "nvme_iov_md": false 00:33:31.862 }, 00:33:31.862 "memory_domains": [ 00:33:31.862 { 00:33:31.862 "dma_device_id": "system", 00:33:31.862 "dma_device_type": 1 00:33:31.862 }, 00:33:31.862 { 00:33:31.862 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:31.862 "dma_device_type": 2 00:33:31.862 } 00:33:31.862 ], 00:33:31.862 "driver_specific": {} 00:33:31.862 } 00:33:31.862 ] 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:33:31.862 11:44:15 blockdev_crypto_qat -- bdev/blockdev.sh@737 -- # rpc_cmd bdev_wait_for_examine 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@559 -- # xtrace_disable 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@10 -- # set +x 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:33:31.862 11:44:15 blockdev_crypto_qat -- bdev/blockdev.sh@740 -- # cat 00:33:31.862 11:44:15 blockdev_crypto_qat -- bdev/blockdev.sh@740 -- # rpc_cmd save_subsystem_config -n accel 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@559 -- # xtrace_disable 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@10 -- # set +x 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:33:31.862 11:44:15 blockdev_crypto_qat -- bdev/blockdev.sh@740 -- # rpc_cmd save_subsystem_config -n bdev 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@559 -- # xtrace_disable 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@10 -- # set +x 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:33:31.862 11:44:15 blockdev_crypto_qat -- bdev/blockdev.sh@740 -- # rpc_cmd save_subsystem_config -n iobuf 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@559 -- # xtrace_disable 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@10 -- # set +x 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:33:31.862 11:44:15 blockdev_crypto_qat -- bdev/blockdev.sh@748 -- # mapfile -t bdevs 00:33:31.862 11:44:15 blockdev_crypto_qat -- bdev/blockdev.sh@748 -- # rpc_cmd bdev_get_bdevs 00:33:31.862 11:44:15 blockdev_crypto_qat -- bdev/blockdev.sh@748 -- # jq -r '.[] | select(.claimed == false)' 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@559 -- # xtrace_disable 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@10 -- # set +x 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:33:31.862 11:44:15 blockdev_crypto_qat -- bdev/blockdev.sh@749 -- # mapfile -t bdevs_name 00:33:31.862 11:44:15 blockdev_crypto_qat -- bdev/blockdev.sh@749 -- # jq -r .name 00:33:31.862 11:44:15 blockdev_crypto_qat -- bdev/blockdev.sh@749 -- # printf '%s\n' '{' ' "name": "crypto_ram",' ' "aliases": [' ' "df965139-bb3b-5c62-b107-9ca32835792e"' ' ],' ' "product_name": "crypto",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "df965139-bb3b-5c62-b107-9ca32835792e",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc0",' ' "name": "crypto_ram",' ' "key_name": "test_dek_qat_cbc"' ' }' ' }' '}' '{' ' "name": "crypto_ram1",' ' "aliases": [' ' "4051e7bd-73e2-518a-b930-d41a7b9aa9d5"' ' ],' ' "product_name": "crypto",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "4051e7bd-73e2-518a-b930-d41a7b9aa9d5",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc1",' ' "name": "crypto_ram1",' ' "key_name": "test_dek_qat_xts"' ' }' ' }' '}' '{' ' "name": "crypto_ram2",' ' "aliases": [' ' "29182e97-bee7-5ae3-92fd-906aa52020db"' ' ],' ' "product_name": "crypto",' ' "block_size": 4096,' ' "num_blocks": 8192,' ' "uuid": "29182e97-bee7-5ae3-92fd-906aa52020db",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc2",' ' "name": "crypto_ram2",' ' "key_name": "test_dek_qat_cbc2"' ' }' ' }' '}' '{' ' "name": "crypto_ram3",' ' "aliases": [' ' "0b993534-dac3-5253-b723-3eff596cf724"' ' ],' ' "product_name": "crypto",' ' "block_size": 4096,' ' "num_blocks": 8192,' ' "uuid": "0b993534-dac3-5253-b723-3eff596cf724",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc3",' ' "name": "crypto_ram3",' ' "key_name": "test_dek_qat_xts2"' ' }' ' }' '}' 00:33:31.862 11:44:15 blockdev_crypto_qat -- bdev/blockdev.sh@750 -- # bdev_list=("${bdevs_name[@]}") 00:33:31.862 11:44:15 blockdev_crypto_qat -- bdev/blockdev.sh@752 -- # hello_world_bdev=crypto_ram 00:33:31.862 11:44:15 blockdev_crypto_qat -- bdev/blockdev.sh@753 -- # trap - SIGINT SIGTERM EXIT 00:33:31.862 11:44:15 blockdev_crypto_qat -- bdev/blockdev.sh@754 -- # killprocess 1060041 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@948 -- # '[' -z 1060041 ']' 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@952 -- # kill -0 1060041 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@953 -- # uname 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1060041 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1060041' 00:33:31.862 killing process with pid 1060041 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@967 -- # kill 1060041 00:33:31.862 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@972 -- # wait 1060041 00:33:32.431 11:44:15 blockdev_crypto_qat -- bdev/blockdev.sh@758 -- # trap cleanup SIGINT SIGTERM EXIT 00:33:32.431 11:44:15 blockdev_crypto_qat -- bdev/blockdev.sh@760 -- # run_test bdev_hello_world /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/hello_bdev --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -b crypto_ram '' 00:33:32.431 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:33:32.431 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@1105 -- # xtrace_disable 00:33:32.431 11:44:15 blockdev_crypto_qat -- common/autotest_common.sh@10 -- # set +x 00:33:32.431 ************************************ 00:33:32.431 START TEST bdev_hello_world 00:33:32.431 ************************************ 00:33:32.431 11:44:15 blockdev_crypto_qat.bdev_hello_world -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/hello_bdev --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -b crypto_ram '' 00:33:32.431 [2024-07-15 11:44:16.022872] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:33:32.431 [2024-07-15 11:44:16.022945] [ DPDK EAL parameters: hello_bdev --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1060679 ] 00:33:32.690 [2024-07-15 11:44:16.151146] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:32.690 [2024-07-15 11:44:16.250600] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:33:32.690 [2024-07-15 11:44:16.271953] accel_dpdk_cryptodev.c: 223:accel_dpdk_cryptodev_set_driver: *NOTICE*: Using driver crypto_qat 00:33:32.690 [2024-07-15 11:44:16.279980] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation encrypt will be assigned to module dpdk_cryptodev 00:33:32.949 [2024-07-15 11:44:16.288002] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation decrypt will be assigned to module dpdk_cryptodev 00:33:32.949 [2024-07-15 11:44:16.401216] accel_dpdk_cryptodev.c:1178:accel_dpdk_cryptodev_init: *NOTICE*: Found crypto devices: 96 00:33:35.483 [2024-07-15 11:44:18.608262] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_cbc" 00:33:35.483 [2024-07-15 11:44:18.608332] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:33:35.483 [2024-07-15 11:44:18.608348] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:35.483 [2024-07-15 11:44:18.616280] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_xts" 00:33:35.483 [2024-07-15 11:44:18.616301] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:33:35.483 [2024-07-15 11:44:18.616313] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:35.483 [2024-07-15 11:44:18.624300] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_cbc2" 00:33:35.483 [2024-07-15 11:44:18.624320] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:33:35.483 [2024-07-15 11:44:18.624332] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:35.483 [2024-07-15 11:44:18.632321] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_xts2" 00:33:35.483 [2024-07-15 11:44:18.632340] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc3 00:33:35.483 [2024-07-15 11:44:18.632352] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:35.483 [2024-07-15 11:44:18.710008] hello_bdev.c: 222:hello_start: *NOTICE*: Successfully started the application 00:33:35.483 [2024-07-15 11:44:18.710054] hello_bdev.c: 231:hello_start: *NOTICE*: Opening the bdev crypto_ram 00:33:35.483 [2024-07-15 11:44:18.710073] hello_bdev.c: 244:hello_start: *NOTICE*: Opening io channel 00:33:35.483 [2024-07-15 11:44:18.711354] hello_bdev.c: 138:hello_write: *NOTICE*: Writing to the bdev 00:33:35.483 [2024-07-15 11:44:18.711423] hello_bdev.c: 117:write_complete: *NOTICE*: bdev io write completed successfully 00:33:35.483 [2024-07-15 11:44:18.711440] hello_bdev.c: 84:hello_read: *NOTICE*: Reading io 00:33:35.483 [2024-07-15 11:44:18.711484] hello_bdev.c: 65:read_complete: *NOTICE*: Read string from bdev : Hello World! 00:33:35.483 00:33:35.483 [2024-07-15 11:44:18.711504] hello_bdev.c: 74:read_complete: *NOTICE*: Stopping app 00:33:35.742 00:33:35.742 real 0m3.156s 00:33:35.742 user 0m2.746s 00:33:35.742 sys 0m0.369s 00:33:35.742 11:44:19 blockdev_crypto_qat.bdev_hello_world -- common/autotest_common.sh@1124 -- # xtrace_disable 00:33:35.742 11:44:19 blockdev_crypto_qat.bdev_hello_world -- common/autotest_common.sh@10 -- # set +x 00:33:35.742 ************************************ 00:33:35.742 END TEST bdev_hello_world 00:33:35.742 ************************************ 00:33:35.742 11:44:19 blockdev_crypto_qat -- common/autotest_common.sh@1142 -- # return 0 00:33:35.742 11:44:19 blockdev_crypto_qat -- bdev/blockdev.sh@761 -- # run_test bdev_bounds bdev_bounds '' 00:33:35.742 11:44:19 blockdev_crypto_qat -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:33:35.742 11:44:19 blockdev_crypto_qat -- common/autotest_common.sh@1105 -- # xtrace_disable 00:33:35.742 11:44:19 blockdev_crypto_qat -- common/autotest_common.sh@10 -- # set +x 00:33:35.742 ************************************ 00:33:35.742 START TEST bdev_bounds 00:33:35.742 ************************************ 00:33:35.742 11:44:19 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@1123 -- # bdev_bounds '' 00:33:35.742 11:44:19 blockdev_crypto_qat.bdev_bounds -- bdev/blockdev.sh@290 -- # bdevio_pid=1061109 00:33:35.742 11:44:19 blockdev_crypto_qat.bdev_bounds -- bdev/blockdev.sh@291 -- # trap 'cleanup; killprocess $bdevio_pid; exit 1' SIGINT SIGTERM EXIT 00:33:35.742 11:44:19 blockdev_crypto_qat.bdev_bounds -- bdev/blockdev.sh@292 -- # echo 'Process bdevio pid: 1061109' 00:33:35.742 Process bdevio pid: 1061109 00:33:35.742 11:44:19 blockdev_crypto_qat.bdev_bounds -- bdev/blockdev.sh@293 -- # waitforlisten 1061109 00:33:35.742 11:44:19 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@829 -- # '[' -z 1061109 ']' 00:33:35.742 11:44:19 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:35.742 11:44:19 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@834 -- # local max_retries=100 00:33:35.742 11:44:19 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:35.742 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:35.742 11:44:19 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@838 -- # xtrace_disable 00:33:35.742 11:44:19 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:33:35.742 11:44:19 blockdev_crypto_qat.bdev_bounds -- bdev/blockdev.sh@289 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevio/bdevio -w -s 0 --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json '' 00:33:35.742 [2024-07-15 11:44:19.258474] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:33:35.742 [2024-07-15 11:44:19.258539] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 0 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1061109 ] 00:33:36.000 [2024-07-15 11:44:19.387532] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:33:36.000 [2024-07-15 11:44:19.496721] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:33:36.000 [2024-07-15 11:44:19.496808] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:33:36.000 [2024-07-15 11:44:19.496813] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:33:36.000 [2024-07-15 11:44:19.518190] accel_dpdk_cryptodev.c: 223:accel_dpdk_cryptodev_set_driver: *NOTICE*: Using driver crypto_qat 00:33:36.000 [2024-07-15 11:44:19.526214] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation encrypt will be assigned to module dpdk_cryptodev 00:33:36.000 [2024-07-15 11:44:19.534237] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation decrypt will be assigned to module dpdk_cryptodev 00:33:36.259 [2024-07-15 11:44:19.636649] accel_dpdk_cryptodev.c:1178:accel_dpdk_cryptodev_init: *NOTICE*: Found crypto devices: 96 00:33:38.795 [2024-07-15 11:44:21.846882] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_cbc" 00:33:38.795 [2024-07-15 11:44:21.846961] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:33:38.795 [2024-07-15 11:44:21.846977] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:38.795 [2024-07-15 11:44:21.854898] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_xts" 00:33:38.795 [2024-07-15 11:44:21.854919] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:33:38.795 [2024-07-15 11:44:21.854937] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:38.795 [2024-07-15 11:44:21.862920] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_cbc2" 00:33:38.795 [2024-07-15 11:44:21.862944] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:33:38.795 [2024-07-15 11:44:21.862956] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:38.795 [2024-07-15 11:44:21.870952] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_xts2" 00:33:38.795 [2024-07-15 11:44:21.870971] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc3 00:33:38.795 [2024-07-15 11:44:21.870982] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:38.795 11:44:21 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:33:38.795 11:44:21 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@862 -- # return 0 00:33:38.795 11:44:21 blockdev_crypto_qat.bdev_bounds -- bdev/blockdev.sh@294 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdevio/tests.py perform_tests 00:33:38.795 I/O targets: 00:33:38.795 crypto_ram: 65536 blocks of 512 bytes (32 MiB) 00:33:38.795 crypto_ram1: 65536 blocks of 512 bytes (32 MiB) 00:33:38.795 crypto_ram2: 8192 blocks of 4096 bytes (32 MiB) 00:33:38.795 crypto_ram3: 8192 blocks of 4096 bytes (32 MiB) 00:33:38.795 00:33:38.795 00:33:38.795 CUnit - A unit testing framework for C - Version 2.1-3 00:33:38.795 http://cunit.sourceforge.net/ 00:33:38.795 00:33:38.795 00:33:38.795 Suite: bdevio tests on: crypto_ram3 00:33:38.795 Test: blockdev write read block ...passed 00:33:38.795 Test: blockdev write zeroes read block ...passed 00:33:38.795 Test: blockdev write zeroes read no split ...passed 00:33:38.795 Test: blockdev write zeroes read split ...passed 00:33:38.795 Test: blockdev write zeroes read split partial ...passed 00:33:38.795 Test: blockdev reset ...passed 00:33:38.795 Test: blockdev write read 8 blocks ...passed 00:33:38.795 Test: blockdev write read size > 128k ...passed 00:33:38.795 Test: blockdev write read invalid size ...passed 00:33:38.795 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:33:38.795 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:33:38.795 Test: blockdev write read max offset ...passed 00:33:38.795 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:33:38.795 Test: blockdev writev readv 8 blocks ...passed 00:33:38.795 Test: blockdev writev readv 30 x 1block ...passed 00:33:38.795 Test: blockdev writev readv block ...passed 00:33:38.795 Test: blockdev writev readv size > 128k ...passed 00:33:38.795 Test: blockdev writev readv size > 128k in two iovs ...passed 00:33:38.795 Test: blockdev comparev and writev ...passed 00:33:38.795 Test: blockdev nvme passthru rw ...passed 00:33:38.795 Test: blockdev nvme passthru vendor specific ...passed 00:33:38.795 Test: blockdev nvme admin passthru ...passed 00:33:38.795 Test: blockdev copy ...passed 00:33:38.795 Suite: bdevio tests on: crypto_ram2 00:33:38.795 Test: blockdev write read block ...passed 00:33:38.795 Test: blockdev write zeroes read block ...passed 00:33:38.795 Test: blockdev write zeroes read no split ...passed 00:33:38.795 Test: blockdev write zeroes read split ...passed 00:33:38.795 Test: blockdev write zeroes read split partial ...passed 00:33:38.795 Test: blockdev reset ...passed 00:33:38.795 Test: blockdev write read 8 blocks ...passed 00:33:38.795 Test: blockdev write read size > 128k ...passed 00:33:38.795 Test: blockdev write read invalid size ...passed 00:33:38.795 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:33:38.795 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:33:38.795 Test: blockdev write read max offset ...passed 00:33:38.795 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:33:38.795 Test: blockdev writev readv 8 blocks ...passed 00:33:38.795 Test: blockdev writev readv 30 x 1block ...passed 00:33:38.795 Test: blockdev writev readv block ...passed 00:33:38.795 Test: blockdev writev readv size > 128k ...passed 00:33:38.795 Test: blockdev writev readv size > 128k in two iovs ...passed 00:33:38.795 Test: blockdev comparev and writev ...passed 00:33:38.795 Test: blockdev nvme passthru rw ...passed 00:33:38.795 Test: blockdev nvme passthru vendor specific ...passed 00:33:38.795 Test: blockdev nvme admin passthru ...passed 00:33:38.795 Test: blockdev copy ...passed 00:33:38.795 Suite: bdevio tests on: crypto_ram1 00:33:38.795 Test: blockdev write read block ...passed 00:33:38.795 Test: blockdev write zeroes read block ...passed 00:33:38.795 Test: blockdev write zeroes read no split ...passed 00:33:38.795 Test: blockdev write zeroes read split ...passed 00:33:38.795 Test: blockdev write zeroes read split partial ...passed 00:33:38.795 Test: blockdev reset ...passed 00:33:38.795 Test: blockdev write read 8 blocks ...passed 00:33:38.795 Test: blockdev write read size > 128k ...passed 00:33:38.795 Test: blockdev write read invalid size ...passed 00:33:38.795 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:33:38.795 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:33:38.795 Test: blockdev write read max offset ...passed 00:33:38.795 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:33:38.795 Test: blockdev writev readv 8 blocks ...passed 00:33:38.795 Test: blockdev writev readv 30 x 1block ...passed 00:33:38.795 Test: blockdev writev readv block ...passed 00:33:38.795 Test: blockdev writev readv size > 128k ...passed 00:33:38.795 Test: blockdev writev readv size > 128k in two iovs ...passed 00:33:38.795 Test: blockdev comparev and writev ...passed 00:33:38.795 Test: blockdev nvme passthru rw ...passed 00:33:38.795 Test: blockdev nvme passthru vendor specific ...passed 00:33:38.795 Test: blockdev nvme admin passthru ...passed 00:33:38.795 Test: blockdev copy ...passed 00:33:38.795 Suite: bdevio tests on: crypto_ram 00:33:38.795 Test: blockdev write read block ...passed 00:33:38.795 Test: blockdev write zeroes read block ...passed 00:33:38.795 Test: blockdev write zeroes read no split ...passed 00:33:38.795 Test: blockdev write zeroes read split ...passed 00:33:38.795 Test: blockdev write zeroes read split partial ...passed 00:33:38.795 Test: blockdev reset ...passed 00:33:38.795 Test: blockdev write read 8 blocks ...passed 00:33:38.795 Test: blockdev write read size > 128k ...passed 00:33:38.795 Test: blockdev write read invalid size ...passed 00:33:38.795 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:33:38.795 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:33:38.795 Test: blockdev write read max offset ...passed 00:33:38.795 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:33:38.795 Test: blockdev writev readv 8 blocks ...passed 00:33:38.795 Test: blockdev writev readv 30 x 1block ...passed 00:33:38.795 Test: blockdev writev readv block ...passed 00:33:38.795 Test: blockdev writev readv size > 128k ...passed 00:33:38.795 Test: blockdev writev readv size > 128k in two iovs ...passed 00:33:38.795 Test: blockdev comparev and writev ...passed 00:33:38.795 Test: blockdev nvme passthru rw ...passed 00:33:38.795 Test: blockdev nvme passthru vendor specific ...passed 00:33:38.795 Test: blockdev nvme admin passthru ...passed 00:33:38.795 Test: blockdev copy ...passed 00:33:38.795 00:33:38.795 Run Summary: Type Total Ran Passed Failed Inactive 00:33:38.795 suites 4 4 n/a 0 0 00:33:38.795 tests 92 92 92 0 0 00:33:38.795 asserts 520 520 520 0 n/a 00:33:38.795 00:33:38.795 Elapsed time = 0.520 seconds 00:33:38.795 0 00:33:38.795 11:44:22 blockdev_crypto_qat.bdev_bounds -- bdev/blockdev.sh@295 -- # killprocess 1061109 00:33:38.796 11:44:22 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@948 -- # '[' -z 1061109 ']' 00:33:38.796 11:44:22 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@952 -- # kill -0 1061109 00:33:38.796 11:44:22 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@953 -- # uname 00:33:38.796 11:44:22 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:33:38.796 11:44:22 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1061109 00:33:39.055 11:44:22 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:33:39.055 11:44:22 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:33:39.055 11:44:22 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1061109' 00:33:39.055 killing process with pid 1061109 00:33:39.055 11:44:22 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@967 -- # kill 1061109 00:33:39.055 11:44:22 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@972 -- # wait 1061109 00:33:39.314 11:44:22 blockdev_crypto_qat.bdev_bounds -- bdev/blockdev.sh@296 -- # trap - SIGINT SIGTERM EXIT 00:33:39.314 00:33:39.315 real 0m3.618s 00:33:39.315 user 0m10.098s 00:33:39.315 sys 0m0.572s 00:33:39.315 11:44:22 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@1124 -- # xtrace_disable 00:33:39.315 11:44:22 blockdev_crypto_qat.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:33:39.315 ************************************ 00:33:39.315 END TEST bdev_bounds 00:33:39.315 ************************************ 00:33:39.315 11:44:22 blockdev_crypto_qat -- common/autotest_common.sh@1142 -- # return 0 00:33:39.315 11:44:22 blockdev_crypto_qat -- bdev/blockdev.sh@762 -- # run_test bdev_nbd nbd_function_test /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json 'crypto_ram crypto_ram1 crypto_ram2 crypto_ram3' '' 00:33:39.315 11:44:22 blockdev_crypto_qat -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:33:39.315 11:44:22 blockdev_crypto_qat -- common/autotest_common.sh@1105 -- # xtrace_disable 00:33:39.315 11:44:22 blockdev_crypto_qat -- common/autotest_common.sh@10 -- # set +x 00:33:39.315 ************************************ 00:33:39.315 START TEST bdev_nbd 00:33:39.315 ************************************ 00:33:39.315 11:44:22 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@1123 -- # nbd_function_test /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json 'crypto_ram crypto_ram1 crypto_ram2 crypto_ram3' '' 00:33:39.315 11:44:22 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@300 -- # uname -s 00:33:39.315 11:44:22 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@300 -- # [[ Linux == Linux ]] 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@302 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@303 -- # local conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@304 -- # bdev_all=('crypto_ram' 'crypto_ram1' 'crypto_ram2' 'crypto_ram3') 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@304 -- # local bdev_all 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@305 -- # local bdev_num=4 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@309 -- # [[ -e /sys/module/nbd ]] 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@311 -- # nbd_all=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@311 -- # local nbd_all 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@312 -- # bdev_num=4 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@314 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11') 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@314 -- # local nbd_list 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@315 -- # bdev_list=('crypto_ram' 'crypto_ram1' 'crypto_ram2' 'crypto_ram3') 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@315 -- # local bdev_list 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@318 -- # nbd_pid=1061610 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@319 -- # trap 'cleanup; killprocess $nbd_pid' SIGINT SIGTERM EXIT 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@317 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-nbd.sock -i 0 --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json '' 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@320 -- # waitforlisten 1061610 /var/tmp/spdk-nbd.sock 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@829 -- # '[' -z 1061610 ']' 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@834 -- # local max_retries=100 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:33:39.574 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@838 -- # xtrace_disable 00:33:39.574 11:44:22 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:33:39.574 [2024-07-15 11:44:22.973473] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:33:39.574 [2024-07-15 11:44:22.973542] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:33:39.574 [2024-07-15 11:44:23.105446] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:39.838 [2024-07-15 11:44:23.204755] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:33:39.838 [2024-07-15 11:44:23.226048] accel_dpdk_cryptodev.c: 223:accel_dpdk_cryptodev_set_driver: *NOTICE*: Using driver crypto_qat 00:33:39.838 [2024-07-15 11:44:23.234086] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation encrypt will be assigned to module dpdk_cryptodev 00:33:39.838 [2024-07-15 11:44:23.242087] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation decrypt will be assigned to module dpdk_cryptodev 00:33:39.838 [2024-07-15 11:44:23.346266] accel_dpdk_cryptodev.c:1178:accel_dpdk_cryptodev_init: *NOTICE*: Found crypto devices: 96 00:33:42.372 [2024-07-15 11:44:25.558771] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_cbc" 00:33:42.372 [2024-07-15 11:44:25.558839] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:33:42.372 [2024-07-15 11:44:25.558854] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:42.372 [2024-07-15 11:44:25.566790] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_xts" 00:33:42.372 [2024-07-15 11:44:25.566810] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:33:42.372 [2024-07-15 11:44:25.566823] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:42.372 [2024-07-15 11:44:25.574810] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_cbc2" 00:33:42.372 [2024-07-15 11:44:25.574828] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:33:42.372 [2024-07-15 11:44:25.574840] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:42.372 [2024-07-15 11:44:25.582830] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_xts2" 00:33:42.372 [2024-07-15 11:44:25.582849] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc3 00:33:42.372 [2024-07-15 11:44:25.582861] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@862 -- # return 0 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@322 -- # nbd_rpc_start_stop_verify /var/tmp/spdk-nbd.sock 'crypto_ram crypto_ram1 crypto_ram2 crypto_ram3' 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@113 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@114 -- # bdev_list=('crypto_ram' 'crypto_ram1' 'crypto_ram2' 'crypto_ram3') 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@114 -- # local bdev_list 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@116 -- # nbd_start_disks_without_nbd_idx /var/tmp/spdk-nbd.sock 'crypto_ram crypto_ram1 crypto_ram2 crypto_ram3' 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@22 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@23 -- # bdev_list=('crypto_ram' 'crypto_ram1' 'crypto_ram2' 'crypto_ram3') 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@23 -- # local bdev_list 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@24 -- # local i 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@25 -- # local nbd_device 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i = 0 )) 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 4 )) 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk crypto_ram 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd0 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd0 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd0 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:33:42.372 11:44:25 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:33:42.373 11:44:25 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:33:42.373 11:44:25 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:33:42.373 1+0 records in 00:33:42.373 1+0 records out 00:33:42.373 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000328715 s, 12.5 MB/s 00:33:42.373 11:44:25 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:33:42.632 11:44:25 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:33:42.632 11:44:25 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:33:42.632 11:44:25 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:33:42.632 11:44:25 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:33:42.632 11:44:25 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:33:42.632 11:44:25 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 4 )) 00:33:42.632 11:44:25 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk crypto_ram1 00:33:42.632 11:44:26 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd1 00:33:42.632 11:44:26 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd1 00:33:42.632 11:44:26 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd1 00:33:42.632 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:33:42.632 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:33:42.632 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:33:42.632 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:33:42.632 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:33:42.891 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:33:42.891 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:33:42.891 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:33:42.891 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:33:42.891 1+0 records in 00:33:42.891 1+0 records out 00:33:42.891 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000321061 s, 12.8 MB/s 00:33:42.891 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:33:42.891 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:33:42.891 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:33:42.891 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:33:42.891 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:33:42.891 11:44:26 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:33:42.891 11:44:26 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 4 )) 00:33:42.891 11:44:26 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk crypto_ram2 00:33:43.153 11:44:26 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd2 00:33:43.153 11:44:26 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd2 00:33:43.153 11:44:26 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd2 00:33:43.153 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd2 00:33:43.153 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:33:43.153 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:33:43.153 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:33:43.153 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd2 /proc/partitions 00:33:43.153 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:33:43.153 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:33:43.153 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:33:43.153 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd2 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:33:43.153 1+0 records in 00:33:43.153 1+0 records out 00:33:43.153 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000367683 s, 11.1 MB/s 00:33:43.153 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:33:43.153 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:33:43.153 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:33:43.153 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:33:43.153 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:33:43.153 11:44:26 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:33:43.153 11:44:26 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 4 )) 00:33:43.153 11:44:26 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@28 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk crypto_ram3 00:33:43.479 11:44:26 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd3 00:33:43.479 11:44:26 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd3 00:33:43.479 11:44:26 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd3 00:33:43.479 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd3 00:33:43.479 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:33:43.479 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:33:43.479 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:33:43.479 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd3 /proc/partitions 00:33:43.479 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:33:43.479 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:33:43.479 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:33:43.479 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd3 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:33:43.479 1+0 records in 00:33:43.479 1+0 records out 00:33:43.479 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000351291 s, 11.7 MB/s 00:33:43.479 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:33:43.479 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:33:43.479 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:33:43.479 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:33:43.479 11:44:26 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:33:43.479 11:44:26 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:33:43.479 11:44:26 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 4 )) 00:33:43.479 11:44:26 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@118 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:33:43.479 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@118 -- # nbd_disks_json='[ 00:33:43.479 { 00:33:43.479 "nbd_device": "/dev/nbd0", 00:33:43.479 "bdev_name": "crypto_ram" 00:33:43.479 }, 00:33:43.479 { 00:33:43.479 "nbd_device": "/dev/nbd1", 00:33:43.479 "bdev_name": "crypto_ram1" 00:33:43.479 }, 00:33:43.479 { 00:33:43.479 "nbd_device": "/dev/nbd2", 00:33:43.479 "bdev_name": "crypto_ram2" 00:33:43.479 }, 00:33:43.479 { 00:33:43.479 "nbd_device": "/dev/nbd3", 00:33:43.479 "bdev_name": "crypto_ram3" 00:33:43.479 } 00:33:43.479 ]' 00:33:43.479 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@119 -- # nbd_disks_name=($(echo "${nbd_disks_json}" | jq -r '.[] | .nbd_device')) 00:33:43.479 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@119 -- # echo '[ 00:33:43.479 { 00:33:43.479 "nbd_device": "/dev/nbd0", 00:33:43.479 "bdev_name": "crypto_ram" 00:33:43.479 }, 00:33:43.479 { 00:33:43.479 "nbd_device": "/dev/nbd1", 00:33:43.479 "bdev_name": "crypto_ram1" 00:33:43.479 }, 00:33:43.479 { 00:33:43.479 "nbd_device": "/dev/nbd2", 00:33:43.479 "bdev_name": "crypto_ram2" 00:33:43.479 }, 00:33:43.479 { 00:33:43.479 "nbd_device": "/dev/nbd3", 00:33:43.479 "bdev_name": "crypto_ram3" 00:33:43.479 } 00:33:43.479 ]' 00:33:43.479 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@119 -- # jq -r '.[] | .nbd_device' 00:33:43.479 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@120 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1 /dev/nbd2 /dev/nbd3' 00:33:43.479 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:33:43.479 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd2' '/dev/nbd3') 00:33:43.479 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:33:43.479 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:33:43.479 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:33:43.479 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:33:43.738 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:33:43.738 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:33:43.738 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:33:43.738 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:33:43.738 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:33:43.738 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:33:43.738 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:33:43.738 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:33:43.738 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:33:43.738 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:33:43.997 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:33:43.997 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:33:43.997 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:33:43.998 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:33:43.998 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:33:43.998 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:33:43.998 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:33:43.998 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:33:43.998 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:33:43.998 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd2 00:33:44.256 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd2 00:33:44.516 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd2 00:33:44.516 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd2 00:33:44.516 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:33:44.516 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:33:44.516 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd2 /proc/partitions 00:33:44.516 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:33:44.516 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:33:44.516 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:33:44.516 11:44:27 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd3 00:33:44.516 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd3 00:33:44.516 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd3 00:33:44.516 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd3 00:33:44.516 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:33:44.516 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:33:44.516 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd3 /proc/partitions 00:33:44.516 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:33:44.516 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:33:44.516 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@122 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:33:44.516 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:33:44.516 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@122 -- # count=0 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@123 -- # '[' 0 -ne 0 ']' 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@127 -- # return 0 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@323 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'crypto_ram crypto_ram1 crypto_ram2 crypto_ram3' '/dev/nbd0 /dev/nbd1 /dev/nbd10 /dev/nbd11' 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@91 -- # bdev_list=('crypto_ram' 'crypto_ram1' 'crypto_ram2' 'crypto_ram3') 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@91 -- # local bdev_list 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11') 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@92 -- # local nbd_list 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'crypto_ram crypto_ram1 crypto_ram2 crypto_ram3' '/dev/nbd0 /dev/nbd1 /dev/nbd10 /dev/nbd11' 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@10 -- # bdev_list=('crypto_ram' 'crypto_ram1' 'crypto_ram2' 'crypto_ram3') 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@10 -- # local bdev_list 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11') 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@11 -- # local nbd_list 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@12 -- # local i 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 4 )) 00:33:44.775 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk crypto_ram /dev/nbd0 00:33:45.035 /dev/nbd0 00:33:45.035 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:33:45.035 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:33:45.035 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:33:45.035 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:33:45.035 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:33:45.035 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:33:45.035 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:33:45.035 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:33:45.035 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:33:45.035 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:33:45.035 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:33:45.035 1+0 records in 00:33:45.035 1+0 records out 00:33:45.035 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000287962 s, 14.2 MB/s 00:33:45.035 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:33:45.035 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:33:45.035 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:33:45.035 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:33:45.035 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:33:45.035 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:33:45.035 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 4 )) 00:33:45.035 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk crypto_ram1 /dev/nbd1 00:33:45.293 /dev/nbd1 00:33:45.293 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:33:45.293 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:33:45.294 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:33:45.294 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:33:45.294 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:33:45.294 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:33:45.294 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:33:45.294 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:33:45.294 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:33:45.294 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:33:45.294 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:33:45.294 1+0 records in 00:33:45.294 1+0 records out 00:33:45.294 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000311129 s, 13.2 MB/s 00:33:45.294 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:33:45.294 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:33:45.294 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:33:45.294 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:33:45.294 11:44:28 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:33:45.294 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:33:45.294 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 4 )) 00:33:45.294 11:44:28 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk crypto_ram2 /dev/nbd10 00:33:45.553 /dev/nbd10 00:33:45.553 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd10 00:33:45.553 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd10 00:33:45.553 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd10 00:33:45.553 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:33:45.553 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:33:45.553 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:33:45.553 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd10 /proc/partitions 00:33:45.553 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:33:45.553 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:33:45.553 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:33:45.553 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd10 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:33:45.553 1+0 records in 00:33:45.553 1+0 records out 00:33:45.553 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000379665 s, 10.8 MB/s 00:33:45.553 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:33:45.553 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:33:45.553 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:33:45.553 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:33:45.553 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:33:45.553 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:33:45.553 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 4 )) 00:33:45.553 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk crypto_ram3 /dev/nbd11 00:33:45.811 /dev/nbd11 00:33:45.811 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd11 00:33:46.070 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd11 00:33:46.070 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@866 -- # local nbd_name=nbd11 00:33:46.070 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@867 -- # local i 00:33:46.070 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:33:46.070 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:33:46.070 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@870 -- # grep -q -w nbd11 /proc/partitions 00:33:46.070 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@871 -- # break 00:33:46.070 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:33:46.070 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:33:46.070 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@883 -- # dd if=/dev/nbd11 of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:33:46.070 1+0 records in 00:33:46.070 1+0 records out 00:33:46.070 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000369851 s, 11.1 MB/s 00:33:46.070 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:33:46.070 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@884 -- # size=4096 00:33:46.070 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdtest 00:33:46.070 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:33:46.070 11:44:29 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@887 -- # return 0 00:33:46.070 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:33:46.070 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 4 )) 00:33:46.070 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:33:46.070 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:33:46.070 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:33:46.330 { 00:33:46.330 "nbd_device": "/dev/nbd0", 00:33:46.330 "bdev_name": "crypto_ram" 00:33:46.330 }, 00:33:46.330 { 00:33:46.330 "nbd_device": "/dev/nbd1", 00:33:46.330 "bdev_name": "crypto_ram1" 00:33:46.330 }, 00:33:46.330 { 00:33:46.330 "nbd_device": "/dev/nbd10", 00:33:46.330 "bdev_name": "crypto_ram2" 00:33:46.330 }, 00:33:46.330 { 00:33:46.330 "nbd_device": "/dev/nbd11", 00:33:46.330 "bdev_name": "crypto_ram3" 00:33:46.330 } 00:33:46.330 ]' 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[ 00:33:46.330 { 00:33:46.330 "nbd_device": "/dev/nbd0", 00:33:46.330 "bdev_name": "crypto_ram" 00:33:46.330 }, 00:33:46.330 { 00:33:46.330 "nbd_device": "/dev/nbd1", 00:33:46.330 "bdev_name": "crypto_ram1" 00:33:46.330 }, 00:33:46.330 { 00:33:46.330 "nbd_device": "/dev/nbd10", 00:33:46.330 "bdev_name": "crypto_ram2" 00:33:46.330 }, 00:33:46.330 { 00:33:46.330 "nbd_device": "/dev/nbd11", 00:33:46.330 "bdev_name": "crypto_ram3" 00:33:46.330 } 00:33:46.330 ]' 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:33:46.330 /dev/nbd1 00:33:46.330 /dev/nbd10 00:33:46.330 /dev/nbd11' 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:33:46.330 /dev/nbd1 00:33:46.330 /dev/nbd10 00:33:46.330 /dev/nbd11' 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=4 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 4 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@95 -- # count=4 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@96 -- # '[' 4 -ne 4 ']' 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1 /dev/nbd10 /dev/nbd11' write 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11') 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=write 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest bs=4096 count=256 00:33:46.330 256+0 records in 00:33:46.330 256+0 records out 00:33:46.330 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0104968 s, 99.9 MB/s 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:33:46.330 256+0 records in 00:33:46.330 256+0 records out 00:33:46.330 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0793245 s, 13.2 MB/s 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:33:46.330 256+0 records in 00:33:46.330 256+0 records out 00:33:46.330 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0659213 s, 15.9 MB/s 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:33:46.330 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd10 bs=4096 count=256 oflag=direct 00:33:46.590 256+0 records in 00:33:46.590 256+0 records out 00:33:46.590 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0580628 s, 18.1 MB/s 00:33:46.590 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:33:46.590 11:44:29 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest of=/dev/nbd11 bs=4096 count=256 oflag=direct 00:33:46.590 256+0 records in 00:33:46.590 256+0 records out 00:33:46.590 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0555518 s, 18.9 MB/s 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1 /dev/nbd10 /dev/nbd11' verify 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11') 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=verify 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd0 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd1 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd10 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest /dev/nbd11 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nbdrandtest 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1 /dev/nbd10 /dev/nbd11' 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11') 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:33:46.590 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:33:46.849 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:33:46.849 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:33:46.849 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:33:46.849 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:33:46.849 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:33:46.849 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:33:46.849 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:33:46.849 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:33:46.849 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:33:46.849 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:33:47.108 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:33:47.108 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:33:47.108 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:33:47.108 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:33:47.108 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:33:47.108 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:33:47.108 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:33:47.108 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:33:47.108 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:33:47.108 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd10 00:33:47.367 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd10 00:33:47.367 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd10 00:33:47.367 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd10 00:33:47.367 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:33:47.367 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:33:47.367 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd10 /proc/partitions 00:33:47.367 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:33:47.367 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:33:47.367 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:33:47.367 11:44:30 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd11 00:33:47.626 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd11 00:33:47.626 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd11 00:33:47.626 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd11 00:33:47.626 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:33:47.626 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:33:47.626 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd11 /proc/partitions 00:33:47.626 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:33:47.626 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:33:47.626 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:33:47.626 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:33:47.626 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:33:47.886 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:33:47.886 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:33:47.886 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:33:47.886 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:33:47.886 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:33:47.886 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:33:47.886 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:33:47.886 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:33:47.886 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:33:47.886 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@104 -- # count=0 00:33:47.886 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:33:47.886 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@109 -- # return 0 00:33:47.886 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@324 -- # nbd_with_lvol_verify /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1 /dev/nbd10 /dev/nbd11' 00:33:47.886 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@131 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:33:47.886 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@132 -- # nbd_list=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11') 00:33:47.886 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@132 -- # local nbd_list 00:33:47.886 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@133 -- # local mkfs_ret 00:33:47.886 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@135 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create -b malloc_lvol_verify 16 512 00:33:48.145 malloc_lvol_verify 00:33:48.145 11:44:31 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@136 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create_lvstore malloc_lvol_verify lvs 00:33:48.713 9a53f2f4-2d43-48b9-8b45-75f96d9346a2 00:33:48.713 11:44:32 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@137 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create lvol 4 -l lvs 00:33:48.972 6afa7fe3-9c08-402c-8c8f-4f7e947924d0 00:33:48.973 11:44:32 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@138 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk lvs/lvol /dev/nbd0 00:33:49.539 /dev/nbd0 00:33:49.539 11:44:32 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@140 -- # mkfs.ext4 /dev/nbd0 00:33:49.539 mke2fs 1.46.5 (30-Dec-2021) 00:33:49.539 Discarding device blocks: 0/4096 done 00:33:49.539 Creating filesystem with 4096 1k blocks and 1024 inodes 00:33:49.539 00:33:49.539 Allocating group tables: 0/1 done 00:33:49.539 Writing inode tables: 0/1 done 00:33:49.539 Creating journal (1024 blocks): done 00:33:49.539 Writing superblocks and filesystem accounting information: 0/1 done 00:33:49.539 00:33:49.539 11:44:32 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@141 -- # mkfs_ret=0 00:33:49.539 11:44:32 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@142 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:33:49.539 11:44:32 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:33:49.539 11:44:32 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:33:49.539 11:44:32 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:33:49.539 11:44:32 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:33:49.539 11:44:32 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:33:49.539 11:44:32 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@143 -- # '[' 0 -ne 0 ']' 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- bdev/nbd_common.sh@147 -- # return 0 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@326 -- # killprocess 1061610 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@948 -- # '[' -z 1061610 ']' 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@952 -- # kill -0 1061610 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@953 -- # uname 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1061610 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1061610' 00:33:49.799 killing process with pid 1061610 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@967 -- # kill 1061610 00:33:49.799 11:44:33 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@972 -- # wait 1061610 00:33:50.365 11:44:33 blockdev_crypto_qat.bdev_nbd -- bdev/blockdev.sh@327 -- # trap - SIGINT SIGTERM EXIT 00:33:50.365 00:33:50.365 real 0m10.847s 00:33:50.365 user 0m14.330s 00:33:50.365 sys 0m4.246s 00:33:50.365 11:44:33 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@1124 -- # xtrace_disable 00:33:50.365 11:44:33 blockdev_crypto_qat.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:33:50.365 ************************************ 00:33:50.365 END TEST bdev_nbd 00:33:50.365 ************************************ 00:33:50.365 11:44:33 blockdev_crypto_qat -- common/autotest_common.sh@1142 -- # return 0 00:33:50.365 11:44:33 blockdev_crypto_qat -- bdev/blockdev.sh@763 -- # [[ y == y ]] 00:33:50.365 11:44:33 blockdev_crypto_qat -- bdev/blockdev.sh@764 -- # '[' crypto_qat = nvme ']' 00:33:50.366 11:44:33 blockdev_crypto_qat -- bdev/blockdev.sh@764 -- # '[' crypto_qat = gpt ']' 00:33:50.366 11:44:33 blockdev_crypto_qat -- bdev/blockdev.sh@768 -- # run_test bdev_fio fio_test_suite '' 00:33:50.366 11:44:33 blockdev_crypto_qat -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:33:50.366 11:44:33 blockdev_crypto_qat -- common/autotest_common.sh@1105 -- # xtrace_disable 00:33:50.366 11:44:33 blockdev_crypto_qat -- common/autotest_common.sh@10 -- # set +x 00:33:50.366 ************************************ 00:33:50.366 START TEST bdev_fio 00:33:50.366 ************************************ 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1123 -- # fio_test_suite '' 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@331 -- # local env_context 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@335 -- # pushd /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev 00:33:50.366 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev /var/jenkins/workspace/crypto-phy-autotest/spdk 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@336 -- # trap 'rm -f ./*.state; popd; exit 1' SIGINT SIGTERM EXIT 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@339 -- # echo '' 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@339 -- # sed s/--env-context=// 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@339 -- # env_context= 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@340 -- # fio_config_gen /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio verify AIO '' 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1280 -- # local config_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1281 -- # local workload=verify 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1282 -- # local bdev_type=AIO 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1283 -- # local env_context= 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1284 -- # local fio_dir=/usr/src/fio 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1286 -- # '[' -e /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio ']' 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1291 -- # '[' -z verify ']' 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -n '' ']' 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1299 -- # touch /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1301 -- # cat 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1313 -- # '[' verify == verify ']' 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1314 -- # cat 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1323 -- # '[' AIO == AIO ']' 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1324 -- # /usr/src/fio/fio --version 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1324 -- # [[ fio-3.35 == *\f\i\o\-\3* ]] 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1325 -- # echo serialize_overlap=1 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_crypto_ram]' 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=crypto_ram 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_crypto_ram1]' 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=crypto_ram1 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_crypto_ram2]' 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=crypto_ram2 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@341 -- # for b in "${bdevs_name[@]}" 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@342 -- # echo '[job_crypto_ram3]' 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@343 -- # echo filename=crypto_ram3 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@347 -- # local 'fio_params=--ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json' 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@349 -- # run_test bdev_fio_rw_verify fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1099 -- # '[' 11 -le 1 ']' 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1105 -- # xtrace_disable 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:33:50.366 ************************************ 00:33:50.366 START TEST bdev_fio_rw_verify 00:33:50.366 ************************************ 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1123 -- # fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1339 -- # local sanitizers 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1341 -- # shift 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # local asan_lib= 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # grep libasan 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # asan_lib= 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:33:50.366 11:44:33 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:33:50.627 11:44:33 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:33:50.627 11:44:33 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:33:50.627 11:44:33 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:33:50.627 11:44:34 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # asan_lib= 00:33:50.627 11:44:34 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:33:50.627 11:44:34 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev' 00:33:50.627 11:44:34 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:33:50.886 job_crypto_ram: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:33:50.886 job_crypto_ram1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:33:50.886 job_crypto_ram2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:33:50.886 job_crypto_ram3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:33:50.886 fio-3.35 00:33:50.886 Starting 4 threads 00:34:05.851 00:34:05.851 job_crypto_ram: (groupid=0, jobs=4): err= 0: pid=1063693: Mon Jul 15 11:44:47 2024 00:34:05.851 read: IOPS=20.5k, BW=80.0MiB/s (83.8MB/s)(800MiB/10001msec) 00:34:05.851 slat (usec): min=17, max=413, avg=68.26, stdev=37.31 00:34:05.851 clat (usec): min=21, max=1683, avg=367.17, stdev=233.45 00:34:05.851 lat (usec): min=39, max=1863, avg=435.43, stdev=253.37 00:34:05.851 clat percentiles (usec): 00:34:05.851 | 50.000th=[ 306], 99.000th=[ 1139], 99.900th=[ 1500], 99.990th=[ 1598], 00:34:05.851 | 99.999th=[ 1647] 00:34:05.851 write: IOPS=22.6k, BW=88.1MiB/s (92.4MB/s)(859MiB/9751msec); 0 zone resets 00:34:05.851 slat (usec): min=27, max=383, avg=80.14, stdev=36.90 00:34:05.851 clat (usec): min=28, max=2399, avg=408.40, stdev=245.80 00:34:05.851 lat (usec): min=79, max=2581, avg=488.54, stdev=265.29 00:34:05.851 clat percentiles (usec): 00:34:05.851 | 50.000th=[ 359], 99.000th=[ 1237], 99.900th=[ 1680], 99.990th=[ 1795], 00:34:05.851 | 99.999th=[ 2057] 00:34:05.851 bw ( KiB/s): min=65232, max=115056, per=97.47%, avg=87925.05, stdev=3730.20, samples=76 00:34:05.851 iops : min=16308, max=28764, avg=21981.26, stdev=932.55, samples=76 00:34:05.851 lat (usec) : 50=0.01%, 100=3.30%, 250=30.23%, 500=39.76%, 750=18.86% 00:34:05.851 lat (usec) : 1000=4.96% 00:34:05.851 lat (msec) : 2=2.88%, 4=0.01% 00:34:05.851 cpu : usr=99.60%, sys=0.00%, ctx=56, majf=0, minf=274 00:34:05.851 IO depths : 1=5.5%, 2=27.0%, 4=54.0%, 8=13.5%, 16=0.0%, 32=0.0%, >=64=0.0% 00:34:05.851 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:34:05.851 complete : 0=0.0%, 4=88.1%, 8=11.9%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:34:05.851 issued rwts: total=204727,219909,0,0 short=0,0,0,0 dropped=0,0,0,0 00:34:05.851 latency : target=0, window=0, percentile=100.00%, depth=8 00:34:05.851 00:34:05.851 Run status group 0 (all jobs): 00:34:05.851 READ: bw=80.0MiB/s (83.8MB/s), 80.0MiB/s-80.0MiB/s (83.8MB/s-83.8MB/s), io=800MiB (839MB), run=10001-10001msec 00:34:05.851 WRITE: bw=88.1MiB/s (92.4MB/s), 88.1MiB/s-88.1MiB/s (92.4MB/s-92.4MB/s), io=859MiB (901MB), run=9751-9751msec 00:34:05.851 00:34:05.851 real 0m13.538s 00:34:05.851 user 0m45.738s 00:34:05.851 sys 0m0.519s 00:34:05.851 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1124 -- # xtrace_disable 00:34:05.851 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@10 -- # set +x 00:34:05.851 ************************************ 00:34:05.851 END TEST bdev_fio_rw_verify 00:34:05.851 ************************************ 00:34:05.851 11:44:47 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1142 -- # return 0 00:34:05.851 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@350 -- # rm -f 00:34:05.851 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@351 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:34:05.851 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@354 -- # fio_config_gen /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio trim '' '' 00:34:05.851 11:44:47 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1280 -- # local config_file=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:34:05.851 11:44:47 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1281 -- # local workload=trim 00:34:05.851 11:44:47 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1282 -- # local bdev_type= 00:34:05.851 11:44:47 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1283 -- # local env_context= 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1284 -- # local fio_dir=/usr/src/fio 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1286 -- # '[' -e /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio ']' 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1291 -- # '[' -z trim ']' 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -n '' ']' 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1299 -- # touch /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1301 -- # cat 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1313 -- # '[' trim == verify ']' 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1328 -- # '[' trim == trim ']' 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1329 -- # echo rw=trimwrite 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@355 -- # jq -r 'select(.supported_io_types.unmap == true) | .name' 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@355 -- # printf '%s\n' '{' ' "name": "crypto_ram",' ' "aliases": [' ' "df965139-bb3b-5c62-b107-9ca32835792e"' ' ],' ' "product_name": "crypto",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "df965139-bb3b-5c62-b107-9ca32835792e",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc0",' ' "name": "crypto_ram",' ' "key_name": "test_dek_qat_cbc"' ' }' ' }' '}' '{' ' "name": "crypto_ram1",' ' "aliases": [' ' "4051e7bd-73e2-518a-b930-d41a7b9aa9d5"' ' ],' ' "product_name": "crypto",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "4051e7bd-73e2-518a-b930-d41a7b9aa9d5",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc1",' ' "name": "crypto_ram1",' ' "key_name": "test_dek_qat_xts"' ' }' ' }' '}' '{' ' "name": "crypto_ram2",' ' "aliases": [' ' "29182e97-bee7-5ae3-92fd-906aa52020db"' ' ],' ' "product_name": "crypto",' ' "block_size": 4096,' ' "num_blocks": 8192,' ' "uuid": "29182e97-bee7-5ae3-92fd-906aa52020db",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc2",' ' "name": "crypto_ram2",' ' "key_name": "test_dek_qat_cbc2"' ' }' ' }' '}' '{' ' "name": "crypto_ram3",' ' "aliases": [' ' "0b993534-dac3-5253-b723-3eff596cf724"' ' ],' ' "product_name": "crypto",' ' "block_size": 4096,' ' "num_blocks": 8192,' ' "uuid": "0b993534-dac3-5253-b723-3eff596cf724",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc3",' ' "name": "crypto_ram3",' ' "key_name": "test_dek_qat_xts2"' ' }' ' }' '}' 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@355 -- # [[ -n crypto_ram 00:34:05.852 crypto_ram1 00:34:05.852 crypto_ram2 00:34:05.852 crypto_ram3 ]] 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@356 -- # jq -r 'select(.supported_io_types.unmap == true) | .name' 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@356 -- # printf '%s\n' '{' ' "name": "crypto_ram",' ' "aliases": [' ' "df965139-bb3b-5c62-b107-9ca32835792e"' ' ],' ' "product_name": "crypto",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "df965139-bb3b-5c62-b107-9ca32835792e",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc0",' ' "name": "crypto_ram",' ' "key_name": "test_dek_qat_cbc"' ' }' ' }' '}' '{' ' "name": "crypto_ram1",' ' "aliases": [' ' "4051e7bd-73e2-518a-b930-d41a7b9aa9d5"' ' ],' ' "product_name": "crypto",' ' "block_size": 512,' ' "num_blocks": 65536,' ' "uuid": "4051e7bd-73e2-518a-b930-d41a7b9aa9d5",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc1",' ' "name": "crypto_ram1",' ' "key_name": "test_dek_qat_xts"' ' }' ' }' '}' '{' ' "name": "crypto_ram2",' ' "aliases": [' ' "29182e97-bee7-5ae3-92fd-906aa52020db"' ' ],' ' "product_name": "crypto",' ' "block_size": 4096,' ' "num_blocks": 8192,' ' "uuid": "29182e97-bee7-5ae3-92fd-906aa52020db",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc2",' ' "name": "crypto_ram2",' ' "key_name": "test_dek_qat_cbc2"' ' }' ' }' '}' '{' ' "name": "crypto_ram3",' ' "aliases": [' ' "0b993534-dac3-5253-b723-3eff596cf724"' ' ],' ' "product_name": "crypto",' ' "block_size": 4096,' ' "num_blocks": 8192,' ' "uuid": "0b993534-dac3-5253-b723-3eff596cf724",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": true,' ' "flush": true,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "memory_domains": [' ' {' ' "dma_device_id": "system",' ' "dma_device_type": 1' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' },' ' {' ' "dma_device_id": "SPDK_ACCEL_DMA_DEVICE",' ' "dma_device_type": 2' ' }' ' ],' ' "driver_specific": {' ' "crypto": {' ' "base_bdev_name": "Malloc3",' ' "name": "crypto_ram3",' ' "key_name": "test_dek_qat_xts2"' ' }' ' }' '}' 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_crypto_ram]' 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=crypto_ram 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_crypto_ram1]' 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=crypto_ram1 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_crypto_ram2]' 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=crypto_ram2 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@356 -- # for b in $(printf '%s\n' "${bdevs[@]}" | jq -r 'select(.supported_io_types.unmap == true) | .name') 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@357 -- # echo '[job_crypto_ram3]' 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@358 -- # echo filename=crypto_ram3 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@367 -- # run_test bdev_fio_trim fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --verify_state_save=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1099 -- # '[' 11 -le 1 ']' 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1105 -- # xtrace_disable 00:34:05.852 11:44:47 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:34:05.852 ************************************ 00:34:05.852 START TEST bdev_fio_trim 00:34:05.852 ************************************ 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1123 -- # fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --verify_state_save=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --verify_state_save=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1339 -- # local sanitizers 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1341 -- # shift 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1343 -- # local asan_lib= 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # grep libasan 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # asan_lib= 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1345 -- # asan_lib= 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/crypto-phy-autotest/spdk/build/fio/spdk_bdev' 00:34:05.853 11:44:47 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json --verify_state_save=0 --aux-path=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:34:05.853 job_crypto_ram: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:34:05.853 job_crypto_ram1: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:34:05.853 job_crypto_ram2: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:34:05.853 job_crypto_ram3: (g=0): rw=trimwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:34:05.853 fio-3.35 00:34:05.853 Starting 4 threads 00:34:18.042 00:34:18.042 job_crypto_ram: (groupid=0, jobs=4): err= 0: pid=1065545: Mon Jul 15 11:45:00 2024 00:34:18.042 write: IOPS=31.8k, BW=124MiB/s (130MB/s)(1244MiB/10001msec); 0 zone resets 00:34:18.042 slat (usec): min=11, max=480, avg=74.77, stdev=50.13 00:34:18.042 clat (usec): min=29, max=1514, avg=262.40, stdev=182.12 00:34:18.042 lat (usec): min=51, max=1683, avg=337.17, stdev=217.87 00:34:18.042 clat percentiles (usec): 00:34:18.042 | 50.000th=[ 215], 99.000th=[ 971], 99.900th=[ 1106], 99.990th=[ 1172], 00:34:18.042 | 99.999th=[ 1254] 00:34:18.042 bw ( KiB/s): min=104448, max=160032, per=99.78%, avg=127041.89, stdev=6269.87, samples=76 00:34:18.042 iops : min=26112, max=40008, avg=31760.47, stdev=1567.41, samples=76 00:34:18.042 trim: IOPS=31.8k, BW=124MiB/s (130MB/s)(1244MiB/10001msec); 0 zone resets 00:34:18.042 slat (usec): min=4, max=1352, avg=19.63, stdev= 9.17 00:34:18.042 clat (usec): min=23, max=1684, avg=337.35, stdev=217.94 00:34:18.042 lat (usec): min=38, max=1732, avg=356.98, stdev=222.54 00:34:18.042 clat percentiles (usec): 00:34:18.042 | 50.000th=[ 281], 99.000th=[ 1172], 99.900th=[ 1319], 99.990th=[ 1385], 00:34:18.042 | 99.999th=[ 1467] 00:34:18.042 bw ( KiB/s): min=104448, max=160032, per=99.78%, avg=127041.89, stdev=6269.44, samples=76 00:34:18.042 iops : min=26112, max=40008, avg=31760.47, stdev=1567.36, samples=76 00:34:18.042 lat (usec) : 50=1.55%, 100=7.70%, 250=41.84%, 500=36.80%, 750=7.59% 00:34:18.042 lat (usec) : 1000=2.88% 00:34:18.042 lat (msec) : 2=1.63% 00:34:18.042 cpu : usr=99.53%, sys=0.00%, ctx=92, majf=0, minf=108 00:34:18.042 IO depths : 1=12.5%, 2=25.0%, 4=50.0%, 8=12.5%, 16=0.0%, 32=0.0%, >=64=0.0% 00:34:18.042 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:34:18.042 complete : 0=0.0%, 4=88.9%, 8=11.1%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:34:18.042 issued rwts: total=0,318346,318348,0 short=0,0,0,0 dropped=0,0,0,0 00:34:18.042 latency : target=0, window=0, percentile=100.00%, depth=8 00:34:18.042 00:34:18.042 Run status group 0 (all jobs): 00:34:18.042 WRITE: bw=124MiB/s (130MB/s), 124MiB/s-124MiB/s (130MB/s-130MB/s), io=1244MiB (1304MB), run=10001-10001msec 00:34:18.042 TRIM: bw=124MiB/s (130MB/s), 124MiB/s-124MiB/s (130MB/s-130MB/s), io=1244MiB (1304MB), run=10001-10001msec 00:34:18.042 00:34:18.042 real 0m13.429s 00:34:18.042 user 0m45.838s 00:34:18.042 sys 0m0.456s 00:34:18.042 11:45:01 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@1124 -- # xtrace_disable 00:34:18.042 11:45:01 blockdev_crypto_qat.bdev_fio.bdev_fio_trim -- common/autotest_common.sh@10 -- # set +x 00:34:18.042 ************************************ 00:34:18.042 END TEST bdev_fio_trim 00:34:18.042 ************************************ 00:34:18.042 11:45:01 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1142 -- # return 0 00:34:18.042 11:45:01 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@368 -- # rm -f 00:34:18.042 11:45:01 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@369 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.fio 00:34:18.042 11:45:01 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@370 -- # popd 00:34:18.042 /var/jenkins/workspace/crypto-phy-autotest/spdk 00:34:18.042 11:45:01 blockdev_crypto_qat.bdev_fio -- bdev/blockdev.sh@371 -- # trap - SIGINT SIGTERM EXIT 00:34:18.042 00:34:18.042 real 0m27.307s 00:34:18.042 user 1m31.744s 00:34:18.042 sys 0m1.169s 00:34:18.042 11:45:01 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@1124 -- # xtrace_disable 00:34:18.042 11:45:01 blockdev_crypto_qat.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:34:18.042 ************************************ 00:34:18.042 END TEST bdev_fio 00:34:18.042 ************************************ 00:34:18.042 11:45:01 blockdev_crypto_qat -- common/autotest_common.sh@1142 -- # return 0 00:34:18.042 11:45:01 blockdev_crypto_qat -- bdev/blockdev.sh@775 -- # trap cleanup SIGINT SIGTERM EXIT 00:34:18.042 11:45:01 blockdev_crypto_qat -- bdev/blockdev.sh@777 -- # run_test bdev_verify /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:34:18.042 11:45:01 blockdev_crypto_qat -- common/autotest_common.sh@1099 -- # '[' 16 -le 1 ']' 00:34:18.042 11:45:01 blockdev_crypto_qat -- common/autotest_common.sh@1105 -- # xtrace_disable 00:34:18.042 11:45:01 blockdev_crypto_qat -- common/autotest_common.sh@10 -- # set +x 00:34:18.042 ************************************ 00:34:18.042 START TEST bdev_verify 00:34:18.042 ************************************ 00:34:18.042 11:45:01 blockdev_crypto_qat.bdev_verify -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:34:18.042 [2024-07-15 11:45:01.248458] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:34:18.042 [2024-07-15 11:45:01.248522] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1066985 ] 00:34:18.042 [2024-07-15 11:45:01.374688] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:34:18.042 [2024-07-15 11:45:01.482029] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:34:18.042 [2024-07-15 11:45:01.482035] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:34:18.042 [2024-07-15 11:45:01.503370] accel_dpdk_cryptodev.c: 223:accel_dpdk_cryptodev_set_driver: *NOTICE*: Using driver crypto_qat 00:34:18.042 [2024-07-15 11:45:01.511403] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation encrypt will be assigned to module dpdk_cryptodev 00:34:18.042 [2024-07-15 11:45:01.519434] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation decrypt will be assigned to module dpdk_cryptodev 00:34:18.042 [2024-07-15 11:45:01.635018] accel_dpdk_cryptodev.c:1178:accel_dpdk_cryptodev_init: *NOTICE*: Found crypto devices: 96 00:34:20.568 [2024-07-15 11:45:03.843655] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_cbc" 00:34:20.568 [2024-07-15 11:45:03.843737] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:34:20.568 [2024-07-15 11:45:03.843752] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:34:20.568 [2024-07-15 11:45:03.851674] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_xts" 00:34:20.568 [2024-07-15 11:45:03.851696] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:34:20.568 [2024-07-15 11:45:03.851707] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:34:20.568 [2024-07-15 11:45:03.859695] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_cbc2" 00:34:20.568 [2024-07-15 11:45:03.859714] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:34:20.568 [2024-07-15 11:45:03.859726] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:34:20.568 [2024-07-15 11:45:03.867718] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_xts2" 00:34:20.568 [2024-07-15 11:45:03.867737] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc3 00:34:20.568 [2024-07-15 11:45:03.867749] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:34:20.568 Running I/O for 5 seconds... 00:34:25.830 00:34:25.830 Latency(us) 00:34:25.830 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:34:25.830 Job: crypto_ram (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:34:25.830 Verification LBA range: start 0x0 length 0x1000 00:34:25.830 crypto_ram : 5.07 484.61 1.89 0.00 0.00 262805.69 1731.01 180537.43 00:34:25.830 Job: crypto_ram (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:34:25.830 Verification LBA range: start 0x1000 length 0x1000 00:34:25.830 crypto_ram : 5.07 490.73 1.92 0.00 0.00 259521.90 2863.64 179625.63 00:34:25.830 Job: crypto_ram1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:34:25.830 Verification LBA range: start 0x0 length 0x1000 00:34:25.830 crypto_ram1 : 5.07 486.04 1.90 0.00 0.00 261267.93 1980.33 165036.74 00:34:25.830 Job: crypto_ram1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:34:25.830 Verification LBA range: start 0x1000 length 0x1000 00:34:25.830 crypto_ram1 : 5.07 493.58 1.93 0.00 0.00 257322.13 4274.09 164124.94 00:34:25.830 Job: crypto_ram2 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:34:25.830 Verification LBA range: start 0x0 length 0x1000 00:34:25.830 crypto_ram2 : 5.06 3809.94 14.88 0.00 0.00 33256.83 2706.92 27354.16 00:34:25.830 Job: crypto_ram2 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:34:25.830 Verification LBA range: start 0x1000 length 0x1000 00:34:25.831 crypto_ram2 : 5.05 3836.96 14.99 0.00 0.00 33022.81 3177.07 27582.11 00:34:25.831 Job: crypto_ram3 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:34:25.831 Verification LBA range: start 0x0 length 0x1000 00:34:25.831 crypto_ram3 : 5.06 3819.21 14.92 0.00 0.00 33102.52 1609.91 27696.08 00:34:25.831 Job: crypto_ram3 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:34:25.831 Verification LBA range: start 0x1000 length 0x1000 00:34:25.831 crypto_ram3 : 5.06 3846.09 15.02 0.00 0.00 32870.74 1759.50 27126.21 00:34:25.831 =================================================================================================================== 00:34:25.831 Total : 17267.15 67.45 0.00 0.00 58837.99 1609.91 180537.43 00:34:26.089 00:34:26.089 real 0m8.287s 00:34:26.089 user 0m15.692s 00:34:26.089 sys 0m0.381s 00:34:26.089 11:45:09 blockdev_crypto_qat.bdev_verify -- common/autotest_common.sh@1124 -- # xtrace_disable 00:34:26.089 11:45:09 blockdev_crypto_qat.bdev_verify -- common/autotest_common.sh@10 -- # set +x 00:34:26.089 ************************************ 00:34:26.089 END TEST bdev_verify 00:34:26.089 ************************************ 00:34:26.089 11:45:09 blockdev_crypto_qat -- common/autotest_common.sh@1142 -- # return 0 00:34:26.089 11:45:09 blockdev_crypto_qat -- bdev/blockdev.sh@778 -- # run_test bdev_verify_big_io /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:34:26.089 11:45:09 blockdev_crypto_qat -- common/autotest_common.sh@1099 -- # '[' 16 -le 1 ']' 00:34:26.089 11:45:09 blockdev_crypto_qat -- common/autotest_common.sh@1105 -- # xtrace_disable 00:34:26.089 11:45:09 blockdev_crypto_qat -- common/autotest_common.sh@10 -- # set +x 00:34:26.089 ************************************ 00:34:26.089 START TEST bdev_verify_big_io 00:34:26.089 ************************************ 00:34:26.089 11:45:09 blockdev_crypto_qat.bdev_verify_big_io -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:34:26.089 [2024-07-15 11:45:09.630922] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:34:26.089 [2024-07-15 11:45:09.631003] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1068506 ] 00:34:26.348 [2024-07-15 11:45:09.762075] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:34:26.348 [2024-07-15 11:45:09.861312] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:34:26.348 [2024-07-15 11:45:09.861317] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:34:26.348 [2024-07-15 11:45:09.882688] accel_dpdk_cryptodev.c: 223:accel_dpdk_cryptodev_set_driver: *NOTICE*: Using driver crypto_qat 00:34:26.348 [2024-07-15 11:45:09.890722] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation encrypt will be assigned to module dpdk_cryptodev 00:34:26.348 [2024-07-15 11:45:09.898748] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation decrypt will be assigned to module dpdk_cryptodev 00:34:26.607 [2024-07-15 11:45:09.997296] accel_dpdk_cryptodev.c:1178:accel_dpdk_cryptodev_init: *NOTICE*: Found crypto devices: 96 00:34:29.139 [2024-07-15 11:45:12.210782] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_cbc" 00:34:29.139 [2024-07-15 11:45:12.210872] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:34:29.139 [2024-07-15 11:45:12.210889] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:34:29.139 [2024-07-15 11:45:12.218798] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_xts" 00:34:29.139 [2024-07-15 11:45:12.218819] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:34:29.139 [2024-07-15 11:45:12.218831] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:34:29.139 [2024-07-15 11:45:12.226820] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_cbc2" 00:34:29.139 [2024-07-15 11:45:12.226840] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:34:29.139 [2024-07-15 11:45:12.226853] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:34:29.139 [2024-07-15 11:45:12.234845] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_xts2" 00:34:29.139 [2024-07-15 11:45:12.234863] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc3 00:34:29.139 [2024-07-15 11:45:12.234875] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:34:29.139 Running I/O for 5 seconds... 00:34:29.711 [2024-07-15 11:45:13.115106] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.115556] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.115944] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.116325] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.116390] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.116435] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.116478] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.116522] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.116965] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.116985] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.117002] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.117018] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.120466] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.120514] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.120558] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.120607] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.121076] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.121123] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.121165] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.121208] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.121640] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.121661] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.121676] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.121692] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.125149] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.125196] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.125241] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.125284] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.125739] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.125784] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.125827] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.125869] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.126286] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.126305] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.126320] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.126336] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.129656] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.129706] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.129747] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.129795] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.130257] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.130316] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.130360] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.130402] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.130841] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.130860] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.130882] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.130897] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.134147] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.134195] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.134237] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.134279] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.134730] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.134777] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.134820] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.134863] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.135303] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.135323] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.135339] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.135355] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.138528] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.138575] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.138620] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.138661] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.139120] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.139167] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.139210] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.139253] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.139671] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.139690] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.139708] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.139724] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.142836] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.142885] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.142934] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.142977] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.143457] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.143503] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.143546] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.143589] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.143944] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.143962] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.143978] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.143994] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.147223] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.147270] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.147314] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.147357] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.147820] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.147877] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.147921] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.147981] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.148352] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.148371] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.148386] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.148402] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.151860] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.151915] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.151964] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.152007] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.152438] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.152485] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.152526] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.152578] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.152933] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.152951] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.152967] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.152986] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.156330] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.156394] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.156450] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.156494] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.156916] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.156966] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.157011] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.157053] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.157473] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.157491] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.157506] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.157523] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.160594] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.160669] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.160711] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.160778] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.161197] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.161244] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.161286] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.161328] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.161756] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.161776] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.161792] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.161808] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.164863] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.164911] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.164959] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.165002] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.165524] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.165574] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.165616] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.165659] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.166088] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.166108] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.711 [2024-07-15 11:45:13.166124] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.166140] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.168982] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.169029] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.169069] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.169110] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.169570] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.169617] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.169659] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.169701] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.170107] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.170126] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.170141] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.170157] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.173339] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.173387] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.173433] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.173477] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.173949] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.173997] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.174039] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.174099] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.174552] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.174571] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.174587] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.174603] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.177508] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.177558] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.177600] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.177642] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.178043] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.178088] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.178131] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.178172] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.178589] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.178608] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.178624] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.178640] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.181713] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.181774] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.181816] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.181859] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.182352] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.182398] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.182441] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.182484] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.182904] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.182930] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.182947] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.182964] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.185783] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.185832] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.185873] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.185915] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.186394] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.186441] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.186483] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.186530] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.186900] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.186919] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.186941] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.186958] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.190002] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.190053] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.190098] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.190142] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.190581] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.190654] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.190709] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.190763] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.191187] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.191206] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.191220] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.191235] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.194203] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.194250] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.194293] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.194335] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.194759] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.194805] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.194849] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.194902] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.195261] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.195280] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.195295] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.195310] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.198578] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.198644] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.198685] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.198729] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.199118] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.199164] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.199208] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.199249] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.199664] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.199683] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.199698] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.199714] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.202687] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.202746] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.202789] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.202857] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.203373] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.203419] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.203461] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.203503] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.203941] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.203964] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.203980] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.203996] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.206915] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.206967] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.207024] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.207069] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.207537] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.207584] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.207627] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.207669] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.208120] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.208139] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.208157] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.208172] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.211007] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.211054] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.211096] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.211138] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.211611] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.211657] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.211700] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.211742] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.212144] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.212174] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.212189] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.212205] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.215045] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.215094] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.215136] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.215178] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.215607] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.215653] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.215695] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.712 [2024-07-15 11:45:13.215737] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.216166] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.216186] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.216202] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.216221] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.218905] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.218958] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.219005] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.219071] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.219547] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.219594] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.219637] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.219679] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.220110] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.220131] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.220148] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.220164] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.222898] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.222954] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.222996] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.223038] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.223512] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.223561] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.223603] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.223646] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.224045] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.224065] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.224082] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.224099] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.226845] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.226894] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.226943] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.226988] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.227459] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.227509] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.227579] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.227634] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.228043] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.228067] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.228081] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.228097] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.230932] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.230984] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.231036] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.231083] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.231495] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.231563] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.231620] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.231663] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.232061] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.232082] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.232099] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.232115] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.234795] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.234842] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.234883] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.234932] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.235282] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.235326] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.235369] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.235417] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.235691] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.235710] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.235725] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.235740] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.237721] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.237785] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.237829] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.237878] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.238195] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.238242] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.238286] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.238327] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.238780] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.238800] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.238816] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.238832] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.241354] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.241402] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.241448] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.241491] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.241809] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.241858] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.241899] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.241948] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.242221] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.242241] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.242256] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.242273] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.244116] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.244163] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.244203] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.244234] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.244593] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.244640] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.244682] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.244744] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.245205] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.245224] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.245245] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.245262] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.249245] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.250986] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.252618] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.253731] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.255659] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.257260] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.258311] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.258704] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.259161] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.259183] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.259199] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.259215] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.263087] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.264659] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.265766] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.267086] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.268942] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.270153] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.270544] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.270937] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.271340] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.271359] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.271375] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.713 [2024-07-15 11:45:13.271391] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.275013] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.275738] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.277065] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.278642] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.280632] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.281034] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.281425] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.281814] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.282234] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.282258] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.282274] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.282290] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.285188] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.286877] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.288458] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.290165] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.291020] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.291414] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.291803] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.292198] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.292640] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.292659] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.292675] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.292691] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.295730] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.297072] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.714 [2024-07-15 11:45:13.298645] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.300231] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.301022] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.301416] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.301807] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.302201] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.302478] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.302497] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.302512] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.302533] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.305629] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.307196] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.308761] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.309538] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.310379] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.310771] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.311170] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.312547] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.312858] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.312878] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.312892] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.312908] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.316267] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.317858] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.319063] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.319452] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.320272] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.320665] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.321641] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.322974] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.323250] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.323269] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.323284] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.323299] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.326731] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.328407] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.328804] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.329196] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.329970] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.330494] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.331828] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.333417] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.333692] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.333712] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.333727] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.333742] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.337120] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.337574] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.975 [2024-07-15 11:45:13.337970] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.338362] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.339193] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.340806] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.342540] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.344200] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.344477] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.344496] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.344512] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.344528] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.347163] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.347562] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.347956] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.348349] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.350163] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.351493] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.353072] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.354648] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.355046] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.355066] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.355081] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.355097] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.357199] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.357594] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.357990] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.358382] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.359985] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.361562] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.363139] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.363858] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.364139] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.364159] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.364175] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.364191] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.366439] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.366835] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.367236] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.368481] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.370374] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.371964] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.372979] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.374717] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.375003] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.375023] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.375038] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.375054] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.377491] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.377887] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.378880] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.380214] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.382077] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.383350] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.384795] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.386154] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.386429] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.386449] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.386463] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.386479] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.389134] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.389849] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.391169] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.392739] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.394582] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.395767] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.397086] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.398665] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.398951] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.398971] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.398986] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.399002] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.401910] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.403300] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.404876] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.406452] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.407698] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.409037] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.410615] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.412193] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.412560] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.412581] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.412598] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.412614] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.417274] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.418979] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.420618] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.422101] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.423731] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.425323] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.426904] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.427867] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.428308] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.428330] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.428346] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.428362] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.432117] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.433742] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.435478] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.436510] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.438398] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.439976] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.976 [2024-07-15 11:45:13.441222] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.441614] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.442049] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.442070] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.442089] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.442106] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.445807] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.447403] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.448142] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.449426] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.451369] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.453030] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.453423] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.453811] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.454234] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.454254] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.454269] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.454284] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.457760] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.458531] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.460038] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.461650] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.463517] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.463919] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.464318] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.464718] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.465156] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.465177] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.465194] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.465211] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.468029] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.469645] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.471165] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.472825] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.473650] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.474052] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.474442] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.474835] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.475240] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.475261] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.475277] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.475293] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.478292] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.479634] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.481211] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.482783] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.483571] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.483970] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.484359] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.485064] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.485345] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.485365] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.485380] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.485396] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.488368] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.489954] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.491536] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.492182] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.493042] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.493448] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.493841] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.495316] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.495598] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.495618] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.495633] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.495648] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.498939] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.500525] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.501489] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.501897] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.502718] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.503118] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.504797] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.506365] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.506647] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.506670] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.506685] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.506701] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.509953] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.511238] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.511635] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.512034] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.512856] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.514339] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.515698] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.517270] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.517553] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.517572] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.517587] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.517603] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.520903] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.521306] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.521697] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.522095] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.523644] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.524987] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.526560] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.528138] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.528488] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.977 [2024-07-15 11:45:13.528508] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.528523] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.528539] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.530475] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.530874] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.531269] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.531663] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.533325] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.534899] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.536488] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.537394] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.537676] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.537695] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.537710] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.537726] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.539763] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.540170] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.540565] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.541406] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.543436] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.545021] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.546458] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.547750] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.548087] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.548108] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.548124] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.548140] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.550329] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.550735] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.551259] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.552600] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.554369] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.554801] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.556235] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.557782] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.558070] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.558090] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.558109] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.558125] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.560902] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.561309] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.561705] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.562102] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.563001] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.563402] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.563798] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.564200] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.564646] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.564666] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.564681] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:29.978 [2024-07-15 11:45:13.564698] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.242 [2024-07-15 11:45:13.567447] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.242 [2024-07-15 11:45:13.567842] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.242 [2024-07-15 11:45:13.568238] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.242 [2024-07-15 11:45:13.568643] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.242 [2024-07-15 11:45:13.569480] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.242 [2024-07-15 11:45:13.569874] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.242 [2024-07-15 11:45:13.570271] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.242 [2024-07-15 11:45:13.570664] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.242 [2024-07-15 11:45:13.571033] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.242 [2024-07-15 11:45:13.571055] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.242 [2024-07-15 11:45:13.571071] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.242 [2024-07-15 11:45:13.571087] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.242 [2024-07-15 11:45:13.573767] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.242 [2024-07-15 11:45:13.574174] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.242 [2024-07-15 11:45:13.574572] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.242 [2024-07-15 11:45:13.574617] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.575479] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.575880] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.576282] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.576701] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.577108] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.577130] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.577145] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.577161] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.579839] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.580254] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.580650] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.581048] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.581097] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.581533] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.581947] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.582345] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.582739] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.583140] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.583582] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.583602] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.583617] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.583634] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.585864] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.585914] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.585968] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.586016] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.586424] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.586481] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.586524] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.586568] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.586609] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.587044] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.587066] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.587083] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.587099] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.589291] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.589340] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.589384] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.589426] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.589842] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.589897] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.589948] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.589991] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.590033] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.590453] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.590472] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.590488] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.590504] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.592705] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.592754] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.592796] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.592836] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.593261] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.593317] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.593361] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.593404] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.593445] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.593784] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.593805] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.593820] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.593837] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.596128] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.596187] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.596229] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.596274] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.596686] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.596753] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.596814] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.596870] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.596915] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.597295] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.597315] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.597330] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.597346] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.599710] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.599759] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.599802] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.599853] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.600255] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.600327] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.600383] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.600426] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.600483] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.600860] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.600883] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.600899] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.600914] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.603422] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.603485] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.603528] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.603572] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.603948] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.604016] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.604065] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.243 [2024-07-15 11:45:13.604109] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.604151] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.604573] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.604594] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.604610] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.604627] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.606965] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.607025] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.607097] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.607153] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.607591] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.607657] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.607700] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.607742] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.607784] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.608225] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.608245] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.608263] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.608279] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.610444] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.610493] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.610539] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.610595] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.611077] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.611134] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.611177] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.611221] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.611263] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.611661] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.611687] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.611703] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.611719] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.613970] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.614018] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.614060] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.614103] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.614540] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.614602] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.614646] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.614691] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.614752] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.615206] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.615226] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.615243] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.615259] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.617549] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.617599] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.617641] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.617682] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.618068] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.618126] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.618169] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.618213] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.618258] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.618692] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.618711] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.618727] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.618744] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.620958] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.621009] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.621059] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.621105] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.621536] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.621596] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.621641] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.621684] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.621728] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.622137] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.622158] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.622173] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.622189] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.624457] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.624504] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.624549] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.624591] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.625038] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.625096] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.625141] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.625196] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.625238] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.625639] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.625659] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.625674] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.625690] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.628000] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.628050] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.628093] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.628136] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.628497] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.628568] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.628616] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.628659] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.628703] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.629079] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.629100] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.629116] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.629132] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.631649] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.631711] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.244 [2024-07-15 11:45:13.631753] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.631811] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.632217] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.632284] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.632359] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.632416] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.632459] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.632847] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.632867] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.632883] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.632899] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.635260] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.635309] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.635351] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.635414] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.635818] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.635885] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.635954] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.635998] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.636041] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.636479] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.636499] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.636521] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.636538] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.638798] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.638858] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.638905] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.638957] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.639369] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.639424] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.639469] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.639512] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.639553] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.640007] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.640028] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.640044] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.640061] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.642438] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.642488] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.642529] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.642570] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.643015] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.643075] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.643118] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.643162] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.643205] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.643583] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.643603] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.643618] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.643635] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.645853] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.645902] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.645960] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.646003] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.646441] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.646500] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.646545] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.646588] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.646630] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.647085] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.647105] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.647121] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.647138] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.649408] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.649457] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.649499] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.649542] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.649943] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.650000] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.650045] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.650087] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.650132] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.650570] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.650591] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.650606] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.650623] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.652910] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.652965] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.653011] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.653052] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.653489] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.653544] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.653589] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.653636] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.653678] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.654027] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.654046] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.654062] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.654077] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.656053] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.656102] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.656143] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.656185] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.656596] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.656654] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.656698] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.656741] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.656782] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.657206] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.245 [2024-07-15 11:45:13.657227] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.657246] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.657262] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.658862] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.658909] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.658958] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.659000] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.659341] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.659406] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.659453] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.659496] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.659537] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.659851] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.659871] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.659890] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.659906] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.661642] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.661690] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.661732] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.661775] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.662226] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.662285] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.662329] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.662375] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.662419] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.662867] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.662888] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.662904] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.662921] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.664613] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.664664] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.664709] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.664751] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.665058] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.665121] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.665167] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.665209] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.665256] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.665530] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.665550] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.665565] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.665581] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.667237] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.667284] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.667326] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.667373] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.667805] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.667867] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.667911] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.667963] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.668006] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.668446] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.668466] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.668481] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.668498] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.670291] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.670339] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.670380] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.670432] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.670704] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.670759] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.670810] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.670853] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.670897] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.671213] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.671233] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.671248] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.671264] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.672884] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.672939] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.672983] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.673026] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.673461] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.673518] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.673562] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.673605] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.673651] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.674029] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.674048] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.674063] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.674080] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.675994] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.676041] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.676086] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.676127] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.676395] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.676458] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.676501] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.676543] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.676585] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.676966] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.676986] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.677002] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.677019] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.678542] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.678590] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.678632] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.678675] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.679097] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.246 [2024-07-15 11:45:13.679159] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.679203] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.679246] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.679291] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.679719] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.679739] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.679754] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.679773] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.681757] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.681810] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.681851] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.681893] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.682171] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.682232] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.682275] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.682317] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.682359] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.682708] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.682728] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.682744] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.682760] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.684313] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.684361] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.684754] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.684800] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.685235] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.685293] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.685338] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.685380] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.685422] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.685836] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.685855] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.685870] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.685887] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.687697] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.687744] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.687784] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.689398] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.689775] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.689839] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.689882] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.689923] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.689972] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.690306] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.690322] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.690336] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.690350] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.692458] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.692853] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.693249] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.694934] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.695207] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.696792] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.698377] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.699120] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.700451] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.700726] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.700745] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.700760] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.700775] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.703085] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.703480] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.704946] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.706286] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.706559] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.708151] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.708934] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.710452] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.712085] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.712360] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.712380] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.712395] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.712411] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.714856] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.715968] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.717296] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.718866] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.719148] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.720265] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.721934] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.723462] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.247 [2024-07-15 11:45:13.725116] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.725395] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.725414] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.725429] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.725444] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.728671] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.730009] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.731587] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.733158] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.733473] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.734919] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.736249] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.737826] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.739418] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.739812] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.739832] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.739848] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.739864] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.743485] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.745065] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.746648] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.747831] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.748120] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.749458] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.751039] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.752618] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.753229] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.753673] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.753693] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.753709] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.753725] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.757507] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.759084] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.760516] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.761822] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.762162] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.763753] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.765325] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.766176] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.766569] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.767021] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.767042] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.767057] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.767072] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.770682] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.772376] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.773375] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.774704] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.774994] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.776594] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.777750] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.778148] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.778539] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.779002] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.779024] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.779039] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.779056] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.782300] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.783057] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.784382] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.785957] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.786233] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.787733] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.788131] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.788522] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.788915] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.789358] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.789378] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.789393] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.789409] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.791677] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.793157] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.794759] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.796342] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.796618] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.797036] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.797430] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.797820] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.798218] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.798514] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.798532] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.798547] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.798563] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.801769] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.803288] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.804907] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.806618] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.806997] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.807405] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.807795] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.808194] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.809390] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.809698] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.809717] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.809732] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.809749] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.812640] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.814222] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.815796] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.816246] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.248 [2024-07-15 11:45:13.816717] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.817129] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.817522] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.818332] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.819651] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.819932] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.819952] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.819967] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.819982] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.823137] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.824725] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.825512] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.825903] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.826337] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.826755] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.827252] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.828604] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.830187] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.830462] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.830481] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.830496] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.249 [2024-07-15 11:45:13.830512] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.833721] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.834829] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.835227] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.835614] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.836069] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.836473] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.838131] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.839812] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.841493] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.841768] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.841786] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.841801] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.841817] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.844970] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.845366] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.845758] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.846153] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.846588] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.847989] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.849315] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.850892] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.852472] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.852845] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.852864] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.852879] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.852895] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.854777] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.855181] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.855572] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.855963] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.856274] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.857595] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.859179] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.860751] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.861556] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.513 [2024-07-15 11:45:13.861832] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.861851] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.861866] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.861881] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.863861] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.864266] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.864658] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.865745] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.866085] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.867675] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.869252] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.870372] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.872012] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.872323] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.872346] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.872362] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.872378] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.874605] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.875007] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.875833] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.877156] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.877429] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.879026] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.880456] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.881819] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.883146] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.883423] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.883442] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.883457] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.883472] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.886091] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.886586] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.887947] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.889527] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.889804] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.891525] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.892586] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.893907] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.895485] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.895763] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.895782] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.895797] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.895813] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.898447] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.900016] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.901691] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.903275] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.903551] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.904438] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.905766] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.907353] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.908939] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.909265] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.909285] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.909302] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.909319] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.913624] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.915290] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.916960] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.918478] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.918831] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.920174] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.921748] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.923328] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.924263] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.924696] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.924716] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.924732] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.924749] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.928301] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.929909] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.931618] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.932654] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.932972] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.934564] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.936142] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.937291] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.937683] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.938145] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.938166] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.938182] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.938200] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.941750] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.943290] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.944021] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.945348] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.945626] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.947306] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.948842] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.949241] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.949632] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.950075] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.950095] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.950110] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.950127] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.953447] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.954308] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.955892] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.514 [2024-07-15 11:45:13.957576] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.957856] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.959448] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.959868] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.960268] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.960658] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.961122] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.961147] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.961163] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.961179] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.963799] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.965510] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.967089] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.968790] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.969074] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.969588] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.969988] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.970379] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.970775] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.971158] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.971177] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.971192] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.971208] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.974474] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.976065] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.977190] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.977585] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.978034] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.978438] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.978829] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.979232] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.979631] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.980094] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.980115] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.980132] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.980149] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.982776] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.983178] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.983573] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.983970] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.984324] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.984744] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.985144] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.985536] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.985932] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.986367] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.986388] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.986404] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.986420] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.988996] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.989394] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.989792] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.990197] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.990649] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.991060] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.991451] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.991848] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.992254] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.992717] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.992737] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.992752] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.992768] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.995357] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.995760] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.996157] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.996552] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.997002] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.997407] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.997805] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.998216] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.998611] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.999073] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.999094] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.999110] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:13.999126] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.001748] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.002152] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.002547] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.002949] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.003352] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.003756] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.004153] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.004542] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.004941] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.005297] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.005319] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.005335] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.005351] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.008120] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.008521] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.008936] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.009325] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.009773] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.010187] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.010585] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.010994] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.011392] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.011800] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.011820] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.515 [2024-07-15 11:45:14.011841] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.011856] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.014491] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.014885] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.014938] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.015327] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.015674] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.016090] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.016485] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.016874] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.017284] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.017720] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.017740] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.017758] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.017775] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.020446] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.020843] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.021251] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.021312] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.021688] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.022101] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.022493] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.022881] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.023286] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.023664] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.023683] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.023700] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.023716] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.025933] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.025981] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.026023] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.026081] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.026422] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.026488] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.026544] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.026588] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.026659] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.027110] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.027130] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.027146] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.027162] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.029473] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.029531] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.029572] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.029614] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.029950] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.030016] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.030060] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.030103] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.030162] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.030610] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.030630] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.030648] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.030664] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.032909] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.032960] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.033017] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.033074] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.033436] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.033492] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.033534] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.033579] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.033620] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.034051] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.034072] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.034088] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.034105] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.036332] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.036379] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.036436] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.036480] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.036933] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.036991] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.037035] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.037078] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.037120] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.037503] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.037522] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.037537] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.037553] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.039881] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.039931] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.039973] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.040015] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.040452] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.040510] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.040557] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.040599] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.040642] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.041111] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.041131] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.041147] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.041169] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.043393] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.043441] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.043483] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.043525] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.043911] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.043970] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.044014] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.516 [2024-07-15 11:45:14.044056] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.044098] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.044536] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.044558] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.044573] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.044589] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.046887] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.046954] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.046996] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.047038] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.047494] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.047550] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.047593] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.047636] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.047680] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.048064] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.048084] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.048100] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.048115] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.050326] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.050373] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.050419] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.050466] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.050910] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.050971] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.051029] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.051071] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.051128] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.051498] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.051517] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.051533] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.051549] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.053998] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.054046] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.054089] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.054131] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.054472] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.054542] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.054586] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.054630] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.054672] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.055067] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.055087] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.055104] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.055121] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.057773] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.057831] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.057888] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.057962] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.058359] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.058428] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.058486] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.058530] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.058581] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.059011] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.059030] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.059046] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.059062] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.061344] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.061391] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.061432] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.061483] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.061819] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.061884] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.061934] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.061977] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.062018] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.062476] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.062496] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.062512] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.062528] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.064550] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.064598] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.064645] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.064686] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.064998] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.065063] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.065106] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.065148] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.065189] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.065558] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.065578] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.517 [2024-07-15 11:45:14.065594] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.065609] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.067993] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.068053] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.068115] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.068158] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.068542] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.068615] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.068662] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.068704] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.068748] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.069143] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.069162] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.069179] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.069194] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.071418] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.071464] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.071506] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.071553] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.071823] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.071882] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.071924] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.071981] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.072024] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.072296] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.072315] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.072330] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.072346] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.074001] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.074047] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.074088] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.074129] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.074401] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.074463] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.074506] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.074548] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.074589] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.074959] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.074979] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.074995] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.075011] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.077277] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.077322] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.077367] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.077408] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.077716] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.077777] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.077821] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.077862] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.077903] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.078176] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.078196] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.078210] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.078226] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.079875] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.079947] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.079992] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.080034] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.080301] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.080367] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.080410] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.080452] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.080493] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.080820] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.080839] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.080856] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.080872] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.083255] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.083306] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.083350] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.083391] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.083659] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.083719] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.083767] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.083810] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.083852] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.084125] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.084145] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.084160] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.084175] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.085774] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.085820] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.085861] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.085903] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.086177] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.086238] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.086284] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.086326] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.086376] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.086644] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.086664] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.086678] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.086694] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.089061] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.089119] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.089166] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.089208] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.089480] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.089539] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.518 [2024-07-15 11:45:14.089582] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.089630] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.089672] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.089951] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.089971] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.089986] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.090002] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.091692] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.091739] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.091780] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.091821] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.092093] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.092157] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.092199] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.092241] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.092282] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.092552] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.092571] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.092586] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.092601] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.094848] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.094895] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.094943] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.094985] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.095356] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.095419] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.095461] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.095503] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.095544] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.095847] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.095866] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.095881] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.095897] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.097492] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.097539] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.097586] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.097630] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.097902] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.097965] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.098016] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.098061] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.098102] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.098367] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.098387] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.098402] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.098418] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.100555] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.100603] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.100649] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.100691] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.101142] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.101196] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.101243] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.101286] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.519 [2024-07-15 11:45:14.101327] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.101617] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.101637] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.101652] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.101667] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.103282] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.103328] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.103369] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.103410] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.103713] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.103779] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.103822] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.103865] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.103906] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.104181] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.104199] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.104214] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.104230] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.106211] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.106258] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.106299] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.106341] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.106770] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.106825] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.106869] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.106910] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.106960] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.107241] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.107259] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.107274] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.107290] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.108957] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.109011] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.109052] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.109093] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.109397] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.109461] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.109503] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.109545] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.109586] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.109850] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.109869] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.109885] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.109901] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.111857] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.111904] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.111973] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.112017] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.112470] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.112525] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.112569] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.112611] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.112654] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.113024] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.113044] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.113059] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.113074] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.782 [2024-07-15 11:45:14.114587] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.114635] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.114677] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.114718] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.114994] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.115054] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.115101] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.115149] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.115193] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.115466] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.115484] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.115499] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.115515] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.117444] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.117491] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.117533] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.117574] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.117952] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.118018] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.118061] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.118104] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.118147] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.118595] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.118617] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.118632] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.118647] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.120197] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.120243] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.120984] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.121032] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.121389] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.121453] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.121496] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.121538] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.121579] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.121846] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.121868] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.121883] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.121898] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.123951] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.123998] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.124044] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.124433] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.124859] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.124919] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.124970] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.125012] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.125053] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.125352] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.125370] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.125385] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.125401] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.128361] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.129948] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.131523] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.131997] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.132461] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.132871] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.133269] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.133896] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.135233] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.135510] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.135528] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.135543] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.135560] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.138729] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.140319] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.141316] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.141723] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.142178] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.142581] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.142979] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.144572] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.146272] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.146549] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.146568] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.146583] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.146599] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.149879] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.151058] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.151454] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.151846] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.152251] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.152655] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.154131] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.155477] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.157056] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.157336] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.157354] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.157369] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.157385] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.160693] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.161101] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.161490] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.161878] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.162316] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.163404] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.164734] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.166315] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.783 [2024-07-15 11:45:14.167889] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.168273] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.168293] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.168308] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.168324] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.170277] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.170678] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.171076] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.171471] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.171834] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.173168] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.174739] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.176319] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.177381] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.177657] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.177676] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.177691] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.177706] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.179714] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.180114] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.180505] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.181228] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.181526] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.183284] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.184862] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.186347] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.187591] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.187901] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.187919] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.187946] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.187963] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.190217] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.190626] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.191069] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.192490] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.192779] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.194416] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.196168] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.197186] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.198517] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.198794] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.198813] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.198828] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.198843] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.201189] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.201589] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.203270] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.204939] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.205218] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.206815] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.207562] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.208897] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.210466] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.210750] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.210770] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.210784] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.210800] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.213410] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.214981] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.216459] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.218070] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.218349] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.219107] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.220442] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.222026] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.223607] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.223886] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.223905] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.223921] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.223943] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.227704] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.229043] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.230606] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.232185] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.232584] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.234043] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.235603] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.237173] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.238494] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.238885] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.238904] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.238920] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.238941] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.242425] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.244001] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.245576] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.246308] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.246590] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.248155] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.249842] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.251480] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.251873] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.252310] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.252330] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.252346] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.252363] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.255972] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.257563] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.258400] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.259960] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.784 [2024-07-15 11:45:14.260240] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.261835] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.263446] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.263847] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.264252] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.264623] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.264645] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.264661] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.264678] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.268113] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.269254] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.270853] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.272342] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.272619] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.274224] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.274821] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.275217] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.275608] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.276081] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.276104] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.276129] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.276149] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.279267] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.280515] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.281855] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.283424] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.283703] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.284681] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.285099] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.285493] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.285893] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.286339] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.286359] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.286375] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.286390] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.288810] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.290146] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.291729] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.293318] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.293646] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.294069] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.294461] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.294862] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.295346] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.295625] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.295644] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.295659] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.295675] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.298600] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.300205] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.301778] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.302837] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.303274] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.303679] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.304080] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.304474] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.306139] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.306441] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.306459] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.306474] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.306489] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.309810] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.311403] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.312726] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.313124] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.313556] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.313964] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.314355] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.315940] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.317401] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.317675] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.317693] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.317708] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.317723] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.320978] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.322712] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.323114] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.323507] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.323861] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.324275] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.325257] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.326585] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.328169] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.328447] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.328465] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.328480] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.328496] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.331715] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.332167] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.332559] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.332956] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.333400] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.334267] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.335594] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.337179] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.338760] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.339085] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.339104] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.339119] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.339135] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.785 [2024-07-15 11:45:14.341367] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.341770] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.342171] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.342566] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.342992] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.344520] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.346186] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.347855] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.348949] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.349304] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.349324] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.349339] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.349358] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.351780] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.352185] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.353761] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.355465] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.355740] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.357329] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.357807] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.359183] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.360787] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.361067] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.361085] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.361101] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.361116] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.363539] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.363939] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.364335] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.364728] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.365166] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.365567] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.365965] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.366357] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.366752] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.367189] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.367208] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.367224] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.367240] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.369954] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.370367] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.370758] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.371155] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.371562] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:30.786 [2024-07-15 11:45:14.371974] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.049 [2024-07-15 11:45:14.372371] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.049 [2024-07-15 11:45:14.372764] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.049 [2024-07-15 11:45:14.373163] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.049 [2024-07-15 11:45:14.373614] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.049 [2024-07-15 11:45:14.373636] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.049 [2024-07-15 11:45:14.373652] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.049 [2024-07-15 11:45:14.373667] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.049 [2024-07-15 11:45:14.376265] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.049 [2024-07-15 11:45:14.376664] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.049 [2024-07-15 11:45:14.377065] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.049 [2024-07-15 11:45:14.377462] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.049 [2024-07-15 11:45:14.377854] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.049 [2024-07-15 11:45:14.378268] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.049 [2024-07-15 11:45:14.378662] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.049 [2024-07-15 11:45:14.379060] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.049 [2024-07-15 11:45:14.379456] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.049 [2024-07-15 11:45:14.379829] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.379848] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.379864] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.379881] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.382900] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.383311] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.383710] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.384106] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.384582] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.384994] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.385392] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.385792] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.386195] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.386599] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.386618] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.386633] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.386649] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.389269] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.389667] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.390066] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.390460] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.390828] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.391240] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.391632] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.392028] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.392417] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.392777] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.392796] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.392811] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.392827] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.395479] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.395877] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.396279] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.396680] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.397141] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.397543] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.397943] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.398340] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.398743] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.399189] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.399209] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.399225] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.399241] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.401855] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.402275] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.402666] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.403061] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.403434] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.403840] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.404247] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.404640] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.405044] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.405466] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.405485] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.405502] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.405519] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.408247] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.408642] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.409058] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.409457] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.409904] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.410315] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.410722] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.411123] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.411520] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.411919] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.411944] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.411959] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.411976] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.414630] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.415054] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.415447] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.415842] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.416303] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.416708] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.417119] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.417522] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.417914] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.418285] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.418304] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.418321] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.418337] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.420907] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.421312] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.421363] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.421750] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.422133] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.422541] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.422941] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.423331] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.423735] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.424179] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.424199] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.050 [2024-07-15 11:45:14.424215] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.424231] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.427097] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.427497] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.427893] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.427956] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.428335] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.428741] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.429136] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.429527] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.429922] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.430324] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.430343] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.430358] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.430373] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.432659] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.432708] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.432750] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.432793] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.433197] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.433263] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.433307] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.433361] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.433404] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.433789] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.433807] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.433823] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.433839] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.436355] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.436418] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.436470] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.436512] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.436876] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.436946] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.436991] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.437034] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.437076] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.437465] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.437483] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.437498] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.437514] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.439840] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.439904] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.439951] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.440007] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.440374] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.440438] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.440481] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.440524] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.440565] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.440989] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.441009] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.441026] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.441042] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.443229] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.443276] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.443319] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.443360] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.443768] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.443832] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.443876] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.443919] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.443969] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.444405] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.444424] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.444439] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.444454] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.446676] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.446726] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.446766] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.446808] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.447085] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.447155] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.447199] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.447242] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.447284] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.447698] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.447717] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.447732] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.447750] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.450050] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.450117] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.450170] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.450240] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.450627] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.450691] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.450734] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.450776] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.450818] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.451253] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.451273] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.451288] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.451304] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.453343] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.453391] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.453432] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.453473] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.453747] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.453807] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.051 [2024-07-15 11:45:14.453850] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.453892] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.453952] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.454228] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.454249] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.454264] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.454280] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.455901] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.455953] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.455999] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.456041] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.456406] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.456465] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.456522] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.456578] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.456622] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.457083] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.457103] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.457120] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.457137] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.459290] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.459338] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.459381] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.459422] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.459690] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.459754] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.459797] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.459837] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.459879] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.460154] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.460171] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.460186] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.460200] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.461834] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.461891] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.461943] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.461984] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.462257] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.462315] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.462359] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.462402] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.462444] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.462824] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.462842] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.462857] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.462873] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.465052] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.465107] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.465150] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.465196] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.465468] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.465526] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.465575] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.465615] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.465658] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.465923] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.465945] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.465959] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.465974] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.467630] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.467677] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.467724] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.467765] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.468042] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.468103] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.468148] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.468189] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.468236] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.468611] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.468635] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.468650] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.468665] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.470919] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.470970] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.471012] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.471053] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.471379] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.471443] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.471486] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.471528] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.471572] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.471850] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.471868] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.471883] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.471898] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.473553] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.473599] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.473641] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.473682] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.473956] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.474017] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.474061] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.474102] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.474143] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.474557] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.474579] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.474595] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.474610] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.477164] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.052 [2024-07-15 11:45:14.477211] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.477252] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.477294] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.477606] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.477666] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.477711] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.477752] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.477794] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.478069] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.478087] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.478102] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.478117] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.479758] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.479805] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.479858] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.479904] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.480357] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.480420] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.480462] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.480504] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.480546] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.480908] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.480932] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.480947] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.480963] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.483292] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.483339] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.483388] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.483429] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.483728] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.483794] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.483836] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.483878] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.483920] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.484199] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.484217] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.484232] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.484247] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.485903] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.485980] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.486022] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.486064] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.486334] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.486396] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.486438] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.486480] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.486521] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.486825] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.486843] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.486858] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.486876] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.489468] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.489519] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.489560] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.489601] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.489892] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.489971] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.490015] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.490061] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.490102] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.490372] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.490391] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.490406] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.490421] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.492021] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.492067] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.492110] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.492152] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.492426] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.492486] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.492529] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.492588] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.492631] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.492899] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.492917] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.492937] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.492956] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.495296] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.495344] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.495390] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.495433] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.495707] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.495767] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.495817] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.495859] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.495905] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.496183] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.496202] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.496220] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.496235] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.497867] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.497913] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.497962] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.498003] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.498273] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.498332] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.498375] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.498416] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.498457] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.053 [2024-07-15 11:45:14.498725] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.498743] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.498758] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.498773] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.501156] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.501204] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.501246] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.501288] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.501619] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.501674] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.501715] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.501756] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.501796] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.502098] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.502117] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.502132] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.502147] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.503764] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.503810] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.503862] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.503908] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.504187] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.504249] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.504291] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.504333] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.504374] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.504640] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.504658] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.504672] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.504687] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.506962] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.507010] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.507052] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.507095] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.507518] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.507578] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.507620] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.507662] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.507703] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.508007] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.508025] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.508040] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.508056] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.509673] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.509718] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.509758] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.509802] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.510094] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.510155] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.510197] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.510242] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.510303] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.510576] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.510595] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.510609] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.510625] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.512825] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.512872] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.512913] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.512964] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.513395] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.513447] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.513491] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.513533] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.513577] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.513848] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.513866] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.513881] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.513896] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.515510] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.515555] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.515599] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.515640] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.515946] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.516007] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.516049] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.516090] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.516132] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.516400] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.054 [2024-07-15 11:45:14.516418] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.516433] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.516452] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.518696] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.518743] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.518784] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.518826] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.519276] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.519333] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.519377] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.519421] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.519462] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.519763] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.519781] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.519796] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.519812] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.521456] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.521503] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.521546] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.521588] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.521855] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.521914] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.521981] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.522023] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.522064] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.522333] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.522351] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.522366] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.522381] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.524417] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.524464] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.524858] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.524912] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.525350] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.525404] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.525452] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.525497] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.525538] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.525808] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.525827] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.525843] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.525858] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.527565] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.527612] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.527653] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.529018] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.529295] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.529356] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.529399] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.529440] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.529490] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.529762] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.529780] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.529794] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.529809] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.533043] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.534377] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.535949] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.537521] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.537835] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.539311] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.540642] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.542197] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.543760] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.544200] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.544220] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.544236] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.544252] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.547860] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.549442] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.551029] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.552086] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.552361] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.553698] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.555275] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.556849] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.557303] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.557743] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.557763] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.557779] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.557795] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.561559] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.563137] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.564540] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.565873] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.566214] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.567801] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.569363] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.570263] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.570660] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.571090] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.571109] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.571125] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.571141] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.574900] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.576589] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.577562] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.578888] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.579167] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.580766] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.055 [2024-07-15 11:45:14.582002] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.582395] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.582788] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.583192] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.583211] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.583226] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.583241] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.586600] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.587350] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.588666] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.590244] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.590520] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.592179] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.592573] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.592970] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.593366] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.593801] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.593820] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.593836] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.593853] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.596436] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.598151] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.599785] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.601494] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.601773] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.602199] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.602594] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.602986] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.603379] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.603735] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.603753] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.603767] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.603782] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.606744] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.608084] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.609659] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.611229] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.611593] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.612006] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.612401] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.612796] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.613517] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.613804] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.613823] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.613838] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.613853] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.616822] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.618409] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.619989] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.620673] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.621148] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.621550] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.621957] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.622367] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.623806] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.624091] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.624109] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.624124] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.624139] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.627361] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.628946] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.629902] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.630314] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.630754] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.631165] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.631559] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.633266] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.634835] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.635118] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.635136] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.635151] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.635167] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.056 [2024-07-15 11:45:14.638424] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.320 [2024-07-15 11:45:14.639686] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.320 [2024-07-15 11:45:14.640089] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.320 [2024-07-15 11:45:14.640481] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.320 [2024-07-15 11:45:14.640869] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.320 [2024-07-15 11:45:14.641284] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.320 [2024-07-15 11:45:14.642746] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.320 [2024-07-15 11:45:14.644060] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.320 [2024-07-15 11:45:14.645621] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.320 [2024-07-15 11:45:14.645894] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.320 [2024-07-15 11:45:14.645913] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.320 [2024-07-15 11:45:14.645934] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.645949] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.649362] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.649764] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.650162] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.650555] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.650998] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.652044] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.653368] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.654939] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.656512] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.656862] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.656881] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.656896] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.656912] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.658939] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.659340] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.659734] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.660133] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.660491] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.661826] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.663405] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.664981] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.666061] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.666339] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.666357] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.666372] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.666387] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.668434] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.668831] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.669229] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.670022] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.670324] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.671981] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.673481] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.675002] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.676227] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.676533] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.676552] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.676567] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.676582] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.678803] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.679223] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.679617] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.681107] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.681382] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.682957] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.684542] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.685397] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.686716] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.686996] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.687014] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.687029] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.687045] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.689427] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.689825] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.691419] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.692889] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.693171] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.694769] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.695512] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.696841] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.698419] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.698698] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.698716] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.698731] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.698746] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.701334] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.702741] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.704077] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.705646] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.705921] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.706763] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.708299] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.709965] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.711543] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.711814] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.711832] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.711847] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.711863] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.714834] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.716413] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.717999] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.718847] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.719151] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.720746] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.722331] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.722881] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.723281] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.723705] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.723723] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.723738] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.723754] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.321 [2024-07-15 11:45:14.727447] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.728957] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.729842] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.731164] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.731438] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.733043] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.734363] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.734759] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.735159] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.735512] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.735530] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.735545] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.735561] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.738197] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.738597] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.738996] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.739394] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.739758] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.740172] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.740565] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.740966] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.741361] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.741733] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.741751] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.741766] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.741781] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.744506] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.744915] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.745323] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.745716] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.746102] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.746517] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.746912] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.747314] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.747712] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.748143] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.748165] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.748182] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.748198] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.750855] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.751259] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.751655] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.752053] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.752425] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.752832] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.753233] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.753631] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.754026] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.754462] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.754481] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.754497] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.754514] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.757304] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.757701] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.758105] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.758503] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.758975] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.759377] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.759773] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.760171] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.760579] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.760965] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.760998] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.761017] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.761033] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.763792] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.764201] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.764598] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.764995] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.765431] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.765834] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.766235] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.766632] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.767030] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.767390] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.767409] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.767425] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.767441] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.770118] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.770524] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.770916] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.771313] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.771692] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.772107] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.772500] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.772891] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.773292] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.773701] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.773721] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.773736] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.773752] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.776490] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.776896] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.777539] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.778786] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.779236] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.779745] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.781122] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.781514] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.322 [2024-07-15 11:45:14.781907] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.782321] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.782340] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.782354] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.782370] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.785487] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.785885] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.786284] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.786681] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.787048] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.788069] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.788461] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.789634] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.790333] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.790770] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.790789] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.790805] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.790820] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.793225] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.794772] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.795177] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.795574] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.795948] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.796354] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.798063] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.798463] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.798853] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.799135] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.799153] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.799167] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.799183] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.802888] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.803297] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.804210] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.805174] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.805604] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.806016] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.806414] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.807430] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.808276] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.808712] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.808731] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.808747] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.808763] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.811240] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.812652] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.813121] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.813515] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.813790] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.814432] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.814827] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.815227] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.815625] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.815903] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.815921] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.815947] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.815963] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.818663] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.819073] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.819545] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.820948] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.821423] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.821826] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.823346] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.823738] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.824141] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.824575] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.824595] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.824610] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.824626] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.827814] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.828220] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.828271] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.828661] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.829054] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.830063] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.830933] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.831328] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.832653] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.833032] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.833052] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.833067] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.833083] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.835682] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.836089] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.837808] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.837868] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.838361] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.838768] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.839172] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.839607] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.841035] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.841490] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.841510] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.841526] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.841542] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.843676] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.843735] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.843779] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.843821] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.323 [2024-07-15 11:45:14.844097] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.844158] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.844201] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.844246] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.844288] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.844632] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.844651] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.844665] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.844683] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.846890] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.846946] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.846989] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.847030] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.847303] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.847368] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.847412] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.847458] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.847499] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.847937] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.847957] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.847973] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.847989] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.850274] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.850339] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.850395] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.850438] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.850879] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.850941] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.850988] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.851033] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.851076] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.851461] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.851481] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.851496] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.851512] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.853853] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.853904] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.853955] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.853997] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.854424] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.854494] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.854540] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.854582] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.854623] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.854939] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.854958] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.854977] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.854992] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.856687] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.856734] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.856786] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.856829] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.857110] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.857169] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.857214] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.857266] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.857308] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.857591] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.857610] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.857626] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.857641] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.859889] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.859945] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.859989] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.860031] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.860473] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.860531] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.860574] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.860616] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.860657] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.861004] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.861024] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.861039] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.861055] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.862736] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.862784] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.862835] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.862882] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.863162] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.863217] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.863268] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.863315] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.863357] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.863625] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.863644] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.863659] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.863674] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.865878] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.865936] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.865983] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.866025] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.866460] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.866514] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.866561] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.866602] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.866642] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.866917] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.866943] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.866958] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.324 [2024-07-15 11:45:14.866974] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.868664] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.868711] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.868752] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.868801] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.869080] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.869141] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.869184] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.869227] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.869284] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.869554] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.869573] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.869588] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.869603] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.871783] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.871830] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.871873] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.871916] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.872356] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.872410] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.872454] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.872497] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.872541] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.872812] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.872831] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.872846] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.872863] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.874574] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.874620] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.874660] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.874702] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.875016] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.875076] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.875118] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.875160] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.875200] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.875471] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.875490] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.875505] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.875524] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.877648] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.877697] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.877741] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.877783] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.878219] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.878272] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.878315] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.878357] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.878400] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.878672] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.878690] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.878705] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.878721] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.880419] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.880464] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.880505] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.880546] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.880878] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.880950] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.880993] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.881035] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.881076] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.881343] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.881361] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.881376] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.881391] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.883624] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.883683] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.883725] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.883767] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.884203] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.884256] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.884300] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.884342] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.884383] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.884710] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.884728] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.884742] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.884757] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.886485] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.886536] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.886578] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.886620] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.325 [2024-07-15 11:45:14.886895] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.886964] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.887011] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.887052] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.887094] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.887363] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.887381] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.887396] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.887412] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.889515] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.889562] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.889619] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.889674] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.890146] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.890201] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.890244] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.890290] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.890339] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.890718] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.890737] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.890751] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.890767] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.894971] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.895022] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.895063] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.895104] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.895380] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.895444] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.895486] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.895548] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.895590] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.895859] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.895877] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.895893] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.895909] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.898779] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.898830] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.898871] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.898912] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.899192] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.899252] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.899295] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.899336] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.899392] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.899659] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.899678] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.899693] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.899708] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.904284] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.904337] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.904380] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.904423] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.904818] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.904886] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.904940] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.904983] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.905025] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.905463] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.905483] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.905499] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.326 [2024-07-15 11:45:14.905515] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.586 [2024-07-15 11:45:14.909812] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.586 [2024-07-15 11:45:14.909873] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.586 [2024-07-15 11:45:14.909920] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.586 [2024-07-15 11:45:14.909975] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.586 [2024-07-15 11:45:14.910249] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.586 [2024-07-15 11:45:14.910303] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.586 [2024-07-15 11:45:14.910355] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.586 [2024-07-15 11:45:14.910397] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.586 [2024-07-15 11:45:14.910440] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.586 [2024-07-15 11:45:14.910708] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.586 [2024-07-15 11:45:14.910727] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.586 [2024-07-15 11:45:14.910741] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.586 [2024-07-15 11:45:14.910756] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.586 [2024-07-15 11:45:14.914586] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.586 [2024-07-15 11:45:14.914638] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.586 [2024-07-15 11:45:14.914687] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.586 [2024-07-15 11:45:14.914730] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.586 [2024-07-15 11:45:14.915014] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.586 [2024-07-15 11:45:14.915072] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.915122] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.915167] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.915210] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.915478] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.915496] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.915512] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.915527] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.920155] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.920219] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.920260] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.920302] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.920745] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.920798] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.920843] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.920885] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.920935] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.921289] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.921308] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.921324] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.921339] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.925078] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.925133] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.925174] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.925216] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.925512] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.925588] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.925632] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.925674] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.925715] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.926002] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.926022] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.926037] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.926053] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.930432] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.930489] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.930534] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.930576] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.930878] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.930950] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.930994] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.931035] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.931076] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.931349] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.931367] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.931382] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.931397] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.936070] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.936121] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.936168] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.936211] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.936563] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.936630] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.936673] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.936714] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.936758] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.937196] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.937216] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.937233] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.937250] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.940749] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.940807] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.940849] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.940897] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.941247] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.941305] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.941347] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.941388] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.941428] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.941747] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.941766] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.941781] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.941796] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.945867] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.945920] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.945971] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.946014] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.946399] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.946455] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.946497] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.946539] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.946579] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.946887] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.946905] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.946920] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.946942] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.951861] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.951913] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.951983] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.952025] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.952295] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.952361] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.952406] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.952448] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.952491] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.952877] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.952895] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.952910] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.587 [2024-07-15 11:45:14.952933] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.957078] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.957131] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.957183] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.957225] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.957496] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.962450] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.962503] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.962555] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.962598] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.963055] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.969525] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.981535] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.982840] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.982898] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.984470] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.987987] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.988663] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.989062] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.989449] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.989950] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.990533] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.990590] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.991880] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.991950] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.993505] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.993562] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.995110] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.995435] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.995454] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.995470] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.995487] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.997525] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.997921] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.998316] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:14.998701] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.000350] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.001907] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.003478] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.004547] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.004821] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.004840] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.004855] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.004871] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.006935] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.007332] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.007723] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.008724] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.010619] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.012195] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.013472] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.014941] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.015244] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.015262] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.015281] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.015297] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.017573] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.017979] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.018865] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.020195] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.022051] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.023464] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.024808] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.026144] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.026417] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.026435] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.026450] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.026466] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.028946] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.029689] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.031009] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.032565] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.034427] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.035611] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.036945] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.038496] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.038770] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.038788] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.038803] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.038818] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.041760] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.043074] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.044632] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.046200] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.047507] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.048840] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.050412] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.051985] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.052353] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.052372] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.052387] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.052402] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.056282] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.057861] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.059450] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.060652] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.062241] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.063817] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.065393] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.066148] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.588 [2024-07-15 11:45:15.066623] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.066643] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.066658] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.066675] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.070408] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.071995] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.073279] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.074754] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.076653] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.078241] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.079018] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.079408] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.079850] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.079869] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.079884] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.079904] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.083488] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.084854] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.086262] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.087580] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.089432] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.090308] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.090700] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.091095] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.091542] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.091561] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.091577] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.091593] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.094901] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.096118] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.097443] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.098987] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.100399] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.100792] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.101187] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.101581] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.102026] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.102046] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.102061] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.102077] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.104569] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.105881] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.107459] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.109035] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.109755] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.110152] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.110553] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.110984] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.111257] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.111275] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.111290] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.111306] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.114462] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.115903] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.117113] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.117791] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.119190] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.120090] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.120481] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.121947] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.122255] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.122274] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.122289] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.122305] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.125682] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.127292] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.128750] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.129162] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.129982] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.130376] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.131718] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.133055] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.133330] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.133348] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.133363] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.133378] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.136696] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.138314] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.138718] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.139116] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.139924] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.140324] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.140722] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.141121] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.141543] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.141562] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.141579] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.141597] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.144163] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.144561] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.144959] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.145353] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.146160] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.146555] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.146951] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.147344] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.147789] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.147809] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.147825] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.147840] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.589 [2024-07-15 11:45:15.150469] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.150868] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.151277] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.151671] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.152484] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.152877] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.153277] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.153675] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.154047] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.154067] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.154082] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.154098] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.156743] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.157147] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.157540] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.157941] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.158766] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.159170] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.159563] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.159958] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.160410] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.160429] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.160444] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.160460] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.163088] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.163482] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.163879] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.164290] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.165049] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.165441] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.165834] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.166232] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.166668] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.166686] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.166701] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.166717] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.169362] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.169765] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.170170] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.170559] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.171378] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.171773] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.172176] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.172585] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.173014] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.173034] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.173049] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.173065] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.175741] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.176145] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.176536] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.176937] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.177740] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.590 [2024-07-15 11:45:15.178140] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.178529] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.178920] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.179318] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.179337] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.179352] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.179367] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.182054] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.182453] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.182850] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.183246] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.184072] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.184463] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.184856] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.185259] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.185724] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.185743] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.185761] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.185778] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.188420] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.188816] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.189209] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.189600] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.190427] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.190827] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.191226] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.191616] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.192055] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.192075] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.192091] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.192110] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.194671] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.195075] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.195469] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.195866] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.196685] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.197079] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.197469] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.197869] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.198235] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.198254] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.198269] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.198285] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.201158] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.201560] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.201958] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.202352] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.203185] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.203595] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.203997] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.204387] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.204788] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.204806] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.204822] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.204838] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.207449] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.207842] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.208238] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.208632] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.209381] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.209773] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.210167] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.210560] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.210954] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.210973] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.210989] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.211004] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.213635] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.214067] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.214467] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.214859] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.215641] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.216083] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.217499] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.218261] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.218537] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.218560] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.218576] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.218592] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.221311] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.221714] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.222114] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.222505] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.223637] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.224051] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.224447] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.224838] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.225319] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.225339] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.225355] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.225374] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.228626] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.229542] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.230859] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.232436] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.234057] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.852 [2024-07-15 11:45:15.234450] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.234840] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.235235] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.235680] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.235701] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.235718] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.235734] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.237997] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.239332] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.240902] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.242449] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.243123] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.243516] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.243905] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.244300] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.244594] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.244612] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.244627] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.244642] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.249841] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.251585] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.251637] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.251675] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.252439] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.252487] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.252872] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.252917] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.253365] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.253384] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.253400] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.253418] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.255802] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.257392] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.257446] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.257488] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.259427] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.259483] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.260960] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.261007] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.261420] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.261438] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.261453] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.261476] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.263718] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.263764] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.263812] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.263853] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.265693] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.265742] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.267438] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.267493] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.267889] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.267908] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.267932] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.267949] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.269546] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.269594] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.269636] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.269693] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.270543] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.270607] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.270998] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.271043] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.271478] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.271497] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.271512] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.271527] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.273076] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.273125] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.273167] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.273210] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.273522] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.273571] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.273613] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.273660] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.273940] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.273960] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.273975] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.273990] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.275860] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.275910] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.275960] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.276002] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.276432] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.276478] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.276520] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.276563] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.276992] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.277012] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.277030] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.277046] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.278592] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.278638] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.278680] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.278721] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.279185] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.279231] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.279272] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.279313] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.853 [2024-07-15 11:45:15.279646] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.279665] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.279682] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.279697] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.281483] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.281531] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.281574] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.281616] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.282099] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.282146] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.282189] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.282229] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.282676] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.282696] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.282712] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.282728] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.284385] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.284433] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.284475] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.284519] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.284883] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.284937] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.284985] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.285025] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.285297] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.285315] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.285331] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.285347] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.287056] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.287102] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.287143] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.287186] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.287646] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.287692] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.287740] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.287782] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.288223] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.288243] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.288259] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.288276] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.290007] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.290052] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.290101] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.290146] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.290455] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.290506] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.290548] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.290590] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.290894] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.290913] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.290934] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.290950] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.292584] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.292637] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.292679] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.292721] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.293200] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.293247] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.293289] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.293332] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.293712] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.293730] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.293746] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.293762] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.295664] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.295715] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.295756] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.295801] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.296117] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.296164] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.296205] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.296245] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.296671] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.296689] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.296704] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.296720] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.298284] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.298332] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.298373] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.298414] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.298857] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.298903] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.298953] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.298996] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.299430] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.299453] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.299469] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.299485] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.301607] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.301654] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.301705] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.301746] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.302071] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.302120] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.302162] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.302208] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.302506] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.302524] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.302540] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.302556] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.854 [2024-07-15 11:45:15.304101] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.304146] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.304187] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.304227] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.304657] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.304703] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.304745] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.304787] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.305211] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.305231] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.305247] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.305263] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.307279] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.307325] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.307366] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.307410] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.307721] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.307766] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.307807] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.307847] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.308117] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.308136] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.308151] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.308166] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.309886] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.309942] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.309987] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.310029] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.310376] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.310422] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.310465] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.310522] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.310969] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.310989] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.311005] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.311021] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.313149] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.313209] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.313250] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.313291] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.313607] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.313651] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.313692] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.313733] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.314005] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.314023] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.314038] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.314053] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.315715] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.315761] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.315801] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.315845] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.316163] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.316208] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.316257] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.316299] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.316705] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.316724] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.316739] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.316755] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.319045] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.319090] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.319130] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.319181] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.319491] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.319535] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.319577] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.319646] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.319916] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.319941] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.319956] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.319971] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.321653] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.321700] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.321742] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.321783] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.322097] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.322142] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.322184] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.322226] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.322673] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.322692] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.322707] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.322722] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.325035] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.325081] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.325121] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.325167] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.325522] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.325567] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.325608] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.325649] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.325917] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.325941] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.325957] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.325973] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.327638] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.327684] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.327744] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.855 [2024-07-15 11:45:15.327786] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.328103] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.328153] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.328195] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.328237] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.328516] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.328534] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.328549] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.328565] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.330998] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.331046] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.331090] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.331135] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.331450] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.331516] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.331560] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.331603] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.331875] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.331893] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.331912] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.331932] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.333537] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.333582] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.333622] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.333663] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.333978] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.334023] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.334065] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.334106] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.334381] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.334399] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.334414] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.334429] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.336849] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.336909] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.336957] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.336999] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.337321] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.337365] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.337407] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.337455] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.337732] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.337750] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.337764] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.337779] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.339433] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.339482] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.339523] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.339564] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.339881] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.339932] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.339974] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.340014] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.340283] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.340301] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.340316] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.340331] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.342578] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.342625] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.342668] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.342717] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.343152] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.343198] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.343240] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.343281] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.343594] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.343613] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.343628] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.343643] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.345257] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.345302] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.345349] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.345390] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.345702] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.345746] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.345797] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.345841] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.346116] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.346135] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.346150] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.346170] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.348376] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.856 [2024-07-15 11:45:15.348422] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.348464] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.348506] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.350910] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.350965] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.351007] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.352375] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.358957] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.364358] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.365818] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.373913] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.373972] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.374025] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.374387] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.377020] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.378356] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.379934] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.381519] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.381860] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.381918] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.382289] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.382334] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.382386] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.382747] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.383206] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.383225] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.386464] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.387209] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.388538] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.390111] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.390439] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.391849] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.391908] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.392278] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.392667] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.393065] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.393514] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.393533] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.393548] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.393564] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.395803] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.397216] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.398774] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.400340] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.401016] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.401409] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.401797] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.402190] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.402505] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.402522] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.402538] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.402552] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.405671] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.407076] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.408646] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.410274] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.411141] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.411533] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.411921] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.412924] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.413244] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.413262] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.413277] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.413292] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.416213] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.417792] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.419368] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.419980] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.420835] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.421233] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.421837] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.423166] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.423442] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.423460] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.423474] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.423490] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.426653] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.428241] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.429241] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.429649] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.430481] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.430874] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.432547] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.434224] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.434501] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.434519] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.434534] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.434549] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.437760] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.439214] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.439607] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.440058] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:31.857 [2024-07-15 11:45:15.440900] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.442521] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.444003] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.445618] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.445897] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.445915] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.445936] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.445952] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.449184] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.449579] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.449973] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.450365] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.451671] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.453247] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.454830] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.455575] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.455853] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.455872] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.455887] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.455902] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.458210] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.458603] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.459482] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.460814] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.462668] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.464070] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.465420] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.466756] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.467044] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.467063] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.467078] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.467093] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.469634] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.470193] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.471511] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.473121] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.475124] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.475793] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.477121] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.478729] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.479009] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.479028] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.479042] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.479058] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.481470] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.481868] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.482273] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.482683] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.483507] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.483899] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.484303] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.484703] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.485164] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.485183] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.485198] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.485214] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.487875] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.488277] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.488668] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.489069] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.489851] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.490260] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.490649] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.491045] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.491487] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.491506] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.491523] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.491539] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.494069] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.494463] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.494851] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.495263] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.496094] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.496485] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.496876] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.497275] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.497649] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.497668] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.497683] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.497698] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.500437] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.120 [2024-07-15 11:45:15.500855] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.501254] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.501646] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.502474] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.502878] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.503282] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.503672] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.504083] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.504106] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.504122] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.504138] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.506696] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.507096] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.507488] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.507880] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.508644] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.509041] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.509430] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.509822] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.510214] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.510233] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.510248] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.510265] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.512878] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.513284] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.513682] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.514084] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.514885] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.515287] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.515691] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.516094] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.516525] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.516545] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.516562] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.516578] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.519140] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.519537] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.519934] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.520320] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.521067] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.521461] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.521850] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.522242] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.522688] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.522711] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.522727] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.522743] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.525290] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.525686] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.526090] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.526487] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.527338] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.527741] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.528140] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.528539] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.528912] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.528935] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.528951] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.528968] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.531581] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.531992] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.532386] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.532777] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.533604] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.534009] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.534406] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.534798] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.535207] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.535226] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.535246] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.535261] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.537872] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.538269] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.538660] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.539058] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.539841] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.540236] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.540632] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.541031] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.541387] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.541406] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.541422] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.541437] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.544132] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.544558] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.544970] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.545364] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.546127] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.546521] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.546922] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.547318] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.547732] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.547751] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.547767] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.547784] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.550349] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.550750] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.551147] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.551537] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.121 [2024-07-15 11:45:15.552306] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.552701] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.553095] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.553503] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.553946] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.553967] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.553984] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.554000] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.556102] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.556494] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.556883] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.557282] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.558090] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.558485] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.558873] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.559271] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.559662] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.559680] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.559695] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.559713] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.562347] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.562744] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.563146] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.564628] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.565476] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.566854] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.567340] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.567731] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.568012] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.568031] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.568046] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.568066] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.571150] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.572767] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.574350] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.575736] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.576548] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.576945] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.577335] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.578772] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.579080] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.579099] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.579114] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.579129] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.582075] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.583654] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.585302] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.586151] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.586825] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.587533] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.588706] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.589102] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.589459] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.589478] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.589492] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.589507] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.592264] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.593596] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.595171] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.596737] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.597542] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.597940] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.598333] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.599107] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.599414] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.599433] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.599448] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.599463] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.602355] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.603938] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.605512] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.606421] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.607357] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.607408] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.607795] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.607843] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.608122] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.608141] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.608156] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.608171] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.611754] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.613450] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.613506] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.614667] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.616556] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.618137] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.618186] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.619115] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.619571] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.619590] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.619606] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.619623] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.623060] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.623117] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.623167] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.623202] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.625055] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.625112] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.626198] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.122 [2024-07-15 11:45:15.627527] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.627803] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.627822] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.627838] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.627853] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.629608] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.629656] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.629699] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.629742] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.630226] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.631828] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.632241] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.632288] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.632665] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.632684] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.632699] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.632714] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.634249] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.634297] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.634339] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.634381] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.635963] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.637541] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.637590] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.639169] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.639535] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.639555] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.639570] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.639588] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.641963] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.642010] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.642051] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.642092] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.643957] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.644007] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.645569] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.645616] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.646032] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.646052] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.646068] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.646083] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.647613] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.647661] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.647709] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.647754] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.649367] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.649428] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.649471] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.649514] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.649958] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.649978] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.649997] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.650013] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.651907] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.651959] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.652013] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.652056] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.652368] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.652414] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.652478] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.652521] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.652791] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.652810] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.652825] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.652840] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.654499] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.654546] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.654586] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.654626] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.654945] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.654991] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.655033] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.655074] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.655439] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.655459] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.655474] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.655491] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.657809] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.657855] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.657896] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.657948] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.658306] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.658352] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.658395] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.658435] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.658711] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.658730] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.658745] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.658760] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.660416] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.660489] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.660532] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.660574] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.660885] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.660937] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.660979] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.661020] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.661353] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.661373] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.661388] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.123 [2024-07-15 11:45:15.661404] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.663283] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.663328] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.663376] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.663417] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.663888] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.663941] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.663984] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.664027] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.664406] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.664425] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.664440] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.664455] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.665977] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.666024] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.666066] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.666113] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.666428] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.666473] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.666522] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.666565] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.666834] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.666853] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.666868] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.666883] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.668808] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.668855] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.668898] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.668951] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.669393] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.669438] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.669483] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.669527] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.669975] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.669996] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.670012] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.670029] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.671565] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.671611] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.671653] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.671703] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.672091] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.672137] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.672180] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.672221] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.672525] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.672547] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.672562] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.672578] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.674183] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.674230] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.674276] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.674319] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.674801] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.674847] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.674891] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.674940] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.675333] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.675352] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.675367] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.675382] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.677309] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.677354] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.677395] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.677435] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.677750] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.677795] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.677838] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.677878] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.678152] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.678171] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.678186] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.678201] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.679828] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.679873] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.679920] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.679977] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.124 [2024-07-15 11:45:15.680296] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.680346] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.680389] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.680432] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.680822] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.680841] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.680856] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.680872] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.683099] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.683145] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.683186] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.683233] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.683545] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.683591] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.683635] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.683684] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.683961] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.683980] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.683995] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.684011] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.685665] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.685710] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.685751] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.685792] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.686104] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.686150] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.686191] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.686231] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.686608] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.686628] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.686643] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.686666] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.688581] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.688636] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.688678] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.688720] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.689187] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.689234] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.689277] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.689320] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.689668] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.689687] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.689702] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.689717] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.691291] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.691337] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.691378] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.691426] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.691731] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.691783] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.691826] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.691870] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.692147] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.692167] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.692181] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.692197] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.694094] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.694143] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.694185] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.694228] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.694702] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.694753] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.694797] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.694840] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.695269] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.695289] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.695305] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.695321] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.696858] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.696905] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.696952] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.697000] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.697386] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.697432] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.697473] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.697515] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.697828] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.697847] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.697862] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.697878] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.699461] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.699507] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.699548] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.699596] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.700102] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.700150] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.700194] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.700236] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.700639] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.700659] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.700674] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.700690] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.702627] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.702673] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.702714] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.702758] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.703076] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.703121] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.703164] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.703205] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.125 [2024-07-15 11:45:15.703472] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.703491] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.703507] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.703522] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.705156] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.705203] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.705249] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.705294] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.705598] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.705650] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.705693] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.705736] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.706160] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.706179] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.706194] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.706209] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.708358] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.708404] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.708445] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.708492] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.708806] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.708852] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.708899] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.708956] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.709224] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.709243] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.709257] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.126 [2024-07-15 11:45:15.709274] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.710909] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.710961] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.711002] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.711043] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.711359] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.711405] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.711447] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.711487] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.711832] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.711851] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.711866] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.711883] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.713780] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.713835] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.713877] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.713923] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.714397] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.714444] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.714487] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.714529] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.714861] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.714880] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.714895] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.714911] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.716482] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.716534] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.716575] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.716621] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.716934] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.716987] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.717030] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.717077] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.717347] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.717366] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.717381] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.717397] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.719304] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.719351] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.719393] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.719436] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.719875] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.719923] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.719973] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.720015] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.720455] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.720474] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.720489] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.720504] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.722027] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.722073] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.722113] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.722160] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.722583] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.723903] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.723956] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.725520] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.725798] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.725818] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.725833] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.725848] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.727877] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.729497] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.729552] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.729941] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.730436] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.730485] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.731857] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.731905] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.732181] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.388 [2024-07-15 11:45:15.732201] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.732216] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.732232] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.733827] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.735414] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.736999] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.737840] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.738241] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.738681] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.738744] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.739139] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.739185] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.739229] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.739623] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.739642] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.739657] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.739673] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.742178] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.743513] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.745082] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.746661] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.747065] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.748579] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.748634] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.748684] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.749080] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.749499] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.749517] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.749531] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.749545] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.755765] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.756897] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.758226] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.759800] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.760080] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.760146] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.760189] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.760922] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.760974] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.761419] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.761440] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.761455] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.761472] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.765256] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.766920] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.768461] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.769658] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.769988] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.770053] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.771614] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.771661] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.773227] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.773632] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.773653] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.773668] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.773684] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.781393] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.783121] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.784744] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.786251] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.786589] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.786651] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.787971] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.789545] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.791125] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.791449] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.791468] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.791483] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.791500] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.795579] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.797065] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.798662] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.800381] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.800763] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.802115] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.803693] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.805274] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.806500] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.806786] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.806804] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.806819] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.806835] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.811390] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.812968] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.814541] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.815285] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.815562] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.817124] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.818792] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.820466] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.820858] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.821301] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.821323] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.821339] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.821356] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.824881] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.826464] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.827428] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.829141] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.389 [2024-07-15 11:45:15.829416] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.830990] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.832551] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.833198] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.834540] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.834984] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.835004] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.835020] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.835036] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.840020] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.841191] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.842492] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.843941] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.844263] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.844673] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.845069] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.845468] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.845875] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.846157] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.846176] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.846191] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.846207] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.849137] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.850715] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.852294] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.853535] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.853829] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.854804] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.855201] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.856479] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.857037] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.857478] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.857499] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.857515] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.857531] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.864003] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.865733] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.867299] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.867693] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.868136] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.868543] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.868944] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.870233] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.871547] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.871822] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.871841] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.871856] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.871871] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.875059] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.876648] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.877335] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.878938] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.879381] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.879792] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.881304] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.881701] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.882103] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.882474] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.882493] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.882508] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.882524] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.886301] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.887322] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.887715] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.888981] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.889366] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.889775] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.890176] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.890570] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.890966] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.891402] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.891426] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.891443] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.891459] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.895157] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.895562] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.895957] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.897573] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.898039] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.898448] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.898846] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.899245] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.899638] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.900054] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.900082] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.900097] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.900113] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.904003] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.905184] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.905577] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.905973] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.906355] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.906762] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.907159] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.907550] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.907948] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.908286] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.908306] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.908322] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.908339] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.390 [2024-07-15 11:45:15.911548] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.912234] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.912627] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.913036] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.913420] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.913827] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.914222] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.914622] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.915018] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.915405] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.915424] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.915439] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.915459] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.918766] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.919180] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.919582] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.919980] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.920349] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.920751] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.921151] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.921546] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.922553] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.922871] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.922889] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.922905] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.922920] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.925379] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.925780] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.926179] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.926572] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.926983] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.927384] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.927783] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.928183] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.929862] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.930332] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.930353] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.930369] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.930384] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.933546] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.933949] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.934341] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.934736] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.935147] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.935797] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.937032] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.937423] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.938466] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.938807] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.938825] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.938841] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.938857] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.941451] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.941849] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.942245] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.942639] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.943008] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.944285] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.944889] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.945283] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.946949] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.947392] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.947416] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.947431] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.947448] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.950824] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.951239] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.951981] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.953109] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.953546] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.954230] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.955426] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.955817] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.956213] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.956610] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.956629] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.956646] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.956662] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.959340] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.959740] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.961204] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.961615] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.962051] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.963509] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.963938] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.964333] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.964734] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.965112] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.965131] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.965146] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.965163] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.969582] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.969985] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.970871] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.971868] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.972293] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.972699] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.973099] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.973492] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.391 [2024-07-15 11:45:15.973882] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.392 [2024-07-15 11:45:15.974299] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.392 [2024-07-15 11:45:15.974319] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.392 [2024-07-15 11:45:15.974334] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.392 [2024-07-15 11:45:15.974350] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.392 [2024-07-15 11:45:15.977180] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.392 [2024-07-15 11:45:15.977580] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.392 [2024-07-15 11:45:15.979089] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.979489] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.979944] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.980348] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.980745] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.981157] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.981548] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.981999] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.982019] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.982036] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.982053] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.987742] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.988148] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.988647] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.990006] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.990411] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.991932] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.992324] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.992713] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.993108] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.993537] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.993558] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.993574] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.993590] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.995888] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.997190] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.997771] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.998168] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.998577] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:15.998991] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:16.000478] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:16.000874] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:16.001275] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:16.001552] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:16.001571] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:16.001586] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:16.001602] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:16.005221] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:16.006574] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:16.006971] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:16.008011] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.655 [2024-07-15 11:45:16.008318] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.009909] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.011480] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.012670] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.014205] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.014518] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.014536] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.014556] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.014572] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.016533] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.016939] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.018515] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.018906] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.019329] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.020832] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.022439] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.024016] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.025390] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.025710] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.025729] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.025744] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.025760] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.030079] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.031628] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.032032] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.032488] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.032764] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.034244] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.035815] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.037490] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.038434] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.038741] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.038760] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.038775] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.038790] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.041344] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.041740] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.043023] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.043611] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.044064] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.045427] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.046754] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.048334] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.049909] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.050293] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.050312] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.050327] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.050343] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.055260] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.056265] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.057150] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.057543] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.057845] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.059186] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.059239] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.060807] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.060855] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.061136] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.061156] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.061171] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.061186] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.062842] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.064459] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.064508] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.065731] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.066116] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.066526] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.067719] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.067772] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.068259] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.068689] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.068710] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.068726] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.068746] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.073753] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.073806] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.073848] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.073889] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.074167] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.075762] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.075811] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.076747] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.078462] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.078902] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.078922] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.078944] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.078963] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.081321] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.081371] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.081412] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.081453] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.081757] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.081822] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.083398] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.084983] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.085031] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.085322] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.085342] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.085358] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.656 [2024-07-15 11:45:16.085379] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.090245] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.090300] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.090341] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.090383] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.090818] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.091445] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.092686] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.092733] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.093128] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.093449] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.093468] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.093484] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.093500] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.095100] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.095148] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.095212] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.095258] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.095529] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.097022] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.097071] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.098760] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.098808] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.099084] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.099104] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.099119] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.099135] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.103628] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.103681] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.103723] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.103774] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.104053] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.105404] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.105452] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.105496] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.105547] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.105818] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.105835] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.105850] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.105865] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.107532] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.107579] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.107620] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.107661] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.107938] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.108001] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.108044] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.108085] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.108127] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.108507] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.108526] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.108541] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.108556] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.113751] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.113802] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.113843] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.113883] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.114187] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.114251] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.114293] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.114335] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.114381] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.114650] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.114670] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.114685] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.114701] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.116359] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.116422] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.116470] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.116511] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.116777] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.116837] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.116880] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.116921] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.116969] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.117289] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.117308] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.117324] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.117340] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.123320] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.123372] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.123413] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.123454] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.123754] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.123816] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.123859] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.123901] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.123950] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.124221] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.124239] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.124254] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.124270] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.125893] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.125947] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.125991] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.126043] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.126312] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.126365] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.126419] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.657 [2024-07-15 11:45:16.126460] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.126502] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.126766] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.126787] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.126802] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.126817] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.131856] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.131908] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.131964] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.132006] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.132272] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.132337] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.132384] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.132426] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.132467] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.132733] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.132751] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.132766] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.132781] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.134395] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.134442] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.134488] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.134529] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.134803] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.134865] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.134907] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.134955] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.135004] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.135274] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.135292] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.135307] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.135321] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.139166] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.139220] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.139265] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.139307] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.139575] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.139632] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.139673] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.139726] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.139768] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.140046] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.140064] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.140080] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.140095] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.141779] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.141827] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.141869] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.141910] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.142188] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.142249] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.142292] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.142333] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.142373] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.142647] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.142665] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.142681] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.142695] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.145750] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.145803] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.145849] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.145892] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.146271] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.146334] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.146376] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.146418] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.146460] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.146761] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.146780] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.146795] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.146811] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.148418] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.148464] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.148518] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.148564] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.148836] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.148891] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.148950] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.148996] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.149037] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.149306] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.149325] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.149341] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.149355] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.152752] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.152805] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.152850] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.152892] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.153344] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.153400] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.153448] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.153490] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.153531] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.153827] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.153847] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.153863] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.153879] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.155500] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.155548] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.155590] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.658 [2024-07-15 11:45:16.155632] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.155952] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.156014] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.156059] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.156101] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.156141] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.156413] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.156432] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.156448] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.156463] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.160333] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.160384] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.160426] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.160468] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.160901] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.160969] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.161014] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.161063] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.161116] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.161385] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.161404] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.161419] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.161434] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.163067] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.163113] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.163157] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.163197] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.163524] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.163585] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.163627] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.163669] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.163712] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.163986] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.164006] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.164021] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.164037] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.167816] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.167875] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.167918] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.167966] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.168396] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.168454] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.168496] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.168539] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.168580] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.168899] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.168919] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.168941] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.168957] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.170572] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.170622] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.170667] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.170709] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.170986] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.171048] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.171098] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.171140] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.171182] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.171451] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.171472] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.171488] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.171505] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.175714] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.175767] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.175808] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.175858] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.176354] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.176414] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.176457] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.176501] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.176543] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.176941] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.176961] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.176976] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.176992] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.178512] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.659 [2024-07-15 11:45:16.178565] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.178613] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.178656] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.178933] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.178992] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.179035] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.179082] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.179126] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.179401] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.179421] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.179436] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.179452] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.184192] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.184244] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.184285] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.184327] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.184726] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.184787] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.184832] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.184875] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.184918] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.185360] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.185382] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.185397] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.185413] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.186949] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.187004] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.187048] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.187091] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.187418] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.187476] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.187526] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.187568] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.187610] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.187938] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.187958] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.187974] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.187990] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.193097] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.193149] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.193190] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.193232] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.193538] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.193602] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.193645] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.193687] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.193729] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.194185] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.194206] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.194222] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.194238] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.195770] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.195816] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.195861] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.195901] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.196409] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.196473] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.196516] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.196557] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.196598] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.196905] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.196934] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.196950] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.196965] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.202010] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.202078] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.202122] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.202165] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.202435] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.202498] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.202541] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.202582] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.202624] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.203050] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.203071] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.203087] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.203103] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.204697] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.204745] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.204787] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.204828] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.205191] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.205257] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.205305] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.205347] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.205388] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.205690] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.205711] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.205726] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.205742] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.210590] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.210645] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.210708] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.210755] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.211036] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.211103] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.211496] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.211542] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.212413] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.212735] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.660 [2024-07-15 11:45:16.212756] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.212771] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.212786] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.215687] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.215741] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.217320] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.217370] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.217642] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.217706] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.217749] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.218368] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.218418] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.218710] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.218729] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.218749] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.218765] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.222851] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.224434] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.225421] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.227161] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.227443] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.227506] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.229089] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.229141] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.229192] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.229466] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.229486] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.229501] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.229516] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.234842] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.236175] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.237736] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.239302] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.239689] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.241441] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.241493] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.241539] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.243080] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.243359] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.243378] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.243393] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.661 [2024-07-15 11:45:16.243408] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.247378] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.248759] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.250093] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.251663] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.251950] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.252015] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.252058] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.252764] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.252812] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.253177] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.253197] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.253216] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.253232] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.259153] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.259780] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.260178] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.261864] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.262150] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.262211] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.263784] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.263840] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.265472] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.265866] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.265885] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.265900] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.265916] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.270391] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.271824] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.272272] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.272689] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.272980] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.273047] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.274684] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.276433] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.278009] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.278406] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.278425] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.278441] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.278457] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.282874] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.284432] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.284839] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.285340] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.285618] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.287167] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.288820] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.290555] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.291587] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.291907] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.291934] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.925 [2024-07-15 11:45:16.291950] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.291966] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.298303] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.298787] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.299189] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.300758] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.301046] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.302649] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.303118] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.304691] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.306279] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.306559] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.306579] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.306595] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.306612] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.312739] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.314084] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.315667] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.317258] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.317710] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.319141] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.320718] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.322311] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.323561] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.323870] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.323889] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.323905] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.323920] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.328562] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.330148] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.331721] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.332456] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.332737] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.334430] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.336096] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.337631] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.338730] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.339095] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.339114] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.339130] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.339145] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.343218] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.344796] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.345404] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.347108] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.347389] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.349011] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.350591] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.350998] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.352595] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.353066] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.353086] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.353103] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.353123] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.357144] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.357544] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.358844] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.359417] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.359854] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.360270] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.360669] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.362173] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.362573] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.363023] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.363043] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.363061] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.363078] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.367731] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.368143] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.369673] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.370071] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.370490] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.370899] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.371409] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.372776] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.373179] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.373556] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.373574] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.373589] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.373605] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.377504] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.378530] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.379374] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.379771] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.380147] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.380557] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.381843] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.382422] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.382813] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.383099] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.383118] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.383133] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.383148] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.386331] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.388080] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.388473] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.388870] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.389279] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.389687] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.391281] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.391676] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.392399] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.392681] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.392700] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.392715] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.392731] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.397119] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.926 [2024-07-15 11:45:16.398086] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.398478] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.398871] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.399244] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.400374] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.401128] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.401523] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.403070] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.403512] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.403533] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.403549] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.403568] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.409719] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.410142] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.410537] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.410949] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.411338] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.413032] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.413424] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.414010] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.415304] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.415759] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.415779] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.415795] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.415811] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.420830] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.421237] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.421633] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.422037] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.422352] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.423496] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.423893] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.424980] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.425773] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.426219] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.426240] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.426256] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.426277] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.430143] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.430543] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.430945] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.431343] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.431625] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.432497] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.432890] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.434216] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.434751] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.435196] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.435216] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.435232] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.435250] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.438903] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.439310] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.439707] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.440114] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.440396] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.441093] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.441485] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.442965] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.443368] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.443805] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.443827] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.443843] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.443859] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.447207] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.447613] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.448029] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.448429] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.448713] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.449312] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.449701] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.451294] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.451699] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.452145] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.452165] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.452181] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.452197] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.455338] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.455740] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.456149] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.456548] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.456826] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.457242] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.457641] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.459312] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.459705] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.460144] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.460165] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.460180] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.460196] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.463139] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.463543] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.464128] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.465233] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.465580] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.466000] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.466945] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.467868] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.468273] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.468612] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.468631] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.468647] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.468663] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.471535] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.471946] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.472346] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.473615] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.474006] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.474415] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.475523] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.476286] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.927 [2024-07-15 11:45:16.476679] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.477027] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.477047] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.477063] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.477080] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.479986] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.480830] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.481846] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.483033] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.483448] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.484860] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.485322] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.485712] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.487403] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.487864] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.487884] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.487899] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.487916] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.493860] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.494277] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.494673] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.495087] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.495484] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.497193] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.497589] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.498151] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.499463] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.499913] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.499943] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.499959] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.499977] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.505106] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.505506] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.505900] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.506304] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.506621] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.507772] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.508174] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.509291] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.510043] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.510489] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.510509] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.510525] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:32.928 [2024-07-15 11:45:16.510542] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.197 [2024-07-15 11:45:16.516658] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.197 [2024-07-15 11:45:16.518244] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.197 [2024-07-15 11:45:16.519972] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.197 [2024-07-15 11:45:16.520842] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.197 [2024-07-15 11:45:16.521141] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.197 [2024-07-15 11:45:16.521552] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.197 [2024-07-15 11:45:16.522385] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.523416] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.523809] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.524132] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.524151] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.524166] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.524182] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.529551] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.531210] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.532050] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.533095] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.533513] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.534250] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.535391] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.535784] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.537023] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.537334] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.537354] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.537369] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.537385] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.543776] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.544398] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.545862] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.546262] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.546673] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.548228] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.548623] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.549457] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.550782] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.551076] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.551097] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.551112] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.551128] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.556571] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.558298] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.558701] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.559188] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.559466] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.559876] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.560397] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.561733] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.563321] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.563600] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.563619] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.563634] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.563649] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.569582] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.570001] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.570395] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.572029] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.572490] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.572892] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.572952] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.574348] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.574396] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.574672] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.574692] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.574707] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.574722] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.579279] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.580225] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.580291] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.581887] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.582363] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.582767] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.584463] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.584524] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.584909] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.585316] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.585335] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.585350] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.585366] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.589641] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.589694] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.589736] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.589776] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.590058] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.591658] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.591707] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.592330] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.593633] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.594080] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.594103] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.594119] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.594136] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.597716] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.597769] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.597810] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.597852] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.598135] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.598201] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.599042] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.600592] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.600644] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.600918] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.600943] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.600958] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.600974] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.605547] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.605600] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.605640] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.605682] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.606087] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.606496] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.607966] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.608015] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.609523] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.609801] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.198 [2024-07-15 11:45:16.609820] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.609836] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.609851] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.614755] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.614807] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.614849] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.614891] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.615179] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.615882] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.615938] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.616323] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.616376] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.616654] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.616677] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.616692] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.616708] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.620465] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.620525] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.620568] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.620611] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.620887] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.622225] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.622275] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.622316] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.622358] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.622631] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.622650] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.622665] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.622680] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.626487] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.626540] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.626582] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.626624] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.627048] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.627105] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.627149] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.627192] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.627235] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.627522] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.627542] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.627557] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.627572] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.631196] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.631259] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.631304] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.631345] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.631618] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.631680] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.631723] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.631765] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.631808] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.632133] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.632154] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.632169] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.632186] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.637980] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.638036] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.638081] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.638123] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.638412] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.638481] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.638527] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.638569] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.638610] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.638883] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.638902] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.638917] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.638940] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.643448] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.643499] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.643540] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.643582] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.643996] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.644059] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.644107] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.644149] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.644190] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.644524] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.644544] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.644560] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.644576] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.647080] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.647131] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.647172] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.647213] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.647483] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.647545] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.647588] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.647636] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.647681] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.647963] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.647982] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.647997] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.648013] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.652438] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.652489] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.652539] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.652583] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.653058] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.653120] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.653165] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.653210] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.653253] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.653638] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.653661] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.653677] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.199 [2024-07-15 11:45:16.653693] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.658031] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.658082] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.658127] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.658169] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.658574] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.658637] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.658686] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.658728] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.658769] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.659056] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.659075] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.659090] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.659106] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.663606] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.663658] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.663701] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.663743] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.664209] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.664266] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.664311] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.664354] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.664395] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.664824] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.664844] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.664859] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.664875] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.669550] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.669600] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.669645] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.669686] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.669962] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.670023] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.670066] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.670107] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.670148] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.670418] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.670437] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.670452] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.670468] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.673388] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.673440] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.673481] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.673522] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.673792] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.673860] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.673902] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.673951] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.673993] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.674267] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.674286] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.674301] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.674316] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.678468] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.678519] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.678564] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.678605] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.679041] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.679098] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.679142] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.679190] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.679233] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.679683] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.679702] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.679718] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.679735] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.683383] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.683435] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.683477] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.683519] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.683820] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.683881] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.683924] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.683973] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.684014] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.684283] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.684303] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.684318] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.684333] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.688686] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.688748] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.688790] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.688832] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.689174] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.689237] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.689279] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.689321] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.689363] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.689633] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.689652] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.689672] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.689687] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.694565] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.694616] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.694659] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.694702] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.695106] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.695167] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.695210] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.695254] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.695297] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.695716] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.695736] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.695752] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.695769] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.699509] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.699562] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.699608] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.699651] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.699932] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.200 [2024-07-15 11:45:16.699991] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.700033] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.700075] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.700122] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.700395] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.700413] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.700428] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.700444] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.704606] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.704658] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.704701] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.704747] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.705028] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.705086] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.705129] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.705174] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.705222] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.705492] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.705511] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.705526] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.705541] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.710236] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.710287] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.710328] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.710369] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.710767] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.710829] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.710888] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.710937] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.710980] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.711431] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.711450] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.711465] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.711482] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.715104] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.715155] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.715196] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.715239] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.715592] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.715658] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.715705] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.715751] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.715793] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.716109] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.716129] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.716144] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.716159] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.720145] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.720197] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.720243] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.720288] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.720709] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.720770] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.720812] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.720853] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.720894] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.721201] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.721221] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.721236] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.721251] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.725875] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.725930] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.725972] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.726013] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.726286] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.726348] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.726391] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.726441] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.726483] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.726864] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.726883] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.726900] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.726920] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.730855] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.730906] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.730953] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.730995] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.731269] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.731330] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.731372] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.731420] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.731462] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.731779] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.731798] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.731813] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.731829] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.736101] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.736152] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.736193] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.201 [2024-07-15 11:45:16.736235] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.736682] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.736741] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.736785] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.736829] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.736872] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.737203] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.737222] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.737237] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.737252] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.740967] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.741017] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.741080] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.741127] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.741399] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.741462] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.741504] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.741546] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.741587] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.741899] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.741918] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.741940] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.741956] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.744697] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.744748] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.744807] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.744852] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.745129] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.745205] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.745252] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.745294] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.745335] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.745628] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.745647] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.745663] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.745679] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.750854] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.750907] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.750958] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.751001] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.751403] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.751457] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.751500] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.751544] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.751591] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.752031] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.752053] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.752070] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.752086] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.757123] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.757175] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.757216] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.757259] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.757531] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.757595] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.759175] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.759223] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.759983] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.760464] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.760484] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.760499] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.760525] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.765834] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.765892] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.767135] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.767183] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.767495] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.767557] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.767599] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.769179] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.769227] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.769498] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.769517] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.769532] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.769547] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.773877] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.775261] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.776627] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.777962] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.778241] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.778306] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.779169] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.779216] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.779259] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.779701] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.779721] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.779737] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.779753] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.784985] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.786078] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.202 [2024-07-15 11:45:16.787398] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.788970] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.789246] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.790372] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.790421] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.790476] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.790862] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.791309] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.791329] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.791343] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.791360] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.797806] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.798814] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.800144] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.801728] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.802014] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.802079] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.802121] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.802942] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.802999] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.803500] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.803519] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.803534] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.803553] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.808991] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.809845] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.811167] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.812742] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.813023] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.813087] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.814263] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.814309] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.814709] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.815155] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.815175] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.815191] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.815211] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.819423] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.821013] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.822721] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.824298] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.824572] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.824637] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.825041] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.825431] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.825821] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.826261] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.826281] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.826296] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.826315] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.829676] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.830084] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.830495] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.830886] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.831359] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.831762] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.832167] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.832563] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.832960] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.833384] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.833402] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.833418] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.833434] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.836772] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.837178] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.837580] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.837979] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.838421] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.838822] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.464 [2024-07-15 11:45:16.839218] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.839611] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.840011] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.840455] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.840475] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.840491] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.840507] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.843958] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.844361] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.844756] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.845156] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.845540] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.845957] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.846347] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.846737] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.847137] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.847563] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.847581] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.847596] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.847611] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.850973] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.851372] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.851762] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.852157] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.852536] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.852949] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.853344] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.853732] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.854142] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.854587] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.854606] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.854622] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.854638] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.858022] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.858426] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.858820] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.859217] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.859647] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.860065] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.860467] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.860864] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.861264] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.861660] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.861679] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.861694] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.861710] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.865020] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.865418] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.865816] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.866216] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.866647] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.867055] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.867447] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.867850] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.868253] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.868714] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.868733] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.868750] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.868766] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.872314] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.872719] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.873120] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.873519] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.873948] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.874353] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.874741] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.875137] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.875530] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.875931] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.875954] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.875971] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.875986] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.879366] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.879764] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.880172] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.880567] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.880916] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.881331] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.881736] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.882131] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.882521] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.882963] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.882985] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.883002] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.883018] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.465 [2024-07-15 11:45:16.886532] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.886943] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.887334] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.887722] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.888156] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.888558] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.888960] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.889353] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.889747] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.890164] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.890184] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.890200] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.890215] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.893571] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.893986] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.894385] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.894777] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.895189] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.895592] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.895991] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.897547] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.898168] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.898441] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.898460] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.898475] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.898490] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.901898] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.902307] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.902694] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.903091] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.903486] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.903893] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.904292] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.904679] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.905074] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.905524] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.905543] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.905559] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.905575] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.909074] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.909481] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.909871] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.910267] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.910699] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.911751] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.913082] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.914648] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.916226] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.916599] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.916618] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.916633] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.916648] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.918767] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.919167] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.919557] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.919952] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.920315] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.921645] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.923222] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.924791] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.925911] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.926191] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.926209] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.926224] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.926240] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.928312] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.928707] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.929100] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.929935] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.930251] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.931905] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.933487] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.934890] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.936252] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.936582] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.936603] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.936619] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.936634] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.938828] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.939241] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.939753] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.941091] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.466 [2024-07-15 11:45:16.941366] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.943002] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.944728] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.945745] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.947079] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.947354] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.947372] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.947388] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.947403] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.949850] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.950246] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.951882] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.953585] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.953861] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.955451] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.956198] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.957530] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.959109] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.959386] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.959404] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.959419] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.959434] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.962018] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.963573] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.965023] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.966595] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.966871] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.967643] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.969001] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.970572] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.972157] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.972430] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.972449] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.972465] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.972480] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.976171] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.977505] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.979081] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.980656] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.981036] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.982603] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.984273] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.985850] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.987285] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.987673] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.987692] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.987707] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.987722] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.991268] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.992840] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.994403] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.995146] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.995420] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.996863] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:16.998436] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.000104] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.000511] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.000952] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.000971] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.000987] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.001004] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.004600] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.006183] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.007124] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.008768] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.009047] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.010633] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.012216] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.012616] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.013011] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.013408] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.013427] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.013442] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.013458] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.016919] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.018144] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.019680] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.021082] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.021356] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.022959] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.023642] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.024038] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.024431] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.024862] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.024885] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.024908] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.024924] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.028098] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.029329] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.467 [2024-07-15 11:45:17.030662] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.032239] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.032515] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.033510] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.033916] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.034309] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.034699] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.035123] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.035141] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.035156] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.035171] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.037496] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.038835] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.040402] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.041948] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.042283] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.042692] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.043088] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.043479] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.043869] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.044150] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.044168] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.044184] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.044199] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.047140] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.048712] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.050298] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.051508] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.051915] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.052323] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.052713] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.053108] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.054782] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.055085] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.055103] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.055118] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.468 [2024-07-15 11:45:17.055133] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.058442] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.060077] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.061572] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.061968] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.062414] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.062814] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.063209] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.064593] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.065941] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.066217] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.066235] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.066250] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.066265] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.069465] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.071151] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.071551] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.071943] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.072340] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.072740] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.072787] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.074119] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.074167] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.074477] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.074495] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.074510] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.074525] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.076173] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.077870] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.077930] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.079488] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.079765] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.080180] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.080574] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.080622] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.081014] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.732 [2024-07-15 11:45:17.081453] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.081472] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.081489] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.081508] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.083110] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.083156] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.083197] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.083245] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.083637] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.084976] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.085025] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.086597] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.088174] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.088510] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.088529] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.088548] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.088564] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.091124] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.091173] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.091213] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.091254] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.091568] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.091630] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.093211] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.094788] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.094835] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.095134] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.095153] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.095169] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.095185] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.096728] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.096773] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.096813] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.096854] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.097294] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.097702] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.098098] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.098143] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.098531] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.098890] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.098908] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.098923] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.098944] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.100535] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.100582] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.100623] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.100676] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.100954] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.102379] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.102427] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.104148] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.104212] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.104480] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.104498] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.104513] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.104528] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.106808] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.106855] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.106897] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.106945] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.107216] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.108541] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.108589] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.108629] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.108676] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.108954] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.108973] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.108987] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.109002] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.110642] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.110688] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.110728] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.110770] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.111042] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.111102] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.111144] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.111186] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.111232] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.111615] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.111634] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.111649] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.111664] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.114100] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.114145] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.114185] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.114226] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.114537] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.114597] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.114640] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.114681] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.114722] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.114992] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.115010] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.115025] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.115040] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.116680] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.116726] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.116779] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.116824] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.117103] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.733 [2024-07-15 11:45:17.117164] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.117210] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.117252] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.117293] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.117571] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.117590] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.117605] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.117626] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.120009] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.120074] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.120118] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.120159] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.120430] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.120489] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.120536] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.120577] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.120618] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.120885] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.120904] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.120918] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.120939] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.122559] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.122604] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.122645] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.122685] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.122958] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.123019] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.123062] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.123102] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.123152] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.123423] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.123441] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.123456] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.123471] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.125811] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.125858] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.125901] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.125948] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.126225] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.126285] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.126336] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.126379] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.126433] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.126701] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.126719] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.126733] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.126748] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.128416] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.128463] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.128508] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.128549] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.128818] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.128877] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.128920] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.128969] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.129010] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.129280] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.129299] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.129313] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.129329] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.131682] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.131729] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.131773] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.131815] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.132136] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.132193] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.132235] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.132277] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.132322] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.132624] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.132643] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.132657] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.132672] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.134329] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.134382] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.134423] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.134465] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.134735] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.134796] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.134838] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.134880] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.134921] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.135199] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.135217] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.135232] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.135248] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.137513] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.137561] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.137603] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.137645] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.138046] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.138108] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.138150] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.138191] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.138232] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.138556] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.138575] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.138591] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.138607] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.734 [2024-07-15 11:45:17.140233] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.140280] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.140331] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.140376] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.140648] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.140703] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.140752] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.140797] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.140838] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.141114] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.141133] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.141149] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.141164] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.143414] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.143462] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.143505] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.143547] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.144010] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.144073] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.144119] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.144161] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.144201] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.144497] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.144516] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.144532] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.144548] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.146169] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.146215] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.146257] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.146298] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.146613] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.146673] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.146715] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.146758] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.146806] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.147081] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.147100] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.147115] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.147131] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.149298] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.149344] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.149385] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.149428] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.149869] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.149935] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.149980] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.150023] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.150068] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.150340] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.150358] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.150373] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.150389] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.152012] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.152057] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.152103] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.152144] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.152442] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.152502] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.152545] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.152587] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.152629] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.152901] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.152919] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.152942] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.152958] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.155151] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.155197] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.155238] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.155279] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.155700] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.155754] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.155797] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.155839] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.155881] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.156217] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.156236] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.156250] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.156266] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.157920] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.157972] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.158012] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.158061] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.158334] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.158389] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.158440] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.158487] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.158529] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.158797] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.158816] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.158832] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.158849] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.161288] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.161340] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.161382] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.161428] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.161699] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.161758] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.161807] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.161862] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.161908] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.735 [2024-07-15 11:45:17.162186] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.162206] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.162221] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.162237] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.163861] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.163906] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.163959] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.164001] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.164269] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.164330] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.164372] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.164415] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.164456] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.164721] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.164740] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.164755] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.164771] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.167146] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.167195] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.167237] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.167280] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.167584] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.167644] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.167686] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.167727] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.167780] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.168064] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.168083] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.168099] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.168115] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.169766] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.169817] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.169857] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.169898] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.170175] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.170234] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.170276] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.170317] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.170359] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.170625] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.170643] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.170658] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.170674] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.172942] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.172990] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.173035] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.173077] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.173465] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.173525] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.173568] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.173609] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.173650] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.173991] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.174014] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.174029] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.174045] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.175662] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.175708] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.175760] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.175803] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.176080] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.176135] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.176180] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.176232] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.176273] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.176540] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.176559] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.176574] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.176590] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.178797] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.178844] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.178888] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.178937] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.179370] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.179424] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.179471] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.179513] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.179554] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.179829] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.179847] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.736 [2024-07-15 11:45:17.179862] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.179879] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.181570] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.181620] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.181661] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.181702] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.182017] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.182077] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.182119] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.182161] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.182202] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.182469] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.182487] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.182502] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.182517] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.184648] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.184711] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.184753] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.184795] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.185231] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.185290] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.185686] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.185736] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.186135] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.186534] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.186553] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.186569] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.186585] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.189296] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.189349] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.189745] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.189790] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.190242] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.190303] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.190351] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.190739] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.190783] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.191211] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.191230] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.191245] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.191260] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.193882] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.194286] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.194682] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.195081] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.195518] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.195576] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.195972] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.196018] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.196061] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.196478] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.196497] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.196513] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.196528] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.199097] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.199497] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.199898] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.200303] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.200738] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.201147] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.201196] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.201238] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.201625] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.201976] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.202001] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.202016] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.202032] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.204694] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.205100] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.205501] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.205893] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.206331] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.206406] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.206452] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.206840] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.206884] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.207304] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.207323] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.207338] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.207353] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.209986] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.210389] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.210786] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.211185] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.211626] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.211684] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.212080] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.212127] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.212515] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.212891] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.212912] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.212934] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.212950] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.215604] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.216012] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.216413] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.216804] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.217211] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.217268] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.737 [2024-07-15 11:45:17.217656] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.218058] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.218453] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.218851] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.218870] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.218885] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.218901] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.221614] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.222019] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.222409] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.222801] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.223250] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.223655] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.224055] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.224458] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.224849] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.225329] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.225349] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.225365] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.225381] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.227940] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.228331] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.228725] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.229126] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.229489] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.229893] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.230289] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.230685] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.231083] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.231440] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.231459] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.231474] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.231490] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.234166] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.234570] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.234969] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.235360] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.235839] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.236245] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.236645] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.237046] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.237442] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.237866] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.237884] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.237899] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.237916] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.240533] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.240933] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.241323] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.241717] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.242095] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.242502] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.242893] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.243284] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.243674] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.244028] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.244048] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.244069] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.244085] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.246668] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.247070] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.247469] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.247864] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.248303] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.248713] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.249110] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.249505] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.249907] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.250357] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.250376] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.250392] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.250408] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.252995] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.253392] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.253782] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.254176] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.254528] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.256008] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.256741] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.257865] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.258261] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.258698] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.258717] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.258733] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.258752] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.261435] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.261833] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.262229] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.262625] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.263001] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.263407] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.263800] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.264193] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.264584] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.265027] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.265046] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.265062] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.265077] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.267616] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.268018] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.268415] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.738 [2024-07-15 11:45:17.268809] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.269246] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.269646] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.270040] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.270431] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.270830] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.271276] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.271296] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.271311] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.271327] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.274762] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.276345] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.277923] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.278731] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.279045] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.280710] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.282288] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.283675] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.284077] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.284509] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.284529] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.284544] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.284560] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.288166] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.289745] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.290468] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.291849] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.292127] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.293738] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.295462] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.295860] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.296300] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.296695] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.296714] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.296729] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.296744] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.300085] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.301215] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.302860] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.304389] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.304663] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.306262] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.306841] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.307237] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.307627] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.308067] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.308086] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.308102] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.308124] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.311253] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.312505] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.313842] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.315420] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.315691] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.316691] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.317104] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.317497] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.317897] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.318353] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.318373] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.318388] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.318404] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:33.739 [2024-07-15 11:45:17.320615] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.321961] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.323528] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.325105] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.325384] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.325791] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.326189] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.326579] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.326980] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.327259] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.327277] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.327292] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.327306] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.330423] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.332061] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.333644] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.335024] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.335404] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.335809] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.336209] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.336602] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.337897] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.338233] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.338251] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.338266] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.338282] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.341176] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.342742] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.344311] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.344717] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.345170] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.345573] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.345970] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.346802] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.348129] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.348405] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.348423] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.348438] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.348453] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.351601] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.353168] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.354020] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.354415] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.354864] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.355276] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.355667] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.357336] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.359015] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.359290] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.359309] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.359323] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.359338] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.362538] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.363946] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.364334] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.364719] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.365113] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.365512] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.366923] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.368247] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.369828] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.370112] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.370130] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.370145] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.370160] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.373391] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.373793] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.374193] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.374586] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.375016] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.375876] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.002 [2024-07-15 11:45:17.377196] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.378774] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.380353] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.380690] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.380709] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.380724] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.380739] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.383098] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.383497] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.383886] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.384281] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.384686] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.386105] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.387664] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.389235] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.390549] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.390869] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.390887] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.390901] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.390916] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.392903] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.393306] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.393701] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.394101] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.394375] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.395712] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.397300] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.398869] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.399698] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.400043] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.400062] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.400077] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.400092] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.402157] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.402552] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.402946] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.404482] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.404807] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.406408] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.407987] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.408713] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.410101] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.410376] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.410394] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.410409] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.410423] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.412695] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.413099] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.414163] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.415480] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.415752] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.417360] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.418477] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.420124] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.421639] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.421913] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.421936] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.421951] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.421967] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.424447] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.425231] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.426551] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.428114] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.428388] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.429884] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.431159] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.432478] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.434045] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.434323] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.434342] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.434356] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.434372] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.437033] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.438583] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.440266] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.441838] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.442117] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.442891] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.444226] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.445810] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.447391] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.447712] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.447731] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.447747] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.447762] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.451658] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.453094] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.454669] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.456341] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.456765] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.458144] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.459716] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.461294] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.462549] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.462917] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.462941] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.462956] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.462971] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.466568] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.468157] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.003 [2024-07-15 11:45:17.469736] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.470452] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.470724] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.472166] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.472215] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.473942] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.473993] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.474267] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.474285] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.474299] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.474315] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.476662] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.477956] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.478003] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.479319] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.479594] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.481192] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.481974] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.482022] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.483348] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.483621] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.483639] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.483654] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.483669] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.485613] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.485660] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.485702] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.485744] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.486112] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.486522] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.486571] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.488131] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.489715] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.489995] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.490014] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.490028] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.490044] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.491705] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.491751] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.491791] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.491832] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.492108] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.492168] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.492961] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.493350] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.493395] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.493801] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.493821] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.493836] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.493853] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.495775] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.495825] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.495866] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.495907] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.496184] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.497553] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.498948] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.498996] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.500421] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.500699] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.500717] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.500732] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.500747] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.502898] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.502961] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.503003] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.503045] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.503472] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.504163] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.504211] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.505536] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.505583] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.505855] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.505874] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.505888] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.505903] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.507526] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.507572] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.507614] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.507662] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.507942] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.509528] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.509579] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.509627] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.509670] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.510049] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.510068] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.510084] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.510101] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.512404] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.512453] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.512494] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.512543] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.512820] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.512877] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.512919] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.512967] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.513026] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.513297] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.004 [2024-07-15 11:45:17.513316] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.513330] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.513345] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.514998] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.515043] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.515083] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.515124] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.515391] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.515450] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.515492] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.515533] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.515574] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.515930] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.515949] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.515964] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.515980] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.518669] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.518718] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.518763] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.518804] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.519106] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.519177] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.519226] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.519268] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.519309] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.519579] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.519597] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.519611] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.519627] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.521210] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.521266] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.521307] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.521348] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.521614] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.521674] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.521716] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.521762] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.521808] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.522083] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.522101] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.522116] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.522133] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.524441] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.524487] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.524529] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.524570] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.524841] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.524898] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.524946] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.524997] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.525039] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.525311] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.525336] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.525351] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.525366] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.527064] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.527113] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.527154] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.527195] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.527464] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.527529] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.527573] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.527614] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.527655] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.527919] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.527944] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.527958] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.527974] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.530155] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.530202] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.530248] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.530292] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.530738] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.530793] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.530835] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.530876] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.530917] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.531237] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.531256] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.531271] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.531285] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.532891] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.532942] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.532988] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.533030] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.533296] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.533359] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.533401] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.533442] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.533483] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.533913] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.533937] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.533953] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.533969] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.536407] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.536453] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.536494] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.536535] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.536845] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.536907] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.005 [2024-07-15 11:45:17.536956] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.536998] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.537039] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.537312] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.537331] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.537346] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.537362] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.538992] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.539038] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.539080] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.539140] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.539411] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.539466] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.539524] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.539565] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.539607] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.539875] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.539894] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.539909] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.539931] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.542223] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.542269] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.542316] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.542358] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.542629] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.542685] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.542728] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.542775] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.542816] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.543092] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.543111] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.543126] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.543141] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.544826] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.544871] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.544912] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.544959] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.545229] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.545289] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.545332] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.545373] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.545415] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.545682] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.545700] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.545718] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.545733] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.547970] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.548017] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.548059] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.548102] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.548497] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.548554] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.548596] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.548637] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.548678] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.548995] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.549013] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.549028] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.549043] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.550647] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.550694] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.550735] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.550782] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.551058] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.551116] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.551159] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.551209] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.551254] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.551526] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.551545] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.551560] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.551575] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.553599] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.553646] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.553695] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.553739] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.554171] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.554224] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.554267] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.006 [2024-07-15 11:45:17.554310] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.554352] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.554625] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.554643] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.554658] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.554674] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.556282] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.556327] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.556368] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.556409] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.556723] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.556785] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.556828] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.556870] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.556912] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.557190] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.557209] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.557223] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.557238] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.559206] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.559254] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.559299] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.559340] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.559798] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.559851] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.559895] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.559948] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.559991] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.560357] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.560375] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.560390] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.560405] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.561936] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.561982] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.562023] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.562065] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.562374] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.562434] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.562476] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.562518] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.562558] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.562856] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.562874] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.562889] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.562905] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.564718] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.564766] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.564813] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.564855] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.565290] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.565363] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.565418] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.565460] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.565501] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.565959] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.565978] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.565998] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.566014] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.568164] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.568211] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.568254] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.568296] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.568744] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.568799] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.568842] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.568883] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.568933] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.569374] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.569393] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.569407] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.569424] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.571868] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.571933] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.571977] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.572018] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.572449] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.572500] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.572545] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.572587] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.572628] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.572998] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.573019] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.573034] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.573049] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.575242] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.575288] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.575332] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.575379] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.575813] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.575871] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.575916] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.575995] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.576039] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.576540] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.576559] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.576575] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.576591] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.578839] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.578887] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.578943] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.007 [2024-07-15 11:45:17.578987] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.579394] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.579460] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.579519] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.579565] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.579608] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.579973] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.579993] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.580009] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.580025] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.582353] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.582404] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.582447] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.582490] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.582852] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.582916] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.582974] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.583029] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.583098] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.583489] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.583508] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.583524] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.583539] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.586042] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.586096] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.586161] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.586216] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.586632] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.586708] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.586766] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.586808] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.586851] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.587260] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.587279] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.587295] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.587311] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.589579] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.589624] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.589667] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.589711] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.590085] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.590155] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.590562] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.590613] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.591014] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.591513] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.591532] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.591550] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.008 [2024-07-15 11:45:17.591571] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.594143] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.594196] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.594580] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.594625] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.595032] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.595099] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.595157] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.595553] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.595609] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.595970] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.595989] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.596005] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.596021] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.598680] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.599087] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.599480] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.599869] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.600320] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.600381] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.600772] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.600822] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.600865] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.601213] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.601234] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.601249] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.601264] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.604075] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.604474] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.604868] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.605264] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.605730] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.606140] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.606218] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.606262] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.606660] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.607045] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.607066] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.607081] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.607098] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.609970] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.610388] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.610784] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.611180] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.611632] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.611692] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.611735] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.612134] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.612186] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.612525] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.612544] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.612559] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.612575] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.269 [2024-07-15 11:45:17.615327] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.615729] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.616129] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.616520] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.616962] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.617022] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.617411] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.617466] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.617865] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.618247] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.618266] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.618280] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.618295] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.621262] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.621665] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.622066] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.622456] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.622890] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.622956] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.623349] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.623744] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.624149] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.624615] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.624634] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.624650] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.624666] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.627232] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.627630] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.628028] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.628419] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.628777] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.629188] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.629585] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.629986] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.630378] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.630818] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.630839] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.630854] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.630870] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.633555] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.633960] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.634357] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.634754] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.635235] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.635635] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.636045] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.636444] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.636841] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.637227] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.637246] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.637260] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.637276] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.640127] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.640532] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.640937] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.641330] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.641780] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.642191] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.643922] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.644425] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.645796] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.646224] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.646243] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.646260] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.646276] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.649023] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.649424] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.649817] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.650217] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.650656] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.651069] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.651466] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.651857] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.652249] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.652719] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.652738] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.652754] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.652771] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.655358] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.655761] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.656165] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.656577] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.656968] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.270 [2024-07-15 11:45:17.657372] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.657765] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.658165] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.658566] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.658983] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.659002] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.659017] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.659032] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.662242] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.663587] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.665168] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.666744] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.667139] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.668721] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.670164] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.671733] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.673411] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.673793] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.673813] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.673829] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.673844] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.677492] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.679086] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.680664] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.681857] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.682145] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.683482] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.685068] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.686650] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.687243] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.687698] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.687718] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.687733] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.687750] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.691572] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.693162] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.694597] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.695944] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.696276] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.697876] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.699453] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.700291] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.700689] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.701131] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.701152] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.701168] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.701184] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.704814] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.706396] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.707211] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.708541] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.708818] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.710428] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.711790] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.712191] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.712590] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.712998] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.713017] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.713033] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.713049] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.716444] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.717200] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.718563] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.720146] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.720424] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.722168] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.722564] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.722963] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.723357] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.723779] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.723798] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.723818] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.723835] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.726688] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.728207] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.729667] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.731252] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.271 [2024-07-15 11:45:17.731530] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.732163] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.732558] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.732957] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.733352] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.733741] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.733759] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.733773] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.733789] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.736562] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.737888] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.739468] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.741051] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.741427] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.741859] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.742258] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.742654] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.743124] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.743398] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.743417] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.743432] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.743448] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.746349] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.747909] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.749483] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.750688] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.751076] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.751481] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.751873] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.752275] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.753732] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.754053] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.754076] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.754091] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.754106] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.757383] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.759086] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.760629] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.761028] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.761463] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.761870] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.762271] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.762851] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.764167] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.764444] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.764463] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.764478] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.272 [2024-07-15 11:45:17.764493] accel_dpdk_cryptodev.c: 468:accel_dpdk_cryptodev_task_alloc_resources: *ERROR*: Failed to get src_mbufs! 00:34:34.839 00:34:34.839 Latency(us) 00:34:34.839 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:34:34.839 Job: crypto_ram (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:34:34.839 Verification LBA range: start 0x0 length 0x100 00:34:34.839 crypto_ram : 6.06 42.25 2.64 0.00 0.00 2944685.63 266247.12 2596821.26 00:34:34.839 Job: crypto_ram (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:34:34.839 Verification LBA range: start 0x100 length 0x100 00:34:34.839 crypto_ram : 5.98 42.82 2.68 0.00 0.00 2899077.57 310013.77 2407165.77 00:34:34.839 Job: crypto_ram1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:34:34.839 Verification LBA range: start 0x0 length 0x100 00:34:34.839 crypto_ram1 : 6.06 42.24 2.64 0.00 0.00 2844999.68 264423.51 2392576.89 00:34:34.839 Job: crypto_ram1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:34:34.839 Verification LBA range: start 0x100 length 0x100 00:34:34.839 crypto_ram1 : 5.98 42.81 2.68 0.00 0.00 2801777.98 282659.62 2217510.29 00:34:34.839 Job: crypto_ram2 (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:34:34.839 Verification LBA range: start 0x0 length 0x100 00:34:34.839 crypto_ram2 : 5.62 263.43 16.46 0.00 0.00 434031.60 56303.97 649205.31 00:34:34.839 Job: crypto_ram2 (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:34:34.839 Verification LBA range: start 0x100 length 0x100 00:34:34.839 crypto_ram2 : 5.60 280.80 17.55 0.00 0.00 407465.31 25986.45 627321.99 00:34:34.839 Job: crypto_ram3 (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:34:34.839 Verification LBA range: start 0x0 length 0x100 00:34:34.839 crypto_ram3 : 5.70 270.51 16.91 0.00 0.00 410737.85 9858.89 366545.70 00:34:34.839 Job: crypto_ram3 (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:34:34.839 Verification LBA range: start 0x100 length 0x100 00:34:34.839 crypto_ram3 : 5.70 291.69 18.23 0.00 0.00 382518.33 63370.46 474138.71 00:34:34.839 =================================================================================================================== 00:34:34.840 Total : 1276.54 79.78 0.00 0.00 754565.27 9858.89 2596821.26 00:34:35.407 00:34:35.407 real 0m9.284s 00:34:35.407 user 0m17.585s 00:34:35.407 sys 0m0.482s 00:34:35.407 11:45:18 blockdev_crypto_qat.bdev_verify_big_io -- common/autotest_common.sh@1124 -- # xtrace_disable 00:34:35.407 11:45:18 blockdev_crypto_qat.bdev_verify_big_io -- common/autotest_common.sh@10 -- # set +x 00:34:35.407 ************************************ 00:34:35.407 END TEST bdev_verify_big_io 00:34:35.407 ************************************ 00:34:35.407 11:45:18 blockdev_crypto_qat -- common/autotest_common.sh@1142 -- # return 0 00:34:35.407 11:45:18 blockdev_crypto_qat -- bdev/blockdev.sh@779 -- # run_test bdev_write_zeroes /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:34:35.407 11:45:18 blockdev_crypto_qat -- common/autotest_common.sh@1099 -- # '[' 13 -le 1 ']' 00:34:35.407 11:45:18 blockdev_crypto_qat -- common/autotest_common.sh@1105 -- # xtrace_disable 00:34:35.407 11:45:18 blockdev_crypto_qat -- common/autotest_common.sh@10 -- # set +x 00:34:35.407 ************************************ 00:34:35.407 START TEST bdev_write_zeroes 00:34:35.407 ************************************ 00:34:35.407 11:45:18 blockdev_crypto_qat.bdev_write_zeroes -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:34:35.665 [2024-07-15 11:45:19.003245] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:34:35.665 [2024-07-15 11:45:19.003313] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1069602 ] 00:34:35.665 [2024-07-15 11:45:19.134714] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:35.665 [2024-07-15 11:45:19.235900] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:34:35.665 [2024-07-15 11:45:19.257198] accel_dpdk_cryptodev.c: 223:accel_dpdk_cryptodev_set_driver: *NOTICE*: Using driver crypto_qat 00:34:35.923 [2024-07-15 11:45:19.265226] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation encrypt will be assigned to module dpdk_cryptodev 00:34:35.923 [2024-07-15 11:45:19.273245] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation decrypt will be assigned to module dpdk_cryptodev 00:34:35.923 [2024-07-15 11:45:19.381557] accel_dpdk_cryptodev.c:1178:accel_dpdk_cryptodev_init: *NOTICE*: Found crypto devices: 96 00:34:38.502 [2024-07-15 11:45:21.601615] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_cbc" 00:34:38.502 [2024-07-15 11:45:21.601681] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc0 00:34:38.502 [2024-07-15 11:45:21.601696] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:34:38.502 [2024-07-15 11:45:21.609633] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_xts" 00:34:38.502 [2024-07-15 11:45:21.609654] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc1 00:34:38.502 [2024-07-15 11:45:21.609667] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:34:38.502 [2024-07-15 11:45:21.617652] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_cbc2" 00:34:38.503 [2024-07-15 11:45:21.617671] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc2 00:34:38.503 [2024-07-15 11:45:21.617682] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:34:38.503 [2024-07-15 11:45:21.625673] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "test_dek_qat_xts2" 00:34:38.503 [2024-07-15 11:45:21.625696] bdev.c:8157:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: Malloc3 00:34:38.503 [2024-07-15 11:45:21.625708] vbdev_crypto.c: 617:create_crypto_disk: *NOTICE*: vbdev creation deferred pending base bdev arrival 00:34:38.503 Running I/O for 1 seconds... 00:34:39.436 00:34:39.436 Latency(us) 00:34:39.436 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:34:39.436 Job: crypto_ram (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:34:39.436 crypto_ram : 1.02 2021.52 7.90 0.00 0.00 62838.02 5584.81 76135.74 00:34:39.436 Job: crypto_ram1 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:34:39.436 crypto_ram1 : 1.03 2027.04 7.92 0.00 0.00 62311.23 5556.31 70664.90 00:34:39.436 Job: crypto_ram2 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:34:39.436 crypto_ram2 : 1.02 15567.30 60.81 0.00 0.00 8096.32 2436.23 10656.72 00:34:39.436 Job: crypto_ram3 (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:34:39.436 crypto_ram3 : 1.02 15599.29 60.93 0.00 0.00 8053.58 2421.98 8491.19 00:34:39.436 =================================================================================================================== 00:34:39.436 Total : 35215.15 137.56 0.00 0.00 14367.47 2421.98 76135.74 00:34:39.694 00:34:39.694 real 0m4.218s 00:34:39.694 user 0m3.791s 00:34:39.694 sys 0m0.376s 00:34:39.694 11:45:23 blockdev_crypto_qat.bdev_write_zeroes -- common/autotest_common.sh@1124 -- # xtrace_disable 00:34:39.694 11:45:23 blockdev_crypto_qat.bdev_write_zeroes -- common/autotest_common.sh@10 -- # set +x 00:34:39.694 ************************************ 00:34:39.694 END TEST bdev_write_zeroes 00:34:39.694 ************************************ 00:34:39.694 11:45:23 blockdev_crypto_qat -- common/autotest_common.sh@1142 -- # return 0 00:34:39.694 11:45:23 blockdev_crypto_qat -- bdev/blockdev.sh@782 -- # run_test bdev_json_nonenclosed /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:34:39.694 11:45:23 blockdev_crypto_qat -- common/autotest_common.sh@1099 -- # '[' 13 -le 1 ']' 00:34:39.694 11:45:23 blockdev_crypto_qat -- common/autotest_common.sh@1105 -- # xtrace_disable 00:34:39.695 11:45:23 blockdev_crypto_qat -- common/autotest_common.sh@10 -- # set +x 00:34:39.695 ************************************ 00:34:39.695 START TEST bdev_json_nonenclosed 00:34:39.695 ************************************ 00:34:39.695 11:45:23 blockdev_crypto_qat.bdev_json_nonenclosed -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:34:39.952 [2024-07-15 11:45:23.310047] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:34:39.952 [2024-07-15 11:45:23.310115] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1070211 ] 00:34:39.952 [2024-07-15 11:45:23.442266] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:40.209 [2024-07-15 11:45:23.547455] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:34:40.209 [2024-07-15 11:45:23.547527] json_config.c: 608:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: not enclosed in {}. 00:34:40.209 [2024-07-15 11:45:23.547548] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:34:40.209 [2024-07-15 11:45:23.547562] app.c:1053:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:34:40.209 00:34:40.209 real 0m0.408s 00:34:40.209 user 0m0.238s 00:34:40.209 sys 0m0.166s 00:34:40.209 11:45:23 blockdev_crypto_qat.bdev_json_nonenclosed -- common/autotest_common.sh@1123 -- # es=234 00:34:40.209 11:45:23 blockdev_crypto_qat.bdev_json_nonenclosed -- common/autotest_common.sh@1124 -- # xtrace_disable 00:34:40.209 11:45:23 blockdev_crypto_qat.bdev_json_nonenclosed -- common/autotest_common.sh@10 -- # set +x 00:34:40.209 ************************************ 00:34:40.209 END TEST bdev_json_nonenclosed 00:34:40.209 ************************************ 00:34:40.209 11:45:23 blockdev_crypto_qat -- common/autotest_common.sh@1142 -- # return 234 00:34:40.209 11:45:23 blockdev_crypto_qat -- bdev/blockdev.sh@782 -- # true 00:34:40.209 11:45:23 blockdev_crypto_qat -- bdev/blockdev.sh@785 -- # run_test bdev_json_nonarray /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:34:40.209 11:45:23 blockdev_crypto_qat -- common/autotest_common.sh@1099 -- # '[' 13 -le 1 ']' 00:34:40.209 11:45:23 blockdev_crypto_qat -- common/autotest_common.sh@1105 -- # xtrace_disable 00:34:40.209 11:45:23 blockdev_crypto_qat -- common/autotest_common.sh@10 -- # set +x 00:34:40.209 ************************************ 00:34:40.209 START TEST bdev_json_nonarray 00:34:40.209 ************************************ 00:34:40.209 11:45:23 blockdev_crypto_qat.bdev_json_nonarray -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf --json /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:34:40.209 [2024-07-15 11:45:23.794454] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:34:40.209 [2024-07-15 11:45:23.794516] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1070331 ] 00:34:40.467 [2024-07-15 11:45:23.923220] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:40.467 [2024-07-15 11:45:24.030580] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:34:40.467 [2024-07-15 11:45:24.030656] json_config.c: 614:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: 'subsystems' should be an array. 00:34:40.467 [2024-07-15 11:45:24.030677] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:34:40.467 [2024-07-15 11:45:24.030690] app.c:1053:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:34:40.725 00:34:40.725 real 0m0.408s 00:34:40.725 user 0m0.245s 00:34:40.725 sys 0m0.159s 00:34:40.725 11:45:24 blockdev_crypto_qat.bdev_json_nonarray -- common/autotest_common.sh@1123 -- # es=234 00:34:40.725 11:45:24 blockdev_crypto_qat.bdev_json_nonarray -- common/autotest_common.sh@1124 -- # xtrace_disable 00:34:40.725 11:45:24 blockdev_crypto_qat.bdev_json_nonarray -- common/autotest_common.sh@10 -- # set +x 00:34:40.725 ************************************ 00:34:40.725 END TEST bdev_json_nonarray 00:34:40.725 ************************************ 00:34:40.725 11:45:24 blockdev_crypto_qat -- common/autotest_common.sh@1142 -- # return 234 00:34:40.725 11:45:24 blockdev_crypto_qat -- bdev/blockdev.sh@785 -- # true 00:34:40.725 11:45:24 blockdev_crypto_qat -- bdev/blockdev.sh@787 -- # [[ crypto_qat == bdev ]] 00:34:40.725 11:45:24 blockdev_crypto_qat -- bdev/blockdev.sh@794 -- # [[ crypto_qat == gpt ]] 00:34:40.725 11:45:24 blockdev_crypto_qat -- bdev/blockdev.sh@798 -- # [[ crypto_qat == crypto_sw ]] 00:34:40.725 11:45:24 blockdev_crypto_qat -- bdev/blockdev.sh@810 -- # trap - SIGINT SIGTERM EXIT 00:34:40.725 11:45:24 blockdev_crypto_qat -- bdev/blockdev.sh@811 -- # cleanup 00:34:40.725 11:45:24 blockdev_crypto_qat -- bdev/blockdev.sh@23 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/aiofile 00:34:40.725 11:45:24 blockdev_crypto_qat -- bdev/blockdev.sh@24 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/bdev.json 00:34:40.725 11:45:24 blockdev_crypto_qat -- bdev/blockdev.sh@26 -- # [[ crypto_qat == rbd ]] 00:34:40.725 11:45:24 blockdev_crypto_qat -- bdev/blockdev.sh@30 -- # [[ crypto_qat == daos ]] 00:34:40.725 11:45:24 blockdev_crypto_qat -- bdev/blockdev.sh@34 -- # [[ crypto_qat = \g\p\t ]] 00:34:40.725 11:45:24 blockdev_crypto_qat -- bdev/blockdev.sh@40 -- # [[ crypto_qat == xnvme ]] 00:34:40.725 00:34:40.725 real 1m12.657s 00:34:40.725 user 2m41.110s 00:34:40.725 sys 0m9.173s 00:34:40.725 11:45:24 blockdev_crypto_qat -- common/autotest_common.sh@1124 -- # xtrace_disable 00:34:40.725 11:45:24 blockdev_crypto_qat -- common/autotest_common.sh@10 -- # set +x 00:34:40.725 ************************************ 00:34:40.725 END TEST blockdev_crypto_qat 00:34:40.725 ************************************ 00:34:40.725 11:45:24 -- common/autotest_common.sh@1142 -- # return 0 00:34:40.725 11:45:24 -- spdk/autotest.sh@360 -- # run_test chaining /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/chaining.sh 00:34:40.725 11:45:24 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:34:40.725 11:45:24 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:34:40.725 11:45:24 -- common/autotest_common.sh@10 -- # set +x 00:34:40.725 ************************************ 00:34:40.725 START TEST chaining 00:34:40.725 ************************************ 00:34:40.725 11:45:24 chaining -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev/chaining.sh 00:34:40.983 * Looking for test storage... 00:34:40.983 * Found test storage at /var/jenkins/workspace/crypto-phy-autotest/spdk/test/bdev 00:34:40.983 11:45:24 chaining -- bdev/chaining.sh@14 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/test/nvmf/common.sh 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@7 -- # uname -s 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:809e2efd-7f71-e711-906e-0017a4403562 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@18 -- # NVME_HOSTID=809e2efd-7f71-e711-906e-0017a4403562 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/common.sh 00:34:40.983 11:45:24 chaining -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:34:40.983 11:45:24 chaining -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:34:40.983 11:45:24 chaining -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:34:40.983 11:45:24 chaining -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:40.983 11:45:24 chaining -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:40.983 11:45:24 chaining -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:40.983 11:45:24 chaining -- paths/export.sh@5 -- # export PATH 00:34:40.983 11:45:24 chaining -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@47 -- # : 0 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@51 -- # have_pci_nics=0 00:34:40.983 11:45:24 chaining -- bdev/chaining.sh@16 -- # nqn=nqn.2016-06.io.spdk:cnode0 00:34:40.983 11:45:24 chaining -- bdev/chaining.sh@17 -- # key0=(00112233445566778899001122334455 11223344556677889900112233445500) 00:34:40.983 11:45:24 chaining -- bdev/chaining.sh@18 -- # key1=(22334455667788990011223344550011 33445566778899001122334455001122) 00:34:40.983 11:45:24 chaining -- bdev/chaining.sh@19 -- # bperfsock=/var/tmp/bperf.sock 00:34:40.983 11:45:24 chaining -- bdev/chaining.sh@20 -- # declare -A stats 00:34:40.983 11:45:24 chaining -- bdev/chaining.sh@66 -- # nvmftestinit 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@448 -- # prepare_net_devs 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@410 -- # local -g is_hw=no 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@412 -- # remove_spdk_ns 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:40.983 11:45:24 chaining -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:34:40.983 11:45:24 chaining -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@414 -- # [[ phy-fallback != virt ]] 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:34:40.983 11:45:24 chaining -- nvmf/common.sh@285 -- # xtrace_disable 00:34:40.983 11:45:24 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@291 -- # pci_devs=() 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@291 -- # local -a pci_devs 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@292 -- # pci_net_devs=() 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@293 -- # pci_drivers=() 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@293 -- # local -A pci_drivers 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@295 -- # net_devs=() 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@295 -- # local -ga net_devs 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@296 -- # e810=() 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@296 -- # local -ga e810 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@297 -- # x722=() 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@297 -- # local -ga x722 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@298 -- # mlx=() 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@298 -- # local -ga mlx 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:34:49.107 11:45:31 chaining -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@327 -- # [[ '' == mlx5 ]] 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@329 -- # [[ '' == e810 ]] 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@331 -- # [[ '' == x722 ]] 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@335 -- # (( 0 == 0 )) 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@336 -- # return 1 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@416 -- # [[ no == yes ]] 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@423 -- # [[ phy-fallback == phy ]] 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@426 -- # [[ phy-fallback == phy-fallback ]] 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@427 -- # echo 'WARNING: No supported devices were found, fallback requested for tcp test' 00:34:49.108 WARNING: No supported devices were found, fallback requested for tcp test 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@431 -- # [[ tcp == tcp ]] 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@432 -- # nvmf_veth_init 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@141 -- # NVMF_INITIATOR_IP=10.0.0.1 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@142 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@143 -- # NVMF_SECOND_TARGET_IP=10.0.0.3 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@144 -- # NVMF_BRIDGE=nvmf_br 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@145 -- # NVMF_INITIATOR_INTERFACE=nvmf_init_if 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@146 -- # NVMF_INITIATOR_BRIDGE=nvmf_init_br 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@147 -- # NVMF_TARGET_NAMESPACE=nvmf_tgt_ns_spdk 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@148 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@149 -- # NVMF_TARGET_INTERFACE=nvmf_tgt_if 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@150 -- # NVMF_TARGET_INTERFACE2=nvmf_tgt_if2 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@151 -- # NVMF_TARGET_BRIDGE=nvmf_tgt_br 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@152 -- # NVMF_TARGET_BRIDGE2=nvmf_tgt_br2 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@154 -- # ip link set nvmf_init_br nomaster 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@155 -- # ip link set nvmf_tgt_br nomaster 00:34:49.108 Cannot find device "nvmf_tgt_br" 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@155 -- # true 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@156 -- # ip link set nvmf_tgt_br2 nomaster 00:34:49.108 Cannot find device "nvmf_tgt_br2" 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@156 -- # true 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@157 -- # ip link set nvmf_init_br down 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@158 -- # ip link set nvmf_tgt_br down 00:34:49.108 Cannot find device "nvmf_tgt_br" 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@158 -- # true 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@159 -- # ip link set nvmf_tgt_br2 down 00:34:49.108 Cannot find device "nvmf_tgt_br2" 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@159 -- # true 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@160 -- # ip link delete nvmf_br type bridge 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@161 -- # ip link delete nvmf_init_if 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@162 -- # ip netns exec nvmf_tgt_ns_spdk ip link delete nvmf_tgt_if 00:34:49.108 Cannot open network namespace "nvmf_tgt_ns_spdk": No such file or directory 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@162 -- # true 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@163 -- # ip netns exec nvmf_tgt_ns_spdk ip link delete nvmf_tgt_if2 00:34:49.108 Cannot open network namespace "nvmf_tgt_ns_spdk": No such file or directory 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@163 -- # true 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@166 -- # ip netns add nvmf_tgt_ns_spdk 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@169 -- # ip link add nvmf_init_if type veth peer name nvmf_init_br 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@170 -- # ip link add nvmf_tgt_if type veth peer name nvmf_tgt_br 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@171 -- # ip link add nvmf_tgt_if2 type veth peer name nvmf_tgt_br2 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@174 -- # ip link set nvmf_tgt_if netns nvmf_tgt_ns_spdk 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@175 -- # ip link set nvmf_tgt_if2 netns nvmf_tgt_ns_spdk 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@178 -- # ip addr add 10.0.0.1/24 dev nvmf_init_if 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@179 -- # ip netns exec nvmf_tgt_ns_spdk ip addr add 10.0.0.2/24 dev nvmf_tgt_if 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@180 -- # ip netns exec nvmf_tgt_ns_spdk ip addr add 10.0.0.3/24 dev nvmf_tgt_if2 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@183 -- # ip link set nvmf_init_if up 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@184 -- # ip link set nvmf_init_br up 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@185 -- # ip link set nvmf_tgt_br up 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@186 -- # ip link set nvmf_tgt_br2 up 00:34:49.108 11:45:31 chaining -- nvmf/common.sh@187 -- # ip netns exec nvmf_tgt_ns_spdk ip link set nvmf_tgt_if up 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@188 -- # ip netns exec nvmf_tgt_ns_spdk ip link set nvmf_tgt_if2 up 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@189 -- # ip netns exec nvmf_tgt_ns_spdk ip link set lo up 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@192 -- # ip link add nvmf_br type bridge 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@193 -- # ip link set nvmf_br up 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@196 -- # ip link set nvmf_init_br master nvmf_br 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@197 -- # ip link set nvmf_tgt_br master nvmf_br 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@198 -- # ip link set nvmf_tgt_br2 master nvmf_br 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@201 -- # iptables -I INPUT 1 -i nvmf_init_if -p tcp --dport 4420 -j ACCEPT 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@202 -- # iptables -A FORWARD -i nvmf_br -o nvmf_br -j ACCEPT 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@205 -- # ping -c 1 10.0.0.2 00:34:49.108 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:34:49.108 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.098 ms 00:34:49.108 00:34:49.108 --- 10.0.0.2 ping statistics --- 00:34:49.108 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:49.108 rtt min/avg/max/mdev = 0.098/0.098/0.098/0.000 ms 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@206 -- # ping -c 1 10.0.0.3 00:34:49.108 PING 10.0.0.3 (10.0.0.3) 56(84) bytes of data. 00:34:49.108 64 bytes from 10.0.0.3: icmp_seq=1 ttl=64 time=0.090 ms 00:34:49.108 00:34:49.108 --- 10.0.0.3 ping statistics --- 00:34:49.108 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:49.108 rtt min/avg/max/mdev = 0.090/0.090/0.090/0.000 ms 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@207 -- # ip netns exec nvmf_tgt_ns_spdk ping -c 1 10.0.0.1 00:34:49.108 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:34:49.108 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.042 ms 00:34:49.108 00:34:49.108 --- 10.0.0.1 ping statistics --- 00:34:49.108 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:49.108 rtt min/avg/max/mdev = 0.042/0.042/0.042/0.000 ms 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@209 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@433 -- # return 0 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:34:49.108 11:45:32 chaining -- bdev/chaining.sh@67 -- # nvmfappstart -m 0x2 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:34:49.108 11:45:32 chaining -- common/autotest_common.sh@722 -- # xtrace_disable 00:34:49.108 11:45:32 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@481 -- # nvmfpid=1074011 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@480 -- # ip netns exec nvmf_tgt_ns_spdk /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:34:49.108 11:45:32 chaining -- nvmf/common.sh@482 -- # waitforlisten 1074011 00:34:49.108 11:45:32 chaining -- common/autotest_common.sh@829 -- # '[' -z 1074011 ']' 00:34:49.108 11:45:32 chaining -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:34:49.108 11:45:32 chaining -- common/autotest_common.sh@834 -- # local max_retries=100 00:34:49.108 11:45:32 chaining -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:34:49.108 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:34:49.108 11:45:32 chaining -- common/autotest_common.sh@838 -- # xtrace_disable 00:34:49.108 11:45:32 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:49.108 [2024-07-15 11:45:32.538034] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:34:49.108 [2024-07-15 11:45:32.538103] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:34:49.108 [2024-07-15 11:45:32.664311] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:49.367 [2024-07-15 11:45:32.767009] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:34:49.367 [2024-07-15 11:45:32.767053] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:34:49.367 [2024-07-15 11:45:32.767068] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:34:49.367 [2024-07-15 11:45:32.767081] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:34:49.367 [2024-07-15 11:45:32.767092] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:34:49.367 [2024-07-15 11:45:32.767121] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:34:49.932 11:45:33 chaining -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:34:49.932 11:45:33 chaining -- common/autotest_common.sh@862 -- # return 0 00:34:49.932 11:45:33 chaining -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:34:49.932 11:45:33 chaining -- common/autotest_common.sh@728 -- # xtrace_disable 00:34:49.932 11:45:33 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:49.932 11:45:33 chaining -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:34:49.932 11:45:33 chaining -- bdev/chaining.sh@69 -- # mktemp 00:34:49.932 11:45:33 chaining -- bdev/chaining.sh@69 -- # input=/tmp/tmp.hBmwJcdLCL 00:34:49.932 11:45:33 chaining -- bdev/chaining.sh@69 -- # mktemp 00:34:49.932 11:45:33 chaining -- bdev/chaining.sh@69 -- # output=/tmp/tmp.3orPxTIUFo 00:34:49.932 11:45:33 chaining -- bdev/chaining.sh@70 -- # trap 'tgtcleanup; exit 1' SIGINT SIGTERM EXIT 00:34:49.932 11:45:33 chaining -- bdev/chaining.sh@72 -- # rpc_cmd 00:34:49.932 11:45:33 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:49.932 11:45:33 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:49.932 malloc0 00:34:49.932 true 00:34:49.932 true 00:34:49.932 [2024-07-15 11:45:33.486895] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "key0" 00:34:49.932 crypto0 00:34:49.932 [2024-07-15 11:45:33.494923] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "key1" 00:34:49.932 crypto1 00:34:49.932 [2024-07-15 11:45:33.503052] tcp.c: 672:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:34:49.932 [2024-07-15 11:45:33.519283] tcp.c: 967:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:34:50.190 11:45:33 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@85 -- # update_stats 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@51 -- # get_stat sequence_executed 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@39 -- # event=sequence_executed 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@39 -- # opcode= 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@40 -- # [[ -z '' ]] 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@41 -- # rpc_cmd accel_get_stats 00:34:50.190 11:45:33 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@41 -- # jq -r .sequence_executed 00:34:50.190 11:45:33 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:50.190 11:45:33 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@51 -- # stats["sequence_executed"]=12 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@52 -- # get_stat executed encrypt 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@39 -- # opcode=encrypt 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@40 -- # [[ -z encrypt ]] 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "encrypt").executed' 00:34:50.190 11:45:33 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:50.190 11:45:33 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:50.190 11:45:33 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@52 -- # stats["encrypt_executed"]= 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@53 -- # get_stat executed decrypt 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@39 -- # opcode=decrypt 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@40 -- # [[ -z decrypt ]] 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "decrypt").executed' 00:34:50.190 11:45:33 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:50.190 11:45:33 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:50.190 11:45:33 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@53 -- # stats["decrypt_executed"]=12 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@54 -- # get_stat executed copy 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@39 -- # opcode=copy 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@40 -- # [[ -z copy ]] 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "copy").executed' 00:34:50.190 11:45:33 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:50.190 11:45:33 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:50.190 11:45:33 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@54 -- # stats["copy_executed"]=4 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@88 -- # dd if=/dev/urandom of=/tmp/tmp.hBmwJcdLCL bs=1K count=64 00:34:50.190 64+0 records in 00:34:50.190 64+0 records out 00:34:50.190 65536 bytes (66 kB, 64 KiB) copied, 0.00107588 s, 60.9 MB/s 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@89 -- # spdk_dd --if /tmp/tmp.hBmwJcdLCL --ob Nvme0n1 --bs 65536 --count 1 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@25 -- # local config 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@31 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/gen_nvme.sh --mode=remote --json-with-subsystems --trid=tcp:10.0.0.2:4420:nqn.2016-06.io.spdk:cnode0 00:34:50.190 11:45:33 chaining -- bdev/chaining.sh@32 -- # jq '.subsystems[0].config[.subsystems[0].config | length] |= 00:34:50.190 {"method": "bdev_set_options", "params": {"bdev_auto_examine": false}}' 00:34:50.449 11:45:33 chaining -- bdev/chaining.sh@31 -- # config='{ 00:34:50.449 "subsystems": [ 00:34:50.449 { 00:34:50.449 "subsystem": "bdev", 00:34:50.449 "config": [ 00:34:50.449 { 00:34:50.449 "method": "bdev_nvme_attach_controller", 00:34:50.449 "params": { 00:34:50.449 "trtype": "tcp", 00:34:50.449 "adrfam": "IPv4", 00:34:50.449 "name": "Nvme0", 00:34:50.449 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:34:50.449 "traddr": "10.0.0.2", 00:34:50.449 "trsvcid": "4420" 00:34:50.449 } 00:34:50.449 }, 00:34:50.449 { 00:34:50.449 "method": "bdev_set_options", 00:34:50.449 "params": { 00:34:50.449 "bdev_auto_examine": false 00:34:50.449 } 00:34:50.449 } 00:34:50.449 ] 00:34:50.449 } 00:34:50.449 ] 00:34:50.449 }' 00:34:50.449 11:45:33 chaining -- bdev/chaining.sh@33 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_dd -c /dev/fd/62 --if /tmp/tmp.hBmwJcdLCL --ob Nvme0n1 --bs 65536 --count 1 00:34:50.449 11:45:33 chaining -- bdev/chaining.sh@33 -- # echo '{ 00:34:50.449 "subsystems": [ 00:34:50.449 { 00:34:50.449 "subsystem": "bdev", 00:34:50.449 "config": [ 00:34:50.449 { 00:34:50.449 "method": "bdev_nvme_attach_controller", 00:34:50.449 "params": { 00:34:50.449 "trtype": "tcp", 00:34:50.449 "adrfam": "IPv4", 00:34:50.449 "name": "Nvme0", 00:34:50.449 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:34:50.449 "traddr": "10.0.0.2", 00:34:50.449 "trsvcid": "4420" 00:34:50.449 } 00:34:50.449 }, 00:34:50.449 { 00:34:50.449 "method": "bdev_set_options", 00:34:50.449 "params": { 00:34:50.449 "bdev_auto_examine": false 00:34:50.449 } 00:34:50.449 } 00:34:50.449 ] 00:34:50.449 } 00:34:50.449 ] 00:34:50.449 }' 00:34:50.449 [2024-07-15 11:45:33.835683] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:34:50.449 [2024-07-15 11:45:33.835748] [ DPDK EAL parameters: spdk_dd --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1074234 ] 00:34:50.449 [2024-07-15 11:45:33.966768] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:50.707 [2024-07-15 11:45:34.064350] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:34:50.965  Copying: 64/64 [kB] (average 20 MBps) 00:34:50.965 00:34:50.965 11:45:34 chaining -- bdev/chaining.sh@90 -- # get_stat sequence_executed 00:34:50.965 11:45:34 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:50.965 11:45:34 chaining -- bdev/chaining.sh@39 -- # event=sequence_executed 00:34:50.965 11:45:34 chaining -- bdev/chaining.sh@39 -- # opcode= 00:34:50.965 11:45:34 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:50.965 11:45:34 chaining -- bdev/chaining.sh@40 -- # [[ -z '' ]] 00:34:50.965 11:45:34 chaining -- bdev/chaining.sh@41 -- # rpc_cmd accel_get_stats 00:34:50.965 11:45:34 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:50.965 11:45:34 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:50.965 11:45:34 chaining -- bdev/chaining.sh@41 -- # jq -r .sequence_executed 00:34:50.965 11:45:34 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:50.965 11:45:34 chaining -- bdev/chaining.sh@90 -- # (( 13 == stats[sequence_executed] + 1 )) 00:34:50.965 11:45:34 chaining -- bdev/chaining.sh@91 -- # get_stat executed encrypt 00:34:50.965 11:45:34 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:50.965 11:45:34 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:50.965 11:45:34 chaining -- bdev/chaining.sh@39 -- # opcode=encrypt 00:34:50.965 11:45:34 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:50.965 11:45:34 chaining -- bdev/chaining.sh@40 -- # [[ -z encrypt ]] 00:34:50.965 11:45:34 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:50.965 11:45:34 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "encrypt").executed' 00:34:50.965 11:45:34 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:50.965 11:45:34 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:51.223 11:45:34 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@91 -- # (( 2 == stats[encrypt_executed] + 2 )) 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@92 -- # get_stat executed decrypt 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@39 -- # opcode=decrypt 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@40 -- # [[ -z decrypt ]] 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "decrypt").executed' 00:34:51.223 11:45:34 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:51.223 11:45:34 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:51.223 11:45:34 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@92 -- # (( 12 == stats[decrypt_executed] )) 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@95 -- # get_stat executed copy 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@39 -- # opcode=copy 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@40 -- # [[ -z copy ]] 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "copy").executed' 00:34:51.223 11:45:34 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:51.223 11:45:34 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:51.223 11:45:34 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@95 -- # (( 4 == stats[copy_executed] )) 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@96 -- # update_stats 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@51 -- # get_stat sequence_executed 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@39 -- # event=sequence_executed 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@39 -- # opcode= 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@40 -- # [[ -z '' ]] 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@41 -- # jq -r .sequence_executed 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@41 -- # rpc_cmd accel_get_stats 00:34:51.223 11:45:34 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:51.223 11:45:34 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:51.223 11:45:34 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@51 -- # stats["sequence_executed"]=13 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@52 -- # get_stat executed encrypt 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@39 -- # opcode=encrypt 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@40 -- # [[ -z encrypt ]] 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "encrypt").executed' 00:34:51.223 11:45:34 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:51.223 11:45:34 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:51.223 11:45:34 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@52 -- # stats["encrypt_executed"]=2 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@53 -- # get_stat executed decrypt 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@39 -- # opcode=decrypt 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@40 -- # [[ -z decrypt ]] 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:51.223 11:45:34 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "decrypt").executed' 00:34:51.223 11:45:34 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:51.224 11:45:34 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:51.224 11:45:34 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:51.482 11:45:34 chaining -- bdev/chaining.sh@53 -- # stats["decrypt_executed"]=12 00:34:51.482 11:45:34 chaining -- bdev/chaining.sh@54 -- # get_stat executed copy 00:34:51.482 11:45:34 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:51.482 11:45:34 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:51.482 11:45:34 chaining -- bdev/chaining.sh@39 -- # opcode=copy 00:34:51.482 11:45:34 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:51.482 11:45:34 chaining -- bdev/chaining.sh@40 -- # [[ -z copy ]] 00:34:51.482 11:45:34 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:51.482 11:45:34 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "copy").executed' 00:34:51.482 11:45:34 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:51.482 11:45:34 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:51.482 11:45:34 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:51.482 11:45:34 chaining -- bdev/chaining.sh@54 -- # stats["copy_executed"]=4 00:34:51.482 11:45:34 chaining -- bdev/chaining.sh@99 -- # spdk_dd --of /tmp/tmp.3orPxTIUFo --ib Nvme0n1 --bs 65536 --count 1 00:34:51.482 11:45:34 chaining -- bdev/chaining.sh@25 -- # local config 00:34:51.482 11:45:34 chaining -- bdev/chaining.sh@31 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/gen_nvme.sh --mode=remote --json-with-subsystems --trid=tcp:10.0.0.2:4420:nqn.2016-06.io.spdk:cnode0 00:34:51.482 11:45:34 chaining -- bdev/chaining.sh@32 -- # jq '.subsystems[0].config[.subsystems[0].config | length] |= 00:34:51.482 {"method": "bdev_set_options", "params": {"bdev_auto_examine": false}}' 00:34:51.482 11:45:34 chaining -- bdev/chaining.sh@31 -- # config='{ 00:34:51.482 "subsystems": [ 00:34:51.482 { 00:34:51.482 "subsystem": "bdev", 00:34:51.482 "config": [ 00:34:51.482 { 00:34:51.482 "method": "bdev_nvme_attach_controller", 00:34:51.482 "params": { 00:34:51.482 "trtype": "tcp", 00:34:51.482 "adrfam": "IPv4", 00:34:51.482 "name": "Nvme0", 00:34:51.482 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:34:51.482 "traddr": "10.0.0.2", 00:34:51.482 "trsvcid": "4420" 00:34:51.482 } 00:34:51.482 }, 00:34:51.482 { 00:34:51.482 "method": "bdev_set_options", 00:34:51.482 "params": { 00:34:51.482 "bdev_auto_examine": false 00:34:51.482 } 00:34:51.482 } 00:34:51.482 ] 00:34:51.482 } 00:34:51.482 ] 00:34:51.482 }' 00:34:51.482 11:45:34 chaining -- bdev/chaining.sh@33 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_dd -c /dev/fd/62 --of /tmp/tmp.3orPxTIUFo --ib Nvme0n1 --bs 65536 --count 1 00:34:51.482 11:45:34 chaining -- bdev/chaining.sh@33 -- # echo '{ 00:34:51.482 "subsystems": [ 00:34:51.482 { 00:34:51.482 "subsystem": "bdev", 00:34:51.482 "config": [ 00:34:51.482 { 00:34:51.482 "method": "bdev_nvme_attach_controller", 00:34:51.482 "params": { 00:34:51.482 "trtype": "tcp", 00:34:51.482 "adrfam": "IPv4", 00:34:51.482 "name": "Nvme0", 00:34:51.482 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:34:51.482 "traddr": "10.0.0.2", 00:34:51.482 "trsvcid": "4420" 00:34:51.482 } 00:34:51.482 }, 00:34:51.482 { 00:34:51.482 "method": "bdev_set_options", 00:34:51.482 "params": { 00:34:51.482 "bdev_auto_examine": false 00:34:51.482 } 00:34:51.482 } 00:34:51.482 ] 00:34:51.482 } 00:34:51.482 ] 00:34:51.482 }' 00:34:51.482 [2024-07-15 11:45:34.978946] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:34:51.482 [2024-07-15 11:45:34.979010] [ DPDK EAL parameters: spdk_dd --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1074444 ] 00:34:51.741 [2024-07-15 11:45:35.108015] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:51.741 [2024-07-15 11:45:35.204201] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:34:52.258  Copying: 64/64 [kB] (average 15 MBps) 00:34:52.258 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@100 -- # get_stat sequence_executed 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@39 -- # event=sequence_executed 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@39 -- # opcode= 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@40 -- # [[ -z '' ]] 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@41 -- # rpc_cmd accel_get_stats 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@41 -- # jq -r .sequence_executed 00:34:52.258 11:45:35 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:52.258 11:45:35 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:52.258 11:45:35 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@100 -- # (( 14 == stats[sequence_executed] + 1 )) 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@101 -- # get_stat executed encrypt 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@39 -- # opcode=encrypt 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@40 -- # [[ -z encrypt ]] 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:52.258 11:45:35 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:52.258 11:45:35 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "encrypt").executed' 00:34:52.258 11:45:35 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@101 -- # (( 2 == stats[encrypt_executed] )) 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@102 -- # get_stat executed decrypt 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@39 -- # opcode=decrypt 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@40 -- # [[ -z decrypt ]] 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "decrypt").executed' 00:34:52.258 11:45:35 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:52.258 11:45:35 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:52.258 11:45:35 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@102 -- # (( 14 == stats[decrypt_executed] + 2 )) 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@103 -- # get_stat executed copy 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@39 -- # opcode=copy 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@40 -- # [[ -z copy ]] 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:52.258 11:45:35 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:52.258 11:45:35 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "copy").executed' 00:34:52.258 11:45:35 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@103 -- # (( 4 == stats[copy_executed] )) 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@104 -- # cmp /tmp/tmp.hBmwJcdLCL /tmp/tmp.3orPxTIUFo 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@105 -- # spdk_dd --if /dev/zero --ob Nvme0n1 --bs 65536 --count 1 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@25 -- # local config 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@31 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/gen_nvme.sh --mode=remote --json-with-subsystems --trid=tcp:10.0.0.2:4420:nqn.2016-06.io.spdk:cnode0 00:34:52.258 11:45:35 chaining -- bdev/chaining.sh@32 -- # jq '.subsystems[0].config[.subsystems[0].config | length] |= 00:34:52.258 {"method": "bdev_set_options", "params": {"bdev_auto_examine": false}}' 00:34:52.516 11:45:35 chaining -- bdev/chaining.sh@31 -- # config='{ 00:34:52.516 "subsystems": [ 00:34:52.516 { 00:34:52.516 "subsystem": "bdev", 00:34:52.516 "config": [ 00:34:52.517 { 00:34:52.517 "method": "bdev_nvme_attach_controller", 00:34:52.517 "params": { 00:34:52.517 "trtype": "tcp", 00:34:52.517 "adrfam": "IPv4", 00:34:52.517 "name": "Nvme0", 00:34:52.517 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:34:52.517 "traddr": "10.0.0.2", 00:34:52.517 "trsvcid": "4420" 00:34:52.517 } 00:34:52.517 }, 00:34:52.517 { 00:34:52.517 "method": "bdev_set_options", 00:34:52.517 "params": { 00:34:52.517 "bdev_auto_examine": false 00:34:52.517 } 00:34:52.517 } 00:34:52.517 ] 00:34:52.517 } 00:34:52.517 ] 00:34:52.517 }' 00:34:52.517 11:45:35 chaining -- bdev/chaining.sh@33 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_dd -c /dev/fd/62 --if /dev/zero --ob Nvme0n1 --bs 65536 --count 1 00:34:52.517 11:45:35 chaining -- bdev/chaining.sh@33 -- # echo '{ 00:34:52.517 "subsystems": [ 00:34:52.517 { 00:34:52.517 "subsystem": "bdev", 00:34:52.517 "config": [ 00:34:52.517 { 00:34:52.517 "method": "bdev_nvme_attach_controller", 00:34:52.517 "params": { 00:34:52.517 "trtype": "tcp", 00:34:52.517 "adrfam": "IPv4", 00:34:52.517 "name": "Nvme0", 00:34:52.517 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:34:52.517 "traddr": "10.0.0.2", 00:34:52.517 "trsvcid": "4420" 00:34:52.517 } 00:34:52.517 }, 00:34:52.517 { 00:34:52.517 "method": "bdev_set_options", 00:34:52.517 "params": { 00:34:52.517 "bdev_auto_examine": false 00:34:52.517 } 00:34:52.517 } 00:34:52.517 ] 00:34:52.517 } 00:34:52.517 ] 00:34:52.517 }' 00:34:52.517 [2024-07-15 11:45:35.954655] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:34:52.517 [2024-07-15 11:45:35.954723] [ DPDK EAL parameters: spdk_dd --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1074586 ] 00:34:52.517 [2024-07-15 11:45:36.083396] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:52.775 [2024-07-15 11:45:36.187451] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:34:53.033  Copying: 64/64 [kB] (average 10 MBps) 00:34:53.033 00:34:53.033 11:45:36 chaining -- bdev/chaining.sh@106 -- # update_stats 00:34:53.033 11:45:36 chaining -- bdev/chaining.sh@51 -- # get_stat sequence_executed 00:34:53.033 11:45:36 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:53.033 11:45:36 chaining -- bdev/chaining.sh@39 -- # event=sequence_executed 00:34:53.033 11:45:36 chaining -- bdev/chaining.sh@39 -- # opcode= 00:34:53.033 11:45:36 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:53.033 11:45:36 chaining -- bdev/chaining.sh@40 -- # [[ -z '' ]] 00:34:53.033 11:45:36 chaining -- bdev/chaining.sh@41 -- # rpc_cmd accel_get_stats 00:34:53.033 11:45:36 chaining -- bdev/chaining.sh@41 -- # jq -r .sequence_executed 00:34:53.033 11:45:36 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:53.033 11:45:36 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:53.291 11:45:36 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@51 -- # stats["sequence_executed"]=15 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@52 -- # get_stat executed encrypt 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@39 -- # opcode=encrypt 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@40 -- # [[ -z encrypt ]] 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "encrypt").executed' 00:34:53.291 11:45:36 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:53.291 11:45:36 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:53.291 11:45:36 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@52 -- # stats["encrypt_executed"]=4 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@53 -- # get_stat executed decrypt 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@39 -- # opcode=decrypt 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@40 -- # [[ -z decrypt ]] 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "decrypt").executed' 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:53.291 11:45:36 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:53.291 11:45:36 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:53.291 11:45:36 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@53 -- # stats["decrypt_executed"]=14 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@54 -- # get_stat executed copy 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@39 -- # opcode=copy 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@40 -- # [[ -z copy ]] 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:53.291 11:45:36 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "copy").executed' 00:34:53.291 11:45:36 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:53.291 11:45:36 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@54 -- # stats["copy_executed"]=4 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@109 -- # spdk_dd --if /tmp/tmp.hBmwJcdLCL --ob Nvme0n1 --bs 4096 --count 16 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@25 -- # local config 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@31 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/gen_nvme.sh --mode=remote --json-with-subsystems --trid=tcp:10.0.0.2:4420:nqn.2016-06.io.spdk:cnode0 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@32 -- # jq '.subsystems[0].config[.subsystems[0].config | length] |= 00:34:53.291 {"method": "bdev_set_options", "params": {"bdev_auto_examine": false}}' 00:34:53.291 11:45:36 chaining -- bdev/chaining.sh@31 -- # config='{ 00:34:53.291 "subsystems": [ 00:34:53.291 { 00:34:53.291 "subsystem": "bdev", 00:34:53.291 "config": [ 00:34:53.291 { 00:34:53.291 "method": "bdev_nvme_attach_controller", 00:34:53.291 "params": { 00:34:53.291 "trtype": "tcp", 00:34:53.291 "adrfam": "IPv4", 00:34:53.291 "name": "Nvme0", 00:34:53.291 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:34:53.291 "traddr": "10.0.0.2", 00:34:53.291 "trsvcid": "4420" 00:34:53.291 } 00:34:53.291 }, 00:34:53.291 { 00:34:53.291 "method": "bdev_set_options", 00:34:53.291 "params": { 00:34:53.291 "bdev_auto_examine": false 00:34:53.291 } 00:34:53.291 } 00:34:53.291 ] 00:34:53.291 } 00:34:53.291 ] 00:34:53.292 }' 00:34:53.292 11:45:36 chaining -- bdev/chaining.sh@33 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_dd -c /dev/fd/62 --if /tmp/tmp.hBmwJcdLCL --ob Nvme0n1 --bs 4096 --count 16 00:34:53.292 11:45:36 chaining -- bdev/chaining.sh@33 -- # echo '{ 00:34:53.292 "subsystems": [ 00:34:53.292 { 00:34:53.292 "subsystem": "bdev", 00:34:53.292 "config": [ 00:34:53.292 { 00:34:53.292 "method": "bdev_nvme_attach_controller", 00:34:53.292 "params": { 00:34:53.292 "trtype": "tcp", 00:34:53.292 "adrfam": "IPv4", 00:34:53.292 "name": "Nvme0", 00:34:53.292 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:34:53.292 "traddr": "10.0.0.2", 00:34:53.292 "trsvcid": "4420" 00:34:53.292 } 00:34:53.292 }, 00:34:53.292 { 00:34:53.292 "method": "bdev_set_options", 00:34:53.292 "params": { 00:34:53.292 "bdev_auto_examine": false 00:34:53.292 } 00:34:53.292 } 00:34:53.292 ] 00:34:53.292 } 00:34:53.292 ] 00:34:53.292 }' 00:34:53.550 [2024-07-15 11:45:36.905483] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:34:53.550 [2024-07-15 11:45:36.905545] [ DPDK EAL parameters: spdk_dd --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1074677 ] 00:34:53.550 [2024-07-15 11:45:37.034964] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:53.550 [2024-07-15 11:45:37.131481] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:34:54.066  Copying: 64/64 [kB] (average 9142 kBps) 00:34:54.066 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@110 -- # get_stat sequence_executed 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@39 -- # event=sequence_executed 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@39 -- # opcode= 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@40 -- # [[ -z '' ]] 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@41 -- # rpc_cmd accel_get_stats 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@41 -- # jq -r .sequence_executed 00:34:54.066 11:45:37 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:54.066 11:45:37 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:54.066 11:45:37 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@110 -- # (( 31 == stats[sequence_executed] + 16 )) 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@111 -- # get_stat executed encrypt 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@39 -- # opcode=encrypt 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@40 -- # [[ -z encrypt ]] 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "encrypt").executed' 00:34:54.066 11:45:37 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:54.066 11:45:37 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:54.066 11:45:37 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@111 -- # (( 36 == stats[encrypt_executed] + 32 )) 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@112 -- # get_stat executed decrypt 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@39 -- # opcode=decrypt 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@40 -- # [[ -z decrypt ]] 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:54.066 11:45:37 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:54.066 11:45:37 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:54.066 11:45:37 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "decrypt").executed' 00:34:54.325 11:45:37 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@112 -- # (( 14 == stats[decrypt_executed] )) 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@113 -- # get_stat executed copy 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@39 -- # opcode=copy 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@40 -- # [[ -z copy ]] 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "copy").executed' 00:34:54.325 11:45:37 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:54.325 11:45:37 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:54.325 11:45:37 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@113 -- # (( 4 == stats[copy_executed] )) 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@114 -- # update_stats 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@51 -- # get_stat sequence_executed 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@39 -- # event=sequence_executed 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@39 -- # opcode= 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@40 -- # [[ -z '' ]] 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@41 -- # jq -r .sequence_executed 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@41 -- # rpc_cmd accel_get_stats 00:34:54.325 11:45:37 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:54.325 11:45:37 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:54.325 11:45:37 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@51 -- # stats["sequence_executed"]=31 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@52 -- # get_stat executed encrypt 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@39 -- # opcode=encrypt 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@40 -- # [[ -z encrypt ]] 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:54.325 11:45:37 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:54.325 11:45:37 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "encrypt").executed' 00:34:54.325 11:45:37 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@52 -- # stats["encrypt_executed"]=36 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@53 -- # get_stat executed decrypt 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@39 -- # opcode=decrypt 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@40 -- # [[ -z decrypt ]] 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:54.325 11:45:37 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:54.325 11:45:37 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "decrypt").executed' 00:34:54.325 11:45:37 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@53 -- # stats["decrypt_executed"]=14 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@54 -- # get_stat executed copy 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@39 -- # opcode=copy 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@40 -- # [[ -z copy ]] 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:54.325 11:45:37 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "copy").executed' 00:34:54.325 11:45:37 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:54.325 11:45:37 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:54.325 11:45:37 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:54.583 11:45:37 chaining -- bdev/chaining.sh@54 -- # stats["copy_executed"]=4 00:34:54.583 11:45:37 chaining -- bdev/chaining.sh@117 -- # : 00:34:54.583 11:45:37 chaining -- bdev/chaining.sh@118 -- # spdk_dd --of /tmp/tmp.3orPxTIUFo --ib Nvme0n1 --bs 4096 --count 16 00:34:54.583 11:45:37 chaining -- bdev/chaining.sh@25 -- # local config 00:34:54.583 11:45:37 chaining -- bdev/chaining.sh@32 -- # jq '.subsystems[0].config[.subsystems[0].config | length] |= 00:34:54.583 {"method": "bdev_set_options", "params": {"bdev_auto_examine": false}}' 00:34:54.583 11:45:37 chaining -- bdev/chaining.sh@31 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/gen_nvme.sh --mode=remote --json-with-subsystems --trid=tcp:10.0.0.2:4420:nqn.2016-06.io.spdk:cnode0 00:34:54.583 11:45:37 chaining -- bdev/chaining.sh@31 -- # config='{ 00:34:54.583 "subsystems": [ 00:34:54.583 { 00:34:54.583 "subsystem": "bdev", 00:34:54.583 "config": [ 00:34:54.583 { 00:34:54.583 "method": "bdev_nvme_attach_controller", 00:34:54.583 "params": { 00:34:54.583 "trtype": "tcp", 00:34:54.583 "adrfam": "IPv4", 00:34:54.583 "name": "Nvme0", 00:34:54.583 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:34:54.584 "traddr": "10.0.0.2", 00:34:54.584 "trsvcid": "4420" 00:34:54.584 } 00:34:54.584 }, 00:34:54.584 { 00:34:54.584 "method": "bdev_set_options", 00:34:54.584 "params": { 00:34:54.584 "bdev_auto_examine": false 00:34:54.584 } 00:34:54.584 } 00:34:54.584 ] 00:34:54.584 } 00:34:54.584 ] 00:34:54.584 }' 00:34:54.584 11:45:37 chaining -- bdev/chaining.sh@33 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/spdk_dd -c /dev/fd/62 --of /tmp/tmp.3orPxTIUFo --ib Nvme0n1 --bs 4096 --count 16 00:34:54.584 11:45:37 chaining -- bdev/chaining.sh@33 -- # echo '{ 00:34:54.584 "subsystems": [ 00:34:54.584 { 00:34:54.584 "subsystem": "bdev", 00:34:54.584 "config": [ 00:34:54.584 { 00:34:54.584 "method": "bdev_nvme_attach_controller", 00:34:54.584 "params": { 00:34:54.584 "trtype": "tcp", 00:34:54.584 "adrfam": "IPv4", 00:34:54.584 "name": "Nvme0", 00:34:54.584 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:34:54.584 "traddr": "10.0.0.2", 00:34:54.584 "trsvcid": "4420" 00:34:54.584 } 00:34:54.584 }, 00:34:54.584 { 00:34:54.584 "method": "bdev_set_options", 00:34:54.584 "params": { 00:34:54.584 "bdev_auto_examine": false 00:34:54.584 } 00:34:54.584 } 00:34:54.584 ] 00:34:54.584 } 00:34:54.584 ] 00:34:54.584 }' 00:34:54.584 [2024-07-15 11:45:38.017817] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:34:54.584 [2024-07-15 11:45:38.017885] [ DPDK EAL parameters: spdk_dd --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1074884 ] 00:34:54.584 [2024-07-15 11:45:38.146840] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:54.842 [2024-07-15 11:45:38.246761] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:34:55.358  Copying: 64/64 [kB] (average 1391 kBps) 00:34:55.358 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@119 -- # get_stat sequence_executed 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@39 -- # event=sequence_executed 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@39 -- # opcode= 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@40 -- # [[ -z '' ]] 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@41 -- # rpc_cmd accel_get_stats 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@41 -- # jq -r .sequence_executed 00:34:55.358 11:45:38 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:55.358 11:45:38 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:55.358 11:45:38 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@119 -- # (( 47 == stats[sequence_executed] + 16 )) 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@120 -- # get_stat executed encrypt 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@39 -- # opcode=encrypt 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@40 -- # [[ -z encrypt ]] 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "encrypt").executed' 00:34:55.358 11:45:38 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:55.358 11:45:38 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:55.358 11:45:38 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@120 -- # (( 36 == stats[encrypt_executed] )) 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@121 -- # get_stat executed decrypt 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@39 -- # opcode=decrypt 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@40 -- # [[ -z decrypt ]] 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:55.358 11:45:38 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:55.358 11:45:38 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "decrypt").executed' 00:34:55.358 11:45:38 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@121 -- # (( 46 == stats[decrypt_executed] + 32 )) 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@122 -- # get_stat executed copy 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@39 -- # event=executed 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@39 -- # opcode=copy 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_cmd 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@40 -- # [[ -z copy ]] 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@43 -- # rpc_cmd accel_get_stats 00:34:55.358 11:45:38 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "copy").executed' 00:34:55.358 11:45:38 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:55.359 11:45:38 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:55.359 11:45:38 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:55.359 11:45:38 chaining -- bdev/chaining.sh@122 -- # (( 4 == stats[copy_executed] )) 00:34:55.359 11:45:38 chaining -- bdev/chaining.sh@123 -- # cmp /tmp/tmp.hBmwJcdLCL /tmp/tmp.3orPxTIUFo 00:34:55.359 11:45:38 chaining -- bdev/chaining.sh@125 -- # trap - SIGINT SIGTERM EXIT 00:34:55.359 11:45:38 chaining -- bdev/chaining.sh@126 -- # tgtcleanup 00:34:55.359 11:45:38 chaining -- bdev/chaining.sh@58 -- # rm -f /tmp/tmp.hBmwJcdLCL /tmp/tmp.3orPxTIUFo 00:34:55.359 11:45:38 chaining -- bdev/chaining.sh@59 -- # nvmftestfini 00:34:55.359 11:45:38 chaining -- nvmf/common.sh@488 -- # nvmfcleanup 00:34:55.359 11:45:38 chaining -- nvmf/common.sh@117 -- # sync 00:34:55.359 11:45:38 chaining -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:34:55.359 11:45:38 chaining -- nvmf/common.sh@120 -- # set +e 00:34:55.359 11:45:38 chaining -- nvmf/common.sh@121 -- # for i in {1..20} 00:34:55.359 11:45:38 chaining -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:34:55.359 rmmod nvme_tcp 00:34:55.618 rmmod nvme_fabrics 00:34:55.618 rmmod nvme_keyring 00:34:55.618 11:45:38 chaining -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:34:55.618 11:45:39 chaining -- nvmf/common.sh@124 -- # set -e 00:34:55.618 11:45:39 chaining -- nvmf/common.sh@125 -- # return 0 00:34:55.618 11:45:39 chaining -- nvmf/common.sh@489 -- # '[' -n 1074011 ']' 00:34:55.618 11:45:39 chaining -- nvmf/common.sh@490 -- # killprocess 1074011 00:34:55.618 11:45:39 chaining -- common/autotest_common.sh@948 -- # '[' -z 1074011 ']' 00:34:55.618 11:45:39 chaining -- common/autotest_common.sh@952 -- # kill -0 1074011 00:34:55.618 11:45:39 chaining -- common/autotest_common.sh@953 -- # uname 00:34:55.618 11:45:39 chaining -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:34:55.618 11:45:39 chaining -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1074011 00:34:55.618 11:45:39 chaining -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:34:55.618 11:45:39 chaining -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:34:55.618 11:45:39 chaining -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1074011' 00:34:55.618 killing process with pid 1074011 00:34:55.618 11:45:39 chaining -- common/autotest_common.sh@967 -- # kill 1074011 00:34:55.618 11:45:39 chaining -- common/autotest_common.sh@972 -- # wait 1074011 00:34:55.875 11:45:39 chaining -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:34:55.875 11:45:39 chaining -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:34:55.875 11:45:39 chaining -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:34:55.875 11:45:39 chaining -- nvmf/common.sh@274 -- # [[ nvmf_tgt_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:34:55.875 11:45:39 chaining -- nvmf/common.sh@278 -- # remove_spdk_ns 00:34:55.875 11:45:39 chaining -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:55.875 11:45:39 chaining -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:34:55.875 11:45:39 chaining -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:55.875 11:45:39 chaining -- nvmf/common.sh@279 -- # ip -4 addr flush nvmf_init_if 00:34:55.875 11:45:39 chaining -- bdev/chaining.sh@129 -- # trap 'bperfcleanup; exit 1' SIGINT SIGTERM EXIT 00:34:55.875 11:45:39 chaining -- bdev/chaining.sh@132 -- # bperfpid=1075095 00:34:55.875 11:45:39 chaining -- bdev/chaining.sh@131 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -t 5 -w verify -o 4096 -q 256 --wait-for-rpc -z 00:34:55.875 11:45:39 chaining -- bdev/chaining.sh@134 -- # waitforlisten 1075095 00:34:55.875 11:45:39 chaining -- common/autotest_common.sh@829 -- # '[' -z 1075095 ']' 00:34:55.875 11:45:39 chaining -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:34:55.875 11:45:39 chaining -- common/autotest_common.sh@834 -- # local max_retries=100 00:34:55.875 11:45:39 chaining -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:34:55.875 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:34:55.875 11:45:39 chaining -- common/autotest_common.sh@838 -- # xtrace_disable 00:34:55.875 11:45:39 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:55.875 [2024-07-15 11:45:39.387377] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:34:55.876 [2024-07-15 11:45:39.387442] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1075095 ] 00:34:56.133 [2024-07-15 11:45:39.518436] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:56.133 [2024-07-15 11:45:39.620314] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:34:56.700 11:45:40 chaining -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:34:56.700 11:45:40 chaining -- common/autotest_common.sh@862 -- # return 0 00:34:56.700 11:45:40 chaining -- bdev/chaining.sh@135 -- # rpc_cmd 00:34:56.700 11:45:40 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:34:56.700 11:45:40 chaining -- common/autotest_common.sh@10 -- # set +x 00:34:56.959 malloc0 00:34:56.959 true 00:34:56.959 true 00:34:56.959 [2024-07-15 11:45:40.388986] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "key0" 00:34:56.959 crypto0 00:34:56.959 [2024-07-15 11:45:40.397011] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "key1" 00:34:56.959 crypto1 00:34:56.959 11:45:40 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:34:56.959 11:45:40 chaining -- bdev/chaining.sh@145 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:34:56.959 Running I/O for 5 seconds... 00:35:02.266 00:35:02.266 Latency(us) 00:35:02.266 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:35:02.266 Job: crypto1 (Core Mask 0x1, workload: verify, depth: 256, IO size: 4096) 00:35:02.266 Verification LBA range: start 0x0 length 0x2000 00:35:02.266 crypto1 : 5.01 11457.97 44.76 0.00 0.00 22274.54 1695.39 14303.94 00:35:02.266 =================================================================================================================== 00:35:02.266 Total : 11457.97 44.76 0.00 0.00 22274.54 1695.39 14303.94 00:35:02.266 0 00:35:02.266 11:45:45 chaining -- bdev/chaining.sh@146 -- # killprocess 1075095 00:35:02.266 11:45:45 chaining -- common/autotest_common.sh@948 -- # '[' -z 1075095 ']' 00:35:02.266 11:45:45 chaining -- common/autotest_common.sh@952 -- # kill -0 1075095 00:35:02.266 11:45:45 chaining -- common/autotest_common.sh@953 -- # uname 00:35:02.267 11:45:45 chaining -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:35:02.267 11:45:45 chaining -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1075095 00:35:02.267 11:45:45 chaining -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:35:02.267 11:45:45 chaining -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:35:02.267 11:45:45 chaining -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1075095' 00:35:02.267 killing process with pid 1075095 00:35:02.267 11:45:45 chaining -- common/autotest_common.sh@967 -- # kill 1075095 00:35:02.267 Received shutdown signal, test time was about 5.000000 seconds 00:35:02.267 00:35:02.267 Latency(us) 00:35:02.267 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:35:02.267 =================================================================================================================== 00:35:02.267 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:35:02.267 11:45:45 chaining -- common/autotest_common.sh@972 -- # wait 1075095 00:35:02.267 11:45:45 chaining -- bdev/chaining.sh@152 -- # bperfpid=1075970 00:35:02.267 11:45:45 chaining -- bdev/chaining.sh@151 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -t 5 -w verify -o 4096 -q 256 --wait-for-rpc -z 00:35:02.267 11:45:45 chaining -- bdev/chaining.sh@154 -- # waitforlisten 1075970 00:35:02.267 11:45:45 chaining -- common/autotest_common.sh@829 -- # '[' -z 1075970 ']' 00:35:02.267 11:45:45 chaining -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:35:02.267 11:45:45 chaining -- common/autotest_common.sh@834 -- # local max_retries=100 00:35:02.267 11:45:45 chaining -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:35:02.267 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:35:02.267 11:45:45 chaining -- common/autotest_common.sh@838 -- # xtrace_disable 00:35:02.267 11:45:45 chaining -- common/autotest_common.sh@10 -- # set +x 00:35:02.540 [2024-07-15 11:45:45.881076] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:35:02.540 [2024-07-15 11:45:45.881146] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1075970 ] 00:35:02.540 [2024-07-15 11:45:46.010836] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:35:02.540 [2024-07-15 11:45:46.117960] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:35:03.476 11:45:46 chaining -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:35:03.476 11:45:46 chaining -- common/autotest_common.sh@862 -- # return 0 00:35:03.476 11:45:46 chaining -- bdev/chaining.sh@155 -- # rpc_cmd 00:35:03.476 11:45:46 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:35:03.476 11:45:46 chaining -- common/autotest_common.sh@10 -- # set +x 00:35:03.476 malloc0 00:35:03.476 true 00:35:03.476 true 00:35:03.476 [2024-07-15 11:45:46.950524] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:35:03.476 [2024-07-15 11:45:46.950570] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:03.476 [2024-07-15 11:45:46.950593] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x1c8f730 00:35:03.476 [2024-07-15 11:45:46.950608] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:03.476 [2024-07-15 11:45:46.951685] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:03.476 [2024-07-15 11:45:46.951709] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:35:03.476 pt0 00:35:03.476 [2024-07-15 11:45:46.958555] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "key0" 00:35:03.476 crypto0 00:35:03.476 [2024-07-15 11:45:46.966575] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "key1" 00:35:03.476 crypto1 00:35:03.476 11:45:46 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:35:03.476 11:45:46 chaining -- bdev/chaining.sh@166 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:35:03.476 Running I/O for 5 seconds... 00:35:08.747 00:35:08.747 Latency(us) 00:35:08.747 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:35:08.747 Job: crypto1 (Core Mask 0x1, workload: verify, depth: 256, IO size: 4096) 00:35:08.747 Verification LBA range: start 0x0 length 0x2000 00:35:08.747 crypto1 : 5.01 9087.05 35.50 0.00 0.00 28095.87 6553.60 17324.30 00:35:08.747 =================================================================================================================== 00:35:08.747 Total : 9087.05 35.50 0.00 0.00 28095.87 6553.60 17324.30 00:35:08.747 0 00:35:08.747 11:45:52 chaining -- bdev/chaining.sh@167 -- # killprocess 1075970 00:35:08.747 11:45:52 chaining -- common/autotest_common.sh@948 -- # '[' -z 1075970 ']' 00:35:08.747 11:45:52 chaining -- common/autotest_common.sh@952 -- # kill -0 1075970 00:35:08.747 11:45:52 chaining -- common/autotest_common.sh@953 -- # uname 00:35:08.747 11:45:52 chaining -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:35:08.747 11:45:52 chaining -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1075970 00:35:08.747 11:45:52 chaining -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:35:08.747 11:45:52 chaining -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:35:08.747 11:45:52 chaining -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1075970' 00:35:08.747 killing process with pid 1075970 00:35:08.747 11:45:52 chaining -- common/autotest_common.sh@967 -- # kill 1075970 00:35:08.747 Received shutdown signal, test time was about 5.000000 seconds 00:35:08.747 00:35:08.747 Latency(us) 00:35:08.747 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:35:08.747 =================================================================================================================== 00:35:08.747 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:35:08.747 11:45:52 chaining -- common/autotest_common.sh@972 -- # wait 1075970 00:35:09.006 11:45:52 chaining -- bdev/chaining.sh@169 -- # trap - SIGINT SIGTERM EXIT 00:35:09.006 11:45:52 chaining -- bdev/chaining.sh@170 -- # killprocess 1075970 00:35:09.006 11:45:52 chaining -- common/autotest_common.sh@948 -- # '[' -z 1075970 ']' 00:35:09.006 11:45:52 chaining -- common/autotest_common.sh@952 -- # kill -0 1075970 00:35:09.006 /var/jenkins/workspace/crypto-phy-autotest/spdk/test/common/autotest_common.sh: line 952: kill: (1075970) - No such process 00:35:09.006 11:45:52 chaining -- common/autotest_common.sh@975 -- # echo 'Process with pid 1075970 is not found' 00:35:09.006 Process with pid 1075970 is not found 00:35:09.006 11:45:52 chaining -- bdev/chaining.sh@171 -- # wait 1075970 00:35:09.006 11:45:52 chaining -- bdev/chaining.sh@175 -- # nvmftestinit 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@448 -- # prepare_net_devs 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@410 -- # local -g is_hw=no 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@412 -- # remove_spdk_ns 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:35:09.006 11:45:52 chaining -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:35:09.006 11:45:52 chaining -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@414 -- # [[ phy-fallback != virt ]] 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@285 -- # xtrace_disable 00:35:09.006 11:45:52 chaining -- common/autotest_common.sh@10 -- # set +x 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@291 -- # pci_devs=() 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@291 -- # local -a pci_devs 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@292 -- # pci_net_devs=() 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@293 -- # pci_drivers=() 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@293 -- # local -A pci_drivers 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@295 -- # net_devs=() 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@295 -- # local -ga net_devs 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@296 -- # e810=() 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@296 -- # local -ga e810 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@297 -- # x722=() 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@297 -- # local -ga x722 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@298 -- # mlx=() 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@298 -- # local -ga mlx 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@327 -- # [[ '' == mlx5 ]] 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@329 -- # [[ '' == e810 ]] 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@331 -- # [[ '' == x722 ]] 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@335 -- # (( 0 == 0 )) 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@336 -- # return 1 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@416 -- # [[ no == yes ]] 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@423 -- # [[ phy-fallback == phy ]] 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@426 -- # [[ phy-fallback == phy-fallback ]] 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@427 -- # echo 'WARNING: No supported devices were found, fallback requested for tcp test' 00:35:09.006 WARNING: No supported devices were found, fallback requested for tcp test 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@431 -- # [[ tcp == tcp ]] 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@432 -- # nvmf_veth_init 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@141 -- # NVMF_INITIATOR_IP=10.0.0.1 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@142 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@143 -- # NVMF_SECOND_TARGET_IP=10.0.0.3 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@144 -- # NVMF_BRIDGE=nvmf_br 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@145 -- # NVMF_INITIATOR_INTERFACE=nvmf_init_if 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@146 -- # NVMF_INITIATOR_BRIDGE=nvmf_init_br 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@147 -- # NVMF_TARGET_NAMESPACE=nvmf_tgt_ns_spdk 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@148 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@149 -- # NVMF_TARGET_INTERFACE=nvmf_tgt_if 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@150 -- # NVMF_TARGET_INTERFACE2=nvmf_tgt_if2 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@151 -- # NVMF_TARGET_BRIDGE=nvmf_tgt_br 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@152 -- # NVMF_TARGET_BRIDGE2=nvmf_tgt_br2 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@154 -- # ip link set nvmf_init_br nomaster 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@155 -- # ip link set nvmf_tgt_br nomaster 00:35:09.006 Cannot find device "nvmf_tgt_br" 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@155 -- # true 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@156 -- # ip link set nvmf_tgt_br2 nomaster 00:35:09.006 Cannot find device "nvmf_tgt_br2" 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@156 -- # true 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@157 -- # ip link set nvmf_init_br down 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@158 -- # ip link set nvmf_tgt_br down 00:35:09.006 Cannot find device "nvmf_tgt_br" 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@158 -- # true 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@159 -- # ip link set nvmf_tgt_br2 down 00:35:09.006 Cannot find device "nvmf_tgt_br2" 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@159 -- # true 00:35:09.006 11:45:52 chaining -- nvmf/common.sh@160 -- # ip link delete nvmf_br type bridge 00:35:09.264 11:45:52 chaining -- nvmf/common.sh@161 -- # ip link delete nvmf_init_if 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@162 -- # ip netns exec nvmf_tgt_ns_spdk ip link delete nvmf_tgt_if 00:35:09.265 Cannot open network namespace "nvmf_tgt_ns_spdk": No such file or directory 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@162 -- # true 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@163 -- # ip netns exec nvmf_tgt_ns_spdk ip link delete nvmf_tgt_if2 00:35:09.265 Cannot open network namespace "nvmf_tgt_ns_spdk": No such file or directory 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@163 -- # true 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@166 -- # ip netns add nvmf_tgt_ns_spdk 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@169 -- # ip link add nvmf_init_if type veth peer name nvmf_init_br 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@170 -- # ip link add nvmf_tgt_if type veth peer name nvmf_tgt_br 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@171 -- # ip link add nvmf_tgt_if2 type veth peer name nvmf_tgt_br2 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@174 -- # ip link set nvmf_tgt_if netns nvmf_tgt_ns_spdk 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@175 -- # ip link set nvmf_tgt_if2 netns nvmf_tgt_ns_spdk 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@178 -- # ip addr add 10.0.0.1/24 dev nvmf_init_if 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@179 -- # ip netns exec nvmf_tgt_ns_spdk ip addr add 10.0.0.2/24 dev nvmf_tgt_if 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@180 -- # ip netns exec nvmf_tgt_ns_spdk ip addr add 10.0.0.3/24 dev nvmf_tgt_if2 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@183 -- # ip link set nvmf_init_if up 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@184 -- # ip link set nvmf_init_br up 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@185 -- # ip link set nvmf_tgt_br up 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@186 -- # ip link set nvmf_tgt_br2 up 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@187 -- # ip netns exec nvmf_tgt_ns_spdk ip link set nvmf_tgt_if up 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@188 -- # ip netns exec nvmf_tgt_ns_spdk ip link set nvmf_tgt_if2 up 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@189 -- # ip netns exec nvmf_tgt_ns_spdk ip link set lo up 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@192 -- # ip link add nvmf_br type bridge 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@193 -- # ip link set nvmf_br up 00:35:09.265 11:45:52 chaining -- nvmf/common.sh@196 -- # ip link set nvmf_init_br master nvmf_br 00:35:09.523 11:45:52 chaining -- nvmf/common.sh@197 -- # ip link set nvmf_tgt_br master nvmf_br 00:35:09.523 11:45:53 chaining -- nvmf/common.sh@198 -- # ip link set nvmf_tgt_br2 master nvmf_br 00:35:09.782 11:45:53 chaining -- nvmf/common.sh@201 -- # iptables -I INPUT 1 -i nvmf_init_if -p tcp --dport 4420 -j ACCEPT 00:35:09.782 11:45:53 chaining -- nvmf/common.sh@202 -- # iptables -A FORWARD -i nvmf_br -o nvmf_br -j ACCEPT 00:35:09.782 11:45:53 chaining -- nvmf/common.sh@205 -- # ping -c 1 10.0.0.2 00:35:09.782 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:35:09.782 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.101 ms 00:35:09.782 00:35:09.782 --- 10.0.0.2 ping statistics --- 00:35:09.782 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:35:09.782 rtt min/avg/max/mdev = 0.101/0.101/0.101/0.000 ms 00:35:09.783 11:45:53 chaining -- nvmf/common.sh@206 -- # ping -c 1 10.0.0.3 00:35:09.783 PING 10.0.0.3 (10.0.0.3) 56(84) bytes of data. 00:35:09.783 64 bytes from 10.0.0.3: icmp_seq=1 ttl=64 time=0.072 ms 00:35:09.783 00:35:09.783 --- 10.0.0.3 ping statistics --- 00:35:09.783 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:35:09.783 rtt min/avg/max/mdev = 0.072/0.072/0.072/0.000 ms 00:35:09.783 11:45:53 chaining -- nvmf/common.sh@207 -- # ip netns exec nvmf_tgt_ns_spdk ping -c 1 10.0.0.1 00:35:09.783 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:35:09.783 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.038 ms 00:35:09.783 00:35:09.783 --- 10.0.0.1 ping statistics --- 00:35:09.783 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:35:09.783 rtt min/avg/max/mdev = 0.038/0.038/0.038/0.000 ms 00:35:09.783 11:45:53 chaining -- nvmf/common.sh@209 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:35:09.783 11:45:53 chaining -- nvmf/common.sh@433 -- # return 0 00:35:09.783 11:45:53 chaining -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:35:09.783 11:45:53 chaining -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:35:09.783 11:45:53 chaining -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:35:09.783 11:45:53 chaining -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:35:09.783 11:45:53 chaining -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:35:09.783 11:45:53 chaining -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:35:09.783 11:45:53 chaining -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:35:09.783 11:45:53 chaining -- bdev/chaining.sh@176 -- # nvmfappstart -m 0x2 00:35:09.783 11:45:53 chaining -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:35:09.783 11:45:53 chaining -- common/autotest_common.sh@722 -- # xtrace_disable 00:35:09.783 11:45:53 chaining -- common/autotest_common.sh@10 -- # set +x 00:35:09.783 11:45:53 chaining -- nvmf/common.sh@481 -- # nvmfpid=1077125 00:35:09.783 11:45:53 chaining -- nvmf/common.sh@480 -- # ip netns exec nvmf_tgt_ns_spdk ip netns exec nvmf_tgt_ns_spdk /var/jenkins/workspace/crypto-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:35:09.783 11:45:53 chaining -- nvmf/common.sh@482 -- # waitforlisten 1077125 00:35:09.783 11:45:53 chaining -- common/autotest_common.sh@829 -- # '[' -z 1077125 ']' 00:35:09.783 11:45:53 chaining -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:35:09.783 11:45:53 chaining -- common/autotest_common.sh@834 -- # local max_retries=100 00:35:09.783 11:45:53 chaining -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:35:09.783 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:35:09.783 11:45:53 chaining -- common/autotest_common.sh@838 -- # xtrace_disable 00:35:09.783 11:45:53 chaining -- common/autotest_common.sh@10 -- # set +x 00:35:09.783 [2024-07-15 11:45:53.301448] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:35:09.783 [2024-07-15 11:45:53.301520] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:35:10.042 [2024-07-15 11:45:53.426762] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:35:10.042 [2024-07-15 11:45:53.527272] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:35:10.042 [2024-07-15 11:45:53.527316] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:35:10.042 [2024-07-15 11:45:53.527331] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:35:10.042 [2024-07-15 11:45:53.527344] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:35:10.042 [2024-07-15 11:45:53.527354] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:35:10.042 [2024-07-15 11:45:53.527391] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:35:10.976 11:45:54 chaining -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:35:10.977 11:45:54 chaining -- common/autotest_common.sh@862 -- # return 0 00:35:10.977 11:45:54 chaining -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:35:10.977 11:45:54 chaining -- common/autotest_common.sh@728 -- # xtrace_disable 00:35:10.977 11:45:54 chaining -- common/autotest_common.sh@10 -- # set +x 00:35:10.977 11:45:54 chaining -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:35:10.977 11:45:54 chaining -- bdev/chaining.sh@178 -- # rpc_cmd 00:35:10.977 11:45:54 chaining -- common/autotest_common.sh@559 -- # xtrace_disable 00:35:10.977 11:45:54 chaining -- common/autotest_common.sh@10 -- # set +x 00:35:10.977 malloc0 00:35:10.977 [2024-07-15 11:45:54.291242] tcp.c: 672:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:35:10.977 [2024-07-15 11:45:54.307464] tcp.c: 967:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:35:10.977 11:45:54 chaining -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:35:10.977 11:45:54 chaining -- bdev/chaining.sh@186 -- # trap 'bperfcleanup || :; nvmftestfini || :; exit 1' SIGINT SIGTERM EXIT 00:35:10.977 11:45:54 chaining -- bdev/chaining.sh@189 -- # bperfpid=1077313 00:35:10.977 11:45:54 chaining -- bdev/chaining.sh@187 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bperf.sock -t 5 -w verify -o 4096 -q 256 --wait-for-rpc -z 00:35:10.977 11:45:54 chaining -- bdev/chaining.sh@191 -- # waitforlisten 1077313 /var/tmp/bperf.sock 00:35:10.977 11:45:54 chaining -- common/autotest_common.sh@829 -- # '[' -z 1077313 ']' 00:35:10.977 11:45:54 chaining -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bperf.sock 00:35:10.977 11:45:54 chaining -- common/autotest_common.sh@834 -- # local max_retries=100 00:35:10.977 11:45:54 chaining -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:35:10.977 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:35:10.977 11:45:54 chaining -- common/autotest_common.sh@838 -- # xtrace_disable 00:35:10.977 11:45:54 chaining -- common/autotest_common.sh@10 -- # set +x 00:35:10.977 [2024-07-15 11:45:54.382545] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:35:10.977 [2024-07-15 11:45:54.382606] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1077313 ] 00:35:10.977 [2024-07-15 11:45:54.511603] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:35:11.235 [2024-07-15 11:45:54.610031] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:35:11.803 11:45:55 chaining -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:35:11.803 11:45:55 chaining -- common/autotest_common.sh@862 -- # return 0 00:35:11.803 11:45:55 chaining -- bdev/chaining.sh@192 -- # rpc_bperf 00:35:11.803 11:45:55 chaining -- bdev/chaining.sh@22 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock 00:35:12.370 [2024-07-15 11:45:55.722749] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "key0" 00:35:12.370 nvme0n1 00:35:12.370 true 00:35:12.370 crypto0 00:35:12.370 11:45:55 chaining -- bdev/chaining.sh@201 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:35:12.370 Running I/O for 5 seconds... 00:35:17.638 00:35:17.638 Latency(us) 00:35:17.638 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:35:17.638 Job: crypto0 (Core Mask 0x1, workload: verify, depth: 256, IO size: 4096) 00:35:17.638 Verification LBA range: start 0x0 length 0x2000 00:35:17.638 crypto0 : 5.02 8299.34 32.42 0.00 0.00 30742.56 3960.65 24960.67 00:35:17.638 =================================================================================================================== 00:35:17.638 Total : 8299.34 32.42 0.00 0.00 30742.56 3960.65 24960.67 00:35:17.638 0 00:35:17.638 11:46:00 chaining -- bdev/chaining.sh@205 -- # get_stat_bperf sequence_executed 00:35:17.638 11:46:00 chaining -- bdev/chaining.sh@48 -- # get_stat sequence_executed '' rpc_bperf 00:35:17.638 11:46:00 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:35:17.638 11:46:00 chaining -- bdev/chaining.sh@39 -- # event=sequence_executed 00:35:17.638 11:46:00 chaining -- bdev/chaining.sh@39 -- # opcode= 00:35:17.638 11:46:00 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_bperf 00:35:17.638 11:46:00 chaining -- bdev/chaining.sh@40 -- # [[ -z '' ]] 00:35:17.638 11:46:00 chaining -- bdev/chaining.sh@41 -- # rpc_bperf accel_get_stats 00:35:17.638 11:46:00 chaining -- bdev/chaining.sh@22 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:35:17.638 11:46:00 chaining -- bdev/chaining.sh@41 -- # jq -r .sequence_executed 00:35:17.638 11:46:01 chaining -- bdev/chaining.sh@205 -- # sequence=83332 00:35:17.638 11:46:01 chaining -- bdev/chaining.sh@206 -- # get_stat_bperf executed encrypt 00:35:17.638 11:46:01 chaining -- bdev/chaining.sh@48 -- # get_stat executed encrypt rpc_bperf 00:35:17.638 11:46:01 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:35:17.638 11:46:01 chaining -- bdev/chaining.sh@39 -- # event=executed 00:35:17.638 11:46:01 chaining -- bdev/chaining.sh@39 -- # opcode=encrypt 00:35:17.638 11:46:01 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_bperf 00:35:17.638 11:46:01 chaining -- bdev/chaining.sh@40 -- # [[ -z encrypt ]] 00:35:17.638 11:46:01 chaining -- bdev/chaining.sh@43 -- # rpc_bperf accel_get_stats 00:35:17.638 11:46:01 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "encrypt").executed' 00:35:17.638 11:46:01 chaining -- bdev/chaining.sh@22 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:35:17.897 11:46:01 chaining -- bdev/chaining.sh@206 -- # encrypt=41666 00:35:17.897 11:46:01 chaining -- bdev/chaining.sh@207 -- # get_stat_bperf executed decrypt 00:35:17.897 11:46:01 chaining -- bdev/chaining.sh@48 -- # get_stat executed decrypt rpc_bperf 00:35:17.897 11:46:01 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:35:17.897 11:46:01 chaining -- bdev/chaining.sh@39 -- # event=executed 00:35:17.897 11:46:01 chaining -- bdev/chaining.sh@39 -- # opcode=decrypt 00:35:17.897 11:46:01 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_bperf 00:35:17.897 11:46:01 chaining -- bdev/chaining.sh@40 -- # [[ -z decrypt ]] 00:35:17.897 11:46:01 chaining -- bdev/chaining.sh@43 -- # rpc_bperf accel_get_stats 00:35:17.897 11:46:01 chaining -- bdev/chaining.sh@22 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:35:17.897 11:46:01 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "decrypt").executed' 00:35:18.156 11:46:01 chaining -- bdev/chaining.sh@207 -- # decrypt=41666 00:35:18.156 11:46:01 chaining -- bdev/chaining.sh@208 -- # get_stat_bperf executed crc32c 00:35:18.156 11:46:01 chaining -- bdev/chaining.sh@48 -- # get_stat executed crc32c rpc_bperf 00:35:18.156 11:46:01 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:35:18.156 11:46:01 chaining -- bdev/chaining.sh@39 -- # event=executed 00:35:18.156 11:46:01 chaining -- bdev/chaining.sh@39 -- # opcode=crc32c 00:35:18.156 11:46:01 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_bperf 00:35:18.156 11:46:01 chaining -- bdev/chaining.sh@40 -- # [[ -z crc32c ]] 00:35:18.156 11:46:01 chaining -- bdev/chaining.sh@43 -- # rpc_bperf accel_get_stats 00:35:18.156 11:46:01 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "crc32c").executed' 00:35:18.156 11:46:01 chaining -- bdev/chaining.sh@22 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:35:18.415 11:46:01 chaining -- bdev/chaining.sh@208 -- # crc32c=83332 00:35:18.415 11:46:01 chaining -- bdev/chaining.sh@210 -- # (( sequence > 0 )) 00:35:18.415 11:46:01 chaining -- bdev/chaining.sh@211 -- # (( encrypt + decrypt == sequence )) 00:35:18.415 11:46:01 chaining -- bdev/chaining.sh@212 -- # (( encrypt + decrypt == crc32c )) 00:35:18.415 11:46:01 chaining -- bdev/chaining.sh@214 -- # killprocess 1077313 00:35:18.415 11:46:01 chaining -- common/autotest_common.sh@948 -- # '[' -z 1077313 ']' 00:35:18.415 11:46:01 chaining -- common/autotest_common.sh@952 -- # kill -0 1077313 00:35:18.415 11:46:01 chaining -- common/autotest_common.sh@953 -- # uname 00:35:18.415 11:46:01 chaining -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:35:18.415 11:46:01 chaining -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1077313 00:35:18.415 11:46:01 chaining -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:35:18.415 11:46:01 chaining -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:35:18.415 11:46:01 chaining -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1077313' 00:35:18.415 killing process with pid 1077313 00:35:18.415 11:46:01 chaining -- common/autotest_common.sh@967 -- # kill 1077313 00:35:18.415 Received shutdown signal, test time was about 5.000000 seconds 00:35:18.415 00:35:18.415 Latency(us) 00:35:18.415 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:35:18.415 =================================================================================================================== 00:35:18.415 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:35:18.415 11:46:01 chaining -- common/autotest_common.sh@972 -- # wait 1077313 00:35:18.674 11:46:02 chaining -- bdev/chaining.sh@219 -- # bperfpid=1078325 00:35:18.674 11:46:02 chaining -- bdev/chaining.sh@217 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bperf.sock -t 5 -w verify -o 65536 -q 32 --wait-for-rpc -z 00:35:18.674 11:46:02 chaining -- bdev/chaining.sh@221 -- # waitforlisten 1078325 /var/tmp/bperf.sock 00:35:18.674 11:46:02 chaining -- common/autotest_common.sh@829 -- # '[' -z 1078325 ']' 00:35:18.674 11:46:02 chaining -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bperf.sock 00:35:18.674 11:46:02 chaining -- common/autotest_common.sh@834 -- # local max_retries=100 00:35:18.674 11:46:02 chaining -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:35:18.674 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:35:18.674 11:46:02 chaining -- common/autotest_common.sh@838 -- # xtrace_disable 00:35:18.674 11:46:02 chaining -- common/autotest_common.sh@10 -- # set +x 00:35:18.674 [2024-07-15 11:46:02.236641] Starting SPDK v24.09-pre git sha1 e7cce062d / DPDK 24.03.0 initialization... 00:35:18.674 [2024-07-15 11:46:02.236710] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1078325 ] 00:35:18.933 [2024-07-15 11:46:02.365093] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:35:18.933 [2024-07-15 11:46:02.469748] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:35:19.867 11:46:03 chaining -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:35:19.867 11:46:03 chaining -- common/autotest_common.sh@862 -- # return 0 00:35:19.867 11:46:03 chaining -- bdev/chaining.sh@222 -- # rpc_bperf 00:35:19.867 11:46:03 chaining -- bdev/chaining.sh@22 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock 00:35:20.125 [2024-07-15 11:46:03.594991] vbdev_crypto_rpc.c: 115:rpc_bdev_crypto_create: *NOTICE*: Found key "key0" 00:35:20.125 nvme0n1 00:35:20.125 true 00:35:20.125 crypto0 00:35:20.125 11:46:03 chaining -- bdev/chaining.sh@231 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:35:20.125 Running I/O for 5 seconds... 00:35:25.391 00:35:25.391 Latency(us) 00:35:25.391 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:35:25.391 Job: crypto0 (Core Mask 0x1, workload: verify, depth: 32, IO size: 65536) 00:35:25.391 Verification LBA range: start 0x0 length 0x200 00:35:25.391 crypto0 : 5.01 1684.84 105.30 0.00 0.00 18614.22 1723.88 18919.96 00:35:25.391 =================================================================================================================== 00:35:25.391 Total : 1684.84 105.30 0.00 0.00 18614.22 1723.88 18919.96 00:35:25.391 0 00:35:25.391 11:46:08 chaining -- bdev/chaining.sh@233 -- # get_stat_bperf sequence_executed 00:35:25.391 11:46:08 chaining -- bdev/chaining.sh@48 -- # get_stat sequence_executed '' rpc_bperf 00:35:25.391 11:46:08 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:35:25.391 11:46:08 chaining -- bdev/chaining.sh@39 -- # event=sequence_executed 00:35:25.391 11:46:08 chaining -- bdev/chaining.sh@39 -- # opcode= 00:35:25.391 11:46:08 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_bperf 00:35:25.391 11:46:08 chaining -- bdev/chaining.sh@40 -- # [[ -z '' ]] 00:35:25.391 11:46:08 chaining -- bdev/chaining.sh@41 -- # rpc_bperf accel_get_stats 00:35:25.391 11:46:08 chaining -- bdev/chaining.sh@22 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:35:25.391 11:46:08 chaining -- bdev/chaining.sh@41 -- # jq -r .sequence_executed 00:35:25.648 11:46:09 chaining -- bdev/chaining.sh@233 -- # sequence=16872 00:35:25.648 11:46:09 chaining -- bdev/chaining.sh@234 -- # get_stat_bperf executed encrypt 00:35:25.648 11:46:09 chaining -- bdev/chaining.sh@48 -- # get_stat executed encrypt rpc_bperf 00:35:25.648 11:46:09 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:35:25.648 11:46:09 chaining -- bdev/chaining.sh@39 -- # event=executed 00:35:25.648 11:46:09 chaining -- bdev/chaining.sh@39 -- # opcode=encrypt 00:35:25.648 11:46:09 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_bperf 00:35:25.648 11:46:09 chaining -- bdev/chaining.sh@40 -- # [[ -z encrypt ]] 00:35:25.648 11:46:09 chaining -- bdev/chaining.sh@43 -- # rpc_bperf accel_get_stats 00:35:25.648 11:46:09 chaining -- bdev/chaining.sh@22 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:35:25.648 11:46:09 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "encrypt").executed' 00:35:25.905 11:46:09 chaining -- bdev/chaining.sh@234 -- # encrypt=8436 00:35:25.905 11:46:09 chaining -- bdev/chaining.sh@235 -- # get_stat_bperf executed decrypt 00:35:25.905 11:46:09 chaining -- bdev/chaining.sh@48 -- # get_stat executed decrypt rpc_bperf 00:35:25.905 11:46:09 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:35:25.905 11:46:09 chaining -- bdev/chaining.sh@39 -- # event=executed 00:35:25.905 11:46:09 chaining -- bdev/chaining.sh@39 -- # opcode=decrypt 00:35:25.905 11:46:09 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_bperf 00:35:25.905 11:46:09 chaining -- bdev/chaining.sh@40 -- # [[ -z decrypt ]] 00:35:25.905 11:46:09 chaining -- bdev/chaining.sh@43 -- # rpc_bperf accel_get_stats 00:35:25.905 11:46:09 chaining -- bdev/chaining.sh@22 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:35:25.905 11:46:09 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "decrypt").executed' 00:35:26.163 11:46:09 chaining -- bdev/chaining.sh@235 -- # decrypt=8436 00:35:26.163 11:46:09 chaining -- bdev/chaining.sh@236 -- # get_stat_bperf executed crc32c 00:35:26.163 11:46:09 chaining -- bdev/chaining.sh@48 -- # get_stat executed crc32c rpc_bperf 00:35:26.163 11:46:09 chaining -- bdev/chaining.sh@37 -- # local event opcode rpc 00:35:26.163 11:46:09 chaining -- bdev/chaining.sh@39 -- # event=executed 00:35:26.163 11:46:09 chaining -- bdev/chaining.sh@39 -- # opcode=crc32c 00:35:26.163 11:46:09 chaining -- bdev/chaining.sh@39 -- # rpc=rpc_bperf 00:35:26.163 11:46:09 chaining -- bdev/chaining.sh@40 -- # [[ -z crc32c ]] 00:35:26.163 11:46:09 chaining -- bdev/chaining.sh@43 -- # rpc_bperf accel_get_stats 00:35:26.163 11:46:09 chaining -- bdev/chaining.sh@22 -- # /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:35:26.163 11:46:09 chaining -- bdev/chaining.sh@44 -- # jq -r '.operations[] | select(.opcode == "crc32c").executed' 00:35:26.421 11:46:09 chaining -- bdev/chaining.sh@236 -- # crc32c=16872 00:35:26.421 11:46:09 chaining -- bdev/chaining.sh@238 -- # (( sequence > 0 )) 00:35:26.421 11:46:09 chaining -- bdev/chaining.sh@239 -- # (( encrypt + decrypt == sequence )) 00:35:26.421 11:46:09 chaining -- bdev/chaining.sh@240 -- # (( encrypt + decrypt == crc32c )) 00:35:26.421 11:46:09 chaining -- bdev/chaining.sh@242 -- # killprocess 1078325 00:35:26.421 11:46:09 chaining -- common/autotest_common.sh@948 -- # '[' -z 1078325 ']' 00:35:26.421 11:46:09 chaining -- common/autotest_common.sh@952 -- # kill -0 1078325 00:35:26.421 11:46:09 chaining -- common/autotest_common.sh@953 -- # uname 00:35:26.421 11:46:09 chaining -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:35:26.421 11:46:09 chaining -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1078325 00:35:26.421 11:46:09 chaining -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:35:26.421 11:46:09 chaining -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:35:26.421 11:46:09 chaining -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1078325' 00:35:26.421 killing process with pid 1078325 00:35:26.421 11:46:09 chaining -- common/autotest_common.sh@967 -- # kill 1078325 00:35:26.421 Received shutdown signal, test time was about 5.000000 seconds 00:35:26.421 00:35:26.421 Latency(us) 00:35:26.421 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:35:26.421 =================================================================================================================== 00:35:26.421 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:35:26.421 11:46:09 chaining -- common/autotest_common.sh@972 -- # wait 1078325 00:35:26.680 11:46:10 chaining -- bdev/chaining.sh@243 -- # nvmftestfini 00:35:26.680 11:46:10 chaining -- nvmf/common.sh@488 -- # nvmfcleanup 00:35:26.680 11:46:10 chaining -- nvmf/common.sh@117 -- # sync 00:35:26.680 11:46:10 chaining -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:35:26.680 11:46:10 chaining -- nvmf/common.sh@120 -- # set +e 00:35:26.680 11:46:10 chaining -- nvmf/common.sh@121 -- # for i in {1..20} 00:35:26.680 11:46:10 chaining -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:35:26.680 rmmod nvme_tcp 00:35:26.680 rmmod nvme_fabrics 00:35:26.680 rmmod nvme_keyring 00:35:26.680 11:46:10 chaining -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:35:26.680 11:46:10 chaining -- nvmf/common.sh@124 -- # set -e 00:35:26.680 11:46:10 chaining -- nvmf/common.sh@125 -- # return 0 00:35:26.680 11:46:10 chaining -- nvmf/common.sh@489 -- # '[' -n 1077125 ']' 00:35:26.680 11:46:10 chaining -- nvmf/common.sh@490 -- # killprocess 1077125 00:35:26.680 11:46:10 chaining -- common/autotest_common.sh@948 -- # '[' -z 1077125 ']' 00:35:26.680 11:46:10 chaining -- common/autotest_common.sh@952 -- # kill -0 1077125 00:35:26.680 11:46:10 chaining -- common/autotest_common.sh@953 -- # uname 00:35:26.680 11:46:10 chaining -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:35:26.680 11:46:10 chaining -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 1077125 00:35:26.680 11:46:10 chaining -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:35:26.680 11:46:10 chaining -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:35:26.680 11:46:10 chaining -- common/autotest_common.sh@966 -- # echo 'killing process with pid 1077125' 00:35:26.680 killing process with pid 1077125 00:35:26.680 11:46:10 chaining -- common/autotest_common.sh@967 -- # kill 1077125 00:35:26.680 11:46:10 chaining -- common/autotest_common.sh@972 -- # wait 1077125 00:35:26.938 11:46:10 chaining -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:35:26.938 11:46:10 chaining -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:35:26.938 11:46:10 chaining -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:35:26.938 11:46:10 chaining -- nvmf/common.sh@274 -- # [[ nvmf_tgt_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:35:26.938 11:46:10 chaining -- nvmf/common.sh@278 -- # remove_spdk_ns 00:35:26.938 11:46:10 chaining -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:35:26.938 11:46:10 chaining -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:35:26.938 11:46:10 chaining -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:35:26.938 11:46:10 chaining -- nvmf/common.sh@279 -- # ip -4 addr flush nvmf_init_if 00:35:26.938 11:46:10 chaining -- bdev/chaining.sh@245 -- # trap - SIGINT SIGTERM EXIT 00:35:26.938 00:35:26.938 real 0m46.226s 00:35:26.938 user 0m59.702s 00:35:26.938 sys 0m13.455s 00:35:26.938 11:46:10 chaining -- common/autotest_common.sh@1124 -- # xtrace_disable 00:35:26.938 11:46:10 chaining -- common/autotest_common.sh@10 -- # set +x 00:35:26.938 ************************************ 00:35:26.938 END TEST chaining 00:35:26.938 ************************************ 00:35:27.196 11:46:10 -- common/autotest_common.sh@1142 -- # return 0 00:35:27.196 11:46:10 -- spdk/autotest.sh@363 -- # [[ 0 -eq 1 ]] 00:35:27.196 11:46:10 -- spdk/autotest.sh@367 -- # [[ 0 -eq 1 ]] 00:35:27.196 11:46:10 -- spdk/autotest.sh@371 -- # [[ 0 -eq 1 ]] 00:35:27.196 11:46:10 -- spdk/autotest.sh@375 -- # [[ 0 -eq 1 ]] 00:35:27.196 11:46:10 -- spdk/autotest.sh@380 -- # trap - SIGINT SIGTERM EXIT 00:35:27.196 11:46:10 -- spdk/autotest.sh@382 -- # timing_enter post_cleanup 00:35:27.196 11:46:10 -- common/autotest_common.sh@722 -- # xtrace_disable 00:35:27.196 11:46:10 -- common/autotest_common.sh@10 -- # set +x 00:35:27.196 11:46:10 -- spdk/autotest.sh@383 -- # autotest_cleanup 00:35:27.196 11:46:10 -- common/autotest_common.sh@1392 -- # local autotest_es=0 00:35:27.196 11:46:10 -- common/autotest_common.sh@1393 -- # xtrace_disable 00:35:27.196 11:46:10 -- common/autotest_common.sh@10 -- # set +x 00:35:31.455 INFO: APP EXITING 00:35:31.455 INFO: killing all VMs 00:35:31.455 INFO: killing vhost app 00:35:31.455 INFO: EXIT DONE 00:35:34.735 0000:d7:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:d7:05.5 00:35:34.735 0000:85:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:85:05.5 00:35:34.735 Waiting for block devices as requested 00:35:34.735 0000:5e:00.0 (8086 0b60): vfio-pci -> nvme 00:35:34.735 0000:00:04.7 (8086 2021): vfio-pci -> ioatdma 00:35:34.735 0000:00:04.6 (8086 2021): vfio-pci -> ioatdma 00:35:34.995 0000:00:04.5 (8086 2021): vfio-pci -> ioatdma 00:35:34.995 0000:00:04.4 (8086 2021): vfio-pci -> ioatdma 00:35:34.995 0000:00:04.3 (8086 2021): vfio-pci -> ioatdma 00:35:35.252 0000:00:04.2 (8086 2021): vfio-pci -> ioatdma 00:35:35.252 0000:00:04.1 (8086 2021): vfio-pci -> ioatdma 00:35:35.252 0000:00:04.0 (8086 2021): vfio-pci -> ioatdma 00:35:35.511 0000:80:04.7 (8086 2021): vfio-pci -> ioatdma 00:35:35.511 0000:80:04.6 (8086 2021): vfio-pci -> ioatdma 00:35:35.511 0000:80:04.5 (8086 2021): vfio-pci -> ioatdma 00:35:35.769 0000:80:04.4 (8086 2021): vfio-pci -> ioatdma 00:35:35.769 0000:80:04.3 (8086 2021): vfio-pci -> ioatdma 00:35:35.769 0000:80:04.2 (8086 2021): vfio-pci -> ioatdma 00:35:36.028 0000:80:04.1 (8086 2021): vfio-pci -> ioatdma 00:35:36.028 0000:80:04.0 (8086 2021): vfio-pci -> ioatdma 00:35:40.212 0000:d7:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:d7:05.5 00:35:40.212 0000:85:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:85:05.5 00:35:40.212 Cleaning 00:35:40.212 Removing: /var/run/dpdk/spdk0/config 00:35:40.212 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-0 00:35:40.212 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-1 00:35:40.212 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-2 00:35:40.212 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-3 00:35:40.212 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-0 00:35:40.212 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-1 00:35:40.212 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-2 00:35:40.212 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-3 00:35:40.212 Removing: /var/run/dpdk/spdk0/fbarray_memzone 00:35:40.212 Removing: /var/run/dpdk/spdk0/hugepage_info 00:35:40.212 Removing: /dev/shm/nvmf_trace.0 00:35:40.212 Removing: /dev/shm/spdk_tgt_trace.pid821693 00:35:40.212 Removing: /var/run/dpdk/spdk0 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1002115 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1004531 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1010655 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1012969 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1019539 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1021952 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1026110 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1026468 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1026817 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1027177 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1027769 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1028379 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1029225 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1029669 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1031272 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1032877 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1034545 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1035952 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1037624 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1039331 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1040935 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1042522 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1043288 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1043694 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1045825 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1047678 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1049523 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1050587 00:35:40.212 Removing: /var/run/dpdk/spdk_pid1051797 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1052356 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1052379 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1052562 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1052814 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1052977 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1054075 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1055719 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1057249 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1057972 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1058758 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1059047 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1059069 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1059195 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1060041 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1060679 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1061109 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1063272 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1065125 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1066985 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1068506 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1069602 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1070211 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1070331 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1074234 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1074444 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1074586 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1074677 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1074884 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1075095 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1075970 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1077313 00:35:40.213 Removing: /var/run/dpdk/spdk_pid1078325 00:35:40.213 Removing: /var/run/dpdk/spdk_pid820809 00:35:40.213 Removing: /var/run/dpdk/spdk_pid821693 00:35:40.213 Removing: /var/run/dpdk/spdk_pid822251 00:35:40.213 Removing: /var/run/dpdk/spdk_pid822989 00:35:40.213 Removing: /var/run/dpdk/spdk_pid823174 00:35:40.213 Removing: /var/run/dpdk/spdk_pid823934 00:35:40.213 Removing: /var/run/dpdk/spdk_pid824109 00:35:40.213 Removing: /var/run/dpdk/spdk_pid824391 00:35:40.213 Removing: /var/run/dpdk/spdk_pid827002 00:35:40.213 Removing: /var/run/dpdk/spdk_pid828852 00:35:40.213 Removing: /var/run/dpdk/spdk_pid829087 00:35:40.213 Removing: /var/run/dpdk/spdk_pid829488 00:35:40.213 Removing: /var/run/dpdk/spdk_pid829740 00:35:40.213 Removing: /var/run/dpdk/spdk_pid830141 00:35:40.213 Removing: /var/run/dpdk/spdk_pid830347 00:35:40.213 Removing: /var/run/dpdk/spdk_pid830539 00:35:40.213 Removing: /var/run/dpdk/spdk_pid830767 00:35:40.213 Removing: /var/run/dpdk/spdk_pid831513 00:35:40.213 Removing: /var/run/dpdk/spdk_pid834215 00:35:40.213 Removing: /var/run/dpdk/spdk_pid834414 00:35:40.213 Removing: /var/run/dpdk/spdk_pid834672 00:35:40.213 Removing: /var/run/dpdk/spdk_pid834984 00:35:40.213 Removing: /var/run/dpdk/spdk_pid835053 00:35:40.213 Removing: /var/run/dpdk/spdk_pid835280 00:35:40.213 Removing: /var/run/dpdk/spdk_pid835478 00:35:40.213 Removing: /var/run/dpdk/spdk_pid835674 00:35:40.213 Removing: /var/run/dpdk/spdk_pid835874 00:35:40.213 Removing: /var/run/dpdk/spdk_pid836080 00:35:40.213 Removing: /var/run/dpdk/spdk_pid836398 00:35:40.213 Removing: /var/run/dpdk/spdk_pid836624 00:35:40.213 Removing: /var/run/dpdk/spdk_pid836823 00:35:40.213 Removing: /var/run/dpdk/spdk_pid837017 00:35:40.213 Removing: /var/run/dpdk/spdk_pid837217 00:35:40.213 Removing: /var/run/dpdk/spdk_pid837418 00:35:40.213 Removing: /var/run/dpdk/spdk_pid837716 00:35:40.213 Removing: /var/run/dpdk/spdk_pid837966 00:35:40.213 Removing: /var/run/dpdk/spdk_pid838165 00:35:40.213 Removing: /var/run/dpdk/spdk_pid838359 00:35:40.213 Removing: /var/run/dpdk/spdk_pid838563 00:35:40.213 Removing: /var/run/dpdk/spdk_pid838757 00:35:40.213 Removing: /var/run/dpdk/spdk_pid839016 00:35:40.213 Removing: /var/run/dpdk/spdk_pid839319 00:35:40.213 Removing: /var/run/dpdk/spdk_pid839515 00:35:40.213 Removing: /var/run/dpdk/spdk_pid839713 00:35:40.213 Removing: /var/run/dpdk/spdk_pid839959 00:35:40.213 Removing: /var/run/dpdk/spdk_pid840281 00:35:40.213 Removing: /var/run/dpdk/spdk_pid840644 00:35:40.213 Removing: /var/run/dpdk/spdk_pid840942 00:35:40.213 Removing: /var/run/dpdk/spdk_pid841223 00:35:40.213 Removing: /var/run/dpdk/spdk_pid841590 00:35:40.213 Removing: /var/run/dpdk/spdk_pid841942 00:35:40.213 Removing: /var/run/dpdk/spdk_pid842155 00:35:40.213 Removing: /var/run/dpdk/spdk_pid842389 00:35:40.213 Removing: /var/run/dpdk/spdk_pid842645 00:35:40.213 Removing: /var/run/dpdk/spdk_pid843115 00:35:40.213 Removing: /var/run/dpdk/spdk_pid843491 00:35:40.213 Removing: /var/run/dpdk/spdk_pid843683 00:35:40.213 Removing: /var/run/dpdk/spdk_pid847844 00:35:40.213 Removing: /var/run/dpdk/spdk_pid849547 00:35:40.213 Removing: /var/run/dpdk/spdk_pid851246 00:35:40.213 Removing: /var/run/dpdk/spdk_pid852136 00:35:40.213 Removing: /var/run/dpdk/spdk_pid853204 00:35:40.213 Removing: /var/run/dpdk/spdk_pid853498 00:35:40.213 Removing: /var/run/dpdk/spdk_pid853597 00:35:40.213 Removing: /var/run/dpdk/spdk_pid853638 00:35:40.213 Removing: /var/run/dpdk/spdk_pid858034 00:35:40.213 Removing: /var/run/dpdk/spdk_pid858476 00:35:40.213 Removing: /var/run/dpdk/spdk_pid859514 00:35:40.213 Removing: /var/run/dpdk/spdk_pid859744 00:35:40.213 Removing: /var/run/dpdk/spdk_pid864767 00:35:40.213 Removing: /var/run/dpdk/spdk_pid866365 00:35:40.213 Removing: /var/run/dpdk/spdk_pid867334 00:35:40.213 Removing: /var/run/dpdk/spdk_pid871570 00:35:40.213 Removing: /var/run/dpdk/spdk_pid873212 00:35:40.213 Removing: /var/run/dpdk/spdk_pid874181 00:35:40.213 Removing: /var/run/dpdk/spdk_pid878170 00:35:40.213 Removing: /var/run/dpdk/spdk_pid880509 00:35:40.213 Removing: /var/run/dpdk/spdk_pid881601 00:35:40.213 Removing: /var/run/dpdk/spdk_pid891380 00:35:40.213 Removing: /var/run/dpdk/spdk_pid893564 00:35:40.213 Removing: /var/run/dpdk/spdk_pid894576 00:35:40.213 Removing: /var/run/dpdk/spdk_pid904277 00:35:40.213 Removing: /var/run/dpdk/spdk_pid906368 00:35:40.213 Removing: /var/run/dpdk/spdk_pid907347 00:35:40.213 Removing: /var/run/dpdk/spdk_pid917573 00:35:40.213 Removing: /var/run/dpdk/spdk_pid920841 00:35:40.213 Removing: /var/run/dpdk/spdk_pid921868 00:35:40.213 Removing: /var/run/dpdk/spdk_pid932612 00:35:40.213 Removing: /var/run/dpdk/spdk_pid934993 00:35:40.213 Removing: /var/run/dpdk/spdk_pid936146 00:35:40.213 Removing: /var/run/dpdk/spdk_pid947546 00:35:40.213 Removing: /var/run/dpdk/spdk_pid949991 00:35:40.213 Removing: /var/run/dpdk/spdk_pid950993 00:35:40.213 Removing: /var/run/dpdk/spdk_pid961900 00:35:40.213 Removing: /var/run/dpdk/spdk_pid966088 00:35:40.213 Removing: /var/run/dpdk/spdk_pid967229 00:35:40.213 Removing: /var/run/dpdk/spdk_pid968376 00:35:40.213 Removing: /var/run/dpdk/spdk_pid971434 00:35:40.213 Removing: /var/run/dpdk/spdk_pid976615 00:35:40.213 Removing: /var/run/dpdk/spdk_pid979138 00:35:40.213 Removing: /var/run/dpdk/spdk_pid983787 00:35:40.213 Removing: /var/run/dpdk/spdk_pid987019 00:35:40.213 Removing: /var/run/dpdk/spdk_pid993073 00:35:40.213 Removing: /var/run/dpdk/spdk_pid995796 00:35:40.213 Clean 00:35:40.471 11:46:23 -- common/autotest_common.sh@1451 -- # return 0 00:35:40.471 11:46:23 -- spdk/autotest.sh@384 -- # timing_exit post_cleanup 00:35:40.471 11:46:23 -- common/autotest_common.sh@728 -- # xtrace_disable 00:35:40.471 11:46:23 -- common/autotest_common.sh@10 -- # set +x 00:35:40.471 11:46:23 -- spdk/autotest.sh@386 -- # timing_exit autotest 00:35:40.471 11:46:23 -- common/autotest_common.sh@728 -- # xtrace_disable 00:35:40.471 11:46:23 -- common/autotest_common.sh@10 -- # set +x 00:35:40.471 11:46:23 -- spdk/autotest.sh@387 -- # chmod a+r /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/timing.txt 00:35:40.471 11:46:23 -- spdk/autotest.sh@389 -- # [[ -f /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/udev.log ]] 00:35:40.471 11:46:23 -- spdk/autotest.sh@389 -- # rm -f /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/udev.log 00:35:40.471 11:46:23 -- spdk/autotest.sh@391 -- # hash lcov 00:35:40.471 11:46:23 -- spdk/autotest.sh@391 -- # [[ CC_TYPE=gcc == *\c\l\a\n\g* ]] 00:35:40.471 11:46:23 -- spdk/autotest.sh@393 -- # hostname 00:35:40.471 11:46:23 -- spdk/autotest.sh@393 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -c -d /var/jenkins/workspace/crypto-phy-autotest/spdk -t spdk-wfp-50 -o /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/cov_test.info 00:35:40.729 geninfo: WARNING: invalid characters removed from testname! 00:36:07.262 11:46:48 -- spdk/autotest.sh@394 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -a /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/cov_base.info -a /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/cov_test.info -o /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/cov_total.info 00:36:08.639 11:46:52 -- spdk/autotest.sh@395 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -r /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/cov_total.info '*/dpdk/*' -o /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/cov_total.info 00:36:11.923 11:46:54 -- spdk/autotest.sh@396 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -r /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/cov_total.info '/usr/*' -o /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/cov_total.info 00:36:14.453 11:46:57 -- spdk/autotest.sh@397 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -r /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/cov_total.info '*/examples/vmd/*' -o /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/cov_total.info 00:36:17.042 11:47:00 -- spdk/autotest.sh@398 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -r /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/cov_total.info '*/app/spdk_lspci/*' -o /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/cov_total.info 00:36:19.567 11:47:02 -- spdk/autotest.sh@399 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -r /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/cov_total.info '*/app/spdk_top/*' -o /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/cov_total.info 00:36:22.097 11:47:05 -- spdk/autotest.sh@400 -- # rm -f cov_base.info cov_test.info OLD_STDOUT OLD_STDERR 00:36:22.097 11:47:05 -- common/autobuild_common.sh@15 -- $ source /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/common.sh 00:36:22.097 11:47:05 -- scripts/common.sh@508 -- $ [[ -e /bin/wpdk_common.sh ]] 00:36:22.097 11:47:05 -- scripts/common.sh@516 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:36:22.097 11:47:05 -- scripts/common.sh@517 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:36:22.097 11:47:05 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:22.097 11:47:05 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:22.097 11:47:05 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:22.097 11:47:05 -- paths/export.sh@5 -- $ export PATH 00:36:22.097 11:47:05 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:22.097 11:47:05 -- common/autobuild_common.sh@443 -- $ out=/var/jenkins/workspace/crypto-phy-autotest/spdk/../output 00:36:22.097 11:47:05 -- common/autobuild_common.sh@444 -- $ date +%s 00:36:22.097 11:47:05 -- common/autobuild_common.sh@444 -- $ mktemp -dt spdk_1721036825.XXXXXX 00:36:22.097 11:47:05 -- common/autobuild_common.sh@444 -- $ SPDK_WORKSPACE=/tmp/spdk_1721036825.59vPrH 00:36:22.097 11:47:05 -- common/autobuild_common.sh@446 -- $ [[ -n '' ]] 00:36:22.097 11:47:05 -- common/autobuild_common.sh@450 -- $ '[' -n '' ']' 00:36:22.097 11:47:05 -- common/autobuild_common.sh@453 -- $ scanbuild_exclude='--exclude /var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/' 00:36:22.097 11:47:05 -- common/autobuild_common.sh@457 -- $ scanbuild_exclude+=' --exclude /var/jenkins/workspace/crypto-phy-autotest/spdk/xnvme --exclude /tmp' 00:36:22.097 11:47:05 -- common/autobuild_common.sh@459 -- $ scanbuild='scan-build -o /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/scan-build-tmp --exclude /var/jenkins/workspace/crypto-phy-autotest/spdk/dpdk/ --exclude /var/jenkins/workspace/crypto-phy-autotest/spdk/xnvme --exclude /tmp --status-bugs' 00:36:22.097 11:47:05 -- common/autobuild_common.sh@460 -- $ get_config_params 00:36:22.097 11:47:05 -- common/autotest_common.sh@396 -- $ xtrace_disable 00:36:22.097 11:47:05 -- common/autotest_common.sh@10 -- $ set +x 00:36:22.097 11:47:05 -- common/autobuild_common.sh@460 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --with-vbdev-compress --with-dpdk-compressdev --with-crypto --enable-ubsan --enable-coverage --with-ublk' 00:36:22.097 11:47:05 -- common/autobuild_common.sh@462 -- $ start_monitor_resources 00:36:22.097 11:47:05 -- pm/common@17 -- $ local monitor 00:36:22.097 11:47:05 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:36:22.097 11:47:05 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:36:22.097 11:47:05 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:36:22.097 11:47:05 -- pm/common@21 -- $ date +%s 00:36:22.097 11:47:05 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:36:22.097 11:47:05 -- pm/common@21 -- $ date +%s 00:36:22.097 11:47:05 -- pm/common@25 -- $ sleep 1 00:36:22.097 11:47:05 -- pm/common@21 -- $ date +%s 00:36:22.097 11:47:05 -- pm/common@21 -- $ date +%s 00:36:22.097 11:47:05 -- pm/common@21 -- $ /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm/collect-cpu-load -d /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power -l -p monitor.autopackage.sh.1721036825 00:36:22.097 11:47:05 -- pm/common@21 -- $ /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm/collect-vmstat -d /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power -l -p monitor.autopackage.sh.1721036825 00:36:22.097 11:47:05 -- pm/common@21 -- $ /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm/collect-cpu-temp -d /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power -l -p monitor.autopackage.sh.1721036825 00:36:22.097 11:47:05 -- pm/common@21 -- $ sudo -E /var/jenkins/workspace/crypto-phy-autotest/spdk/scripts/perf/pm/collect-bmc-pm -d /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power -l -p monitor.autopackage.sh.1721036825 00:36:22.097 Redirecting to /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/monitor.autopackage.sh.1721036825_collect-vmstat.pm.log 00:36:22.097 Redirecting to /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/monitor.autopackage.sh.1721036825_collect-cpu-load.pm.log 00:36:22.097 Redirecting to /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/monitor.autopackage.sh.1721036825_collect-cpu-temp.pm.log 00:36:22.097 Redirecting to /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/monitor.autopackage.sh.1721036825_collect-bmc-pm.bmc.pm.log 00:36:23.049 11:47:06 -- common/autobuild_common.sh@463 -- $ trap stop_monitor_resources EXIT 00:36:23.049 11:47:06 -- spdk/autopackage.sh@10 -- $ MAKEFLAGS=-j72 00:36:23.049 11:47:06 -- spdk/autopackage.sh@11 -- $ cd /var/jenkins/workspace/crypto-phy-autotest/spdk 00:36:23.049 11:47:06 -- spdk/autopackage.sh@13 -- $ [[ 0 -eq 1 ]] 00:36:23.049 11:47:06 -- spdk/autopackage.sh@18 -- $ [[ 0 -eq 0 ]] 00:36:23.049 11:47:06 -- spdk/autopackage.sh@19 -- $ timing_finish 00:36:23.049 11:47:06 -- common/autotest_common.sh@734 -- $ flamegraph=/usr/local/FlameGraph/flamegraph.pl 00:36:23.049 11:47:06 -- common/autotest_common.sh@735 -- $ '[' -x /usr/local/FlameGraph/flamegraph.pl ']' 00:36:23.050 11:47:06 -- common/autotest_common.sh@737 -- $ /usr/local/FlameGraph/flamegraph.pl --title 'Build Timing' --nametype Step: --countname seconds /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/timing.txt 00:36:23.050 11:47:06 -- spdk/autopackage.sh@20 -- $ exit 0 00:36:23.050 11:47:06 -- spdk/autopackage.sh@1 -- $ stop_monitor_resources 00:36:23.050 11:47:06 -- pm/common@29 -- $ signal_monitor_resources TERM 00:36:23.050 11:47:06 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:36:23.050 11:47:06 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:36:23.050 11:47:06 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/collect-cpu-load.pid ]] 00:36:23.050 11:47:06 -- pm/common@44 -- $ pid=1088830 00:36:23.050 11:47:06 -- pm/common@50 -- $ kill -TERM 1088830 00:36:23.050 11:47:06 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:36:23.050 11:47:06 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/collect-vmstat.pid ]] 00:36:23.050 11:47:06 -- pm/common@44 -- $ pid=1088832 00:36:23.050 11:47:06 -- pm/common@50 -- $ kill -TERM 1088832 00:36:23.050 11:47:06 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:36:23.050 11:47:06 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/collect-cpu-temp.pid ]] 00:36:23.050 11:47:06 -- pm/common@44 -- $ pid=1088834 00:36:23.050 11:47:06 -- pm/common@50 -- $ kill -TERM 1088834 00:36:23.050 11:47:06 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:36:23.050 11:47:06 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/collect-bmc-pm.pid ]] 00:36:23.050 11:47:06 -- pm/common@44 -- $ pid=1088857 00:36:23.050 11:47:06 -- pm/common@50 -- $ sudo -E kill -TERM 1088857 00:36:23.050 + [[ -n 706194 ]] 00:36:23.050 + sudo kill 706194 00:36:23.063 [Pipeline] } 00:36:23.086 [Pipeline] // stage 00:36:23.092 [Pipeline] } 00:36:23.112 [Pipeline] // timeout 00:36:23.117 [Pipeline] } 00:36:23.136 [Pipeline] // catchError 00:36:23.143 [Pipeline] } 00:36:23.160 [Pipeline] // wrap 00:36:23.167 [Pipeline] } 00:36:23.184 [Pipeline] // catchError 00:36:23.193 [Pipeline] stage 00:36:23.195 [Pipeline] { (Epilogue) 00:36:23.208 [Pipeline] catchError 00:36:23.211 [Pipeline] { 00:36:23.226 [Pipeline] echo 00:36:23.227 Cleanup processes 00:36:23.233 [Pipeline] sh 00:36:23.515 + sudo pgrep -af /var/jenkins/workspace/crypto-phy-autotest/spdk 00:36:23.515 1088933 /usr/bin/ipmitool sdr dump /var/jenkins/workspace/crypto-phy-autotest/spdk/../output/power/sdr.cache 00:36:23.515 1089150 sudo pgrep -af /var/jenkins/workspace/crypto-phy-autotest/spdk 00:36:23.531 [Pipeline] sh 00:36:23.815 ++ sudo pgrep -af /var/jenkins/workspace/crypto-phy-autotest/spdk 00:36:23.815 ++ grep -v 'sudo pgrep' 00:36:23.815 ++ awk '{print $1}' 00:36:23.815 + sudo kill -9 1088933 00:36:23.829 [Pipeline] sh 00:36:24.111 + jbp/jenkins/jjb-config/jobs/scripts/compress_artifacts.sh 00:36:36.332 [Pipeline] sh 00:36:36.621 + jbp/jenkins/jjb-config/jobs/scripts/check_artifacts_size.sh 00:36:36.621 Artifacts sizes are good 00:36:36.637 [Pipeline] archiveArtifacts 00:36:36.645 Archiving artifacts 00:36:36.814 [Pipeline] sh 00:36:37.124 + sudo chown -R sys_sgci /var/jenkins/workspace/crypto-phy-autotest 00:36:37.141 [Pipeline] cleanWs 00:36:37.151 [WS-CLEANUP] Deleting project workspace... 00:36:37.151 [WS-CLEANUP] Deferred wipeout is used... 00:36:37.158 [WS-CLEANUP] done 00:36:37.160 [Pipeline] } 00:36:37.180 [Pipeline] // catchError 00:36:37.194 [Pipeline] sh 00:36:37.476 + logger -p user.info -t JENKINS-CI 00:36:37.485 [Pipeline] } 00:36:37.504 [Pipeline] // stage 00:36:37.510 [Pipeline] } 00:36:37.528 [Pipeline] // node 00:36:37.533 [Pipeline] End of Pipeline 00:36:37.563 Finished: SUCCESS